Mental Attributes Profiling System: Difference between revisions

From Future Worlds Center Wiki
Jump to navigationJump to search
no edit summary
No edit summary
No edit summary
 
(47 intermediate revisions by 3 users not shown)
Line 2: Line 2:
       |acronym=MAPS             
       |acronym=MAPS             
       |logo=MAPS Blue Logo.jpg
       |logo=MAPS Blue Logo.jpg
       |project_title= Mental Attributes Profiling System 
       |project_title=[[Development of methodology and software for the diagnosis and treatement of learning difficulties with emphasis in dyslexia]] <br>[[Educational Hypermedia (Hibernation)]]<br>[[Dynamic Visualizations in Instruction]]<br>[[Unified e-Hoop approach to learning differences]]
       |contract_number=
       |contract_number=
       |funding_period=
       |funding_period=1999-2016
       |coordinator=
       |coordinator=[[Cyprus Neuroscience and Technology Institute]]
       |funding_agency=[[Research Promotion Foundation]]
       |funding_agency=[[Research Promotion Foundation]]<br>Leibniz Association, Germany<br>European Commission
       |total_cost=
       |total_cost=ca. 700K
       |partners=[[Cyprus Dyslexia Association]]
       |partners=[[Cyprus Dyslexia Association]]
       |website=
       |website=http://mentalmaps.org
}}     
}}     


   
   
The '''Mental Attributes Profiling System''' (MAPS) is a language independent screening test, i.e., relies on cognitive rather than language-based measures, capable of predicting children at risk (possible dyslexics) and equipping teachers with a profile of their mental abilities so as to design personalized remediation programs.
The '''Mental Attributes Profiling System''' (MAPS) is a language independent screening test, i.e., relies on cognitive rather than language-based measures, capable of predicting children at risk (possible dyslexics) and equipping teachers with a profile of their mental abilities so as to design personalized remediation programs.
MAPS has been developed as a research project within the [[Cyprus Neuroscience and Technology Institute]].
In 2017, the IP was transferred to a spin off company, [[GNOUS Labs Ltd.]].
        
        
__TOC__
__TOC__
Line 20: Line 23:
The MAPS (Mental Attributes Profiling System) cognitive test is a battery of more than 12 validated computer-based video-game-like tests that assess the learning abilities of pre-elementary and elementary age school children. The non-verbal group consists of language-independent tests. Snap shots of three tests are shown in the figures below.  
The MAPS (Mental Attributes Profiling System) cognitive test is a battery of more than 12 validated computer-based video-game-like tests that assess the learning abilities of pre-elementary and elementary age school children. The non-verbal group consists of language-independent tests. Snap shots of three tests are shown in the figures below.  


<ul>The following eight tests have been used for the multimodal evaluation reported here:</ul>
The following eight tests have been used for the multimodal evaluation reported here:</ul>
 


==Categorization==
==Categorization==
Line 33: Line 37:
# An animal (or fish) to be placed in one of three possible environments: sea, sky, or open fields;
# An animal (or fish) to be placed in one of three possible environments: sea, sky, or open fields;
# Objects usually found in a home environment, to be “placed” in one of three rooms of a house (office, kitchen, or bathroom), i.e., as in the example snapshot shown in the illustration. The software measured the time to response and whether the move was correct or false. For this study we used only the average response time (i.e., the cumulative response time of each test divided by 5: (ctt1+ctt2+...ctt5)/5 in Table 1), as it has been shown previously to have a higher correlation than the Correct/False values.
# Objects usually found in a home environment, to be “placed” in one of three rooms of a house (office, kitchen, or bathroom), i.e., as in the example snapshot shown in the illustration. The software measured the time to response and whether the move was correct or false. For this study we used only the average response time (i.e., the cumulative response time of each test divided by 5: (ctt1+ctt2+...ctt5)/5 in Table 1), as it has been shown previously to have a higher correlation than the Correct/False values.


==Orientation: “Left-Right Game” Lateral awareness Test==
==Orientation: “Left-Right Game” Lateral awareness Test==
[[File:Laterization.jpg|left|thumb|upright=1|alt=Logo.|Ability to orientate: Left/Right]]
[[File:Laterization.jpg|left|thumb|upright=1|alt=Logo.|Ability to orientate: Left/Right]]
This test provides two types of measures. One, it evaluates the child’s ability to make left-right discriminations on his own body. During the first part, the test shows a child “sitting” in the same orientation as the subject (i.e., the subject sees on the screen the back of the child) in front of two objects, one on his/her upper left and the other on his/her upper right visual field. The computer asks the subject to “grab” the left or right object using his/her left or right hand. What is measured is (1) the time taken to select an arm (by clicking at the shoulder), and (2) whether the selected arm was correct or wrong. The same procedure is repeated during the second part of the test, in which the orientation of the child on the screen is reversed, i.e., the child on the screen is facing the subject. The second type of measures are derived from Piaget’s (19) tests to evaluate awareness of right-left relations outside our own body, i.e., in the environment. To evaluate this ability, the test measures (1) the time taken by the subject to decide which object to “grab” and (2) the number of correct/wrong decisions.  
This test provides two types of measures. One, it evaluates the child’s ability to make left-right discriminations on his own body. During the first part, the test shows a child “sitting” in the same orientation as the subject (i.e., the subject sees on the screen the back of the child) in front of two objects, one on his/her upper left and the other on his/her upper right visual field. The computer asks the subject to “grab” the left or right object using his/her left or right hand. What is measured is (1) the time taken to select an arm (by clicking at the shoulder), and (2) whether the selected arm was correct or wrong. The same procedure is repeated during the second part of the test, in which the orientation of the child on the screen is reversed, i.e., the child on the screen is facing the subject. The second type of measures are derived from Piaget’s (19) tests to evaluate awareness of right-left relations outside our own body, i.e., in the environment. To evaluate this ability, the test measures (1) the time taken by the subject to decide which object to “grab” and (2) the number of correct/wrong decisions.  


==Navigation==
==Navigation==
Line 42: Line 50:
The Navigation test consists of an 8x8 matrix of small pieces of cheese and a mouse. The computer instructs the subject to move the mouse in one of eight possible directions to “eat” the corresponding piece of cheese: Up, Right-up, Right, Right-Down, Down, Left-down, Left, and Left-up. The instructions are given both visually (with the aid of a small arrow) and/or orally. For the tests reported here the instructions were given only orally. The software measured the total number of correct responses and the total number of correct responses divided by total number of trials.
The Navigation test consists of an 8x8 matrix of small pieces of cheese and a mouse. The computer instructs the subject to move the mouse in one of eight possible directions to “eat” the corresponding piece of cheese: Up, Right-up, Right, Right-Down, Down, Left-down, Left, and Left-up. The instructions are given both visually (with the aid of a small arrow) and/or orally. For the tests reported here the instructions were given only orally. The software measured the total number of correct responses and the total number of correct responses divided by total number of trials.


<br>


<br>
<br>
Line 47: Line 56:
<br>
<br>


<br>
 
 


==Sequencing==
==Sequencing==
Line 53: Line 63:
[[File:Sequencg2.jpg|right|thumb|upright=1|alt=Logo.|Sequencing of 'broken' events]]
[[File:Sequencg2.jpg|right|thumb|upright=1|alt=Logo.|Sequencing of 'broken' events]]
This test was inspired by the “connect the puzzle” games. Different objects or animals appear in two-, three-, four-, or five pieces and the subject is requested to “drag” the pieces and place them in the right order to complete the picture. The second part of the test presents pictures, which represent different stages of a temporal process. The subject is expected to put them in the correct chronological order. The test measured the total time required to complete each sub-section of the test (the test was comprised of six different types of exercises divided by six (i.e., (stt1+stt2+...stt6)/6 in Tabl.1).
This test was inspired by the “connect the puzzle” games. Different objects or animals appear in two-, three-, four-, or five pieces and the subject is requested to “drag” the pieces and place them in the right order to complete the picture. The second part of the test presents pictures, which represent different stages of a temporal process. The subject is expected to put them in the correct chronological order. The test measured the total time required to complete each sub-section of the test (the test was comprised of six different types of exercises divided by six (i.e., (stt1+stt2+...stt6)/6 in Tabl.1).
<br>
<br>


<br>
<br>
Line 67: Line 81:


<br>
<br>
<br>


==Auditory memory==
==Auditory memory==
Line 77: Line 94:
* AMTc: Total number of correct responses
* AMTc: Total number of correct responses
* AMTcw: weighted number of correct digits (i.e., all correctly types digits even when the sequence was wrong) were used as measures of auditory digit span.
* AMTcw: weighted number of correct digits (i.e., all correctly types digits even when the sequence was wrong) were used as measures of auditory digit span.


==Auditory discrimination test==
==Auditory discrimination test==
Line 93: Line 111:
# γ-γγ/γκ
# γ-γγ/γκ
# ξ-κς.  
# ξ-κς.  


In additional sets, the test also evaluates the ability of the child to differentiate between the same letter combinations when they are embedded in non-sense words. The test kept record of the time and correctness of each instance. In Tabl.1 adtt(1), adtt(2),...adtt(n) are used to denote the time taken to respond to each card presented.
In additional sets, the test also evaluates the ability of the child to differentiate between the same letter combinations when they are embedded in non-sense words. The test kept record of the time and correctness of each instance. In Tabl.1 adtt(1), adtt(2),...adtt(n) are used to denote the time taken to respond to each card presented.


==Visual discrimination==
==Visual discrimination==
Modeled using the popular “Find the differences” game, this test presents three pictures very similar to one another and one that must be matched. The exercise is repeated four times with four different pictures. The total number of correct responses is used as an index of visual discrimination.  
Modeled using the popular “Find the differences” game, this test presents three pictures very similar to one another and one that must be matched. The exercise is repeated four times with four different pictures. The total number of correct responses is used as an index of visual discrimination.  
==Grants and Funding of MAPS==
{| class="wikitable"
! Grants
! Year
! Amount (€)
! Source
|-
| DYSLEXIA
| 1999-2001
| style="text-align:right;" | 44,803
| Research Promotion Foundation
|-
| [[Dynamic Visualizations in Instruction|Educational Hypermedia]]
| 2007
| style="text-align:right;" | 44,460
| Leibniz Association, Germany
|-
| Dynamic Visualizations in Instruction
| 2007-2009
| style="text-align:right;" | 51,300
| Leibniz Association, Germany
|-
| [[E-Hoop|Unified e-Hoop approach to learning differences]]
| 2012-2016
| style="text-align:right;" | 562,599<SUP>1</SUP>
| European Commission
|}
Note 1: Only 25% of this amount was exclusively used for R&D.




Line 105: Line 157:
* Haddad, C. and Laouris, Y. (2011). The ability of children with mild learning disabilities to encode emotions through facial expressions. In Towards Autonomous, adaptive and context-aware multimodal interfaces: Theoretical and Practical Issues. A. Esposito, Esposito, A.M., Martone, R., Müller, V.C., Scarpetta, G  (Eds.): COST 2102 Int. Training School 2010, Lecture Notes In Artificial Intelligence Springer-Verlag, Berlin, Heidelberg 6456, 387-402.
* Haddad, C. and Laouris, Y. (2011). The ability of children with mild learning disabilities to encode emotions through facial expressions. In Towards Autonomous, adaptive and context-aware multimodal interfaces: Theoretical and Practical Issues. A. Esposito, Esposito, A.M., Martone, R., Müller, V.C., Scarpetta, G  (Eds.): COST 2102 Int. Training School 2010, Lecture Notes In Artificial Intelligence Springer-Verlag, Berlin, Heidelberg 6456, 387-402.
* Laouris, Y., Makris, P., Papadopoulos, T. (2011). Validation of MAPS in 16 schools: Computer-based battery of 8 mental attributes tests (in preparation; draft available).
* Laouris, Y., Makris, P., Papadopoulos, T. (2011). Validation of MAPS in 16 schools: Computer-based battery of 8 mental attributes tests (in preparation; draft available).


==MAPS presented in International Conferences==
==MAPS presented in International Conferences==
Line 117: Line 170:
* Loredana Mihalca, L., Laouris, Y., Miclea, M. (2007). Children’s school readiness: predicting school competence in first grade. Proc. EARLI Conference, Budapest, Hungary, 26-31 Aug. 2007.  
* Loredana Mihalca, L., Laouris, Y., Miclea, M. (2007). Children’s school readiness: predicting school competence in first grade. Proc. EARLI Conference, Budapest, Hungary, 26-31 Aug. 2007.  
* Laouris, Y., & Makris, P. (2007). An interactive ICT demonstration of MAPS (Mental Attributes Profiling System). Proc. EARLI Conference, Budapest, Hungary, 26-31 Aug. 2007.
* Laouris, Y., & Makris, P. (2007). An interactive ICT demonstration of MAPS (Mental Attributes Profiling System). Proc. EARLI Conference, Budapest, Hungary, 26-31 Aug. 2007.
* Taraszow, T., Wahl, J., Laouris, Y., Scheiter, K., & Gerjets, P. (2007). Using Dynamic Visualization with Written Explanation to By-Pass Information Processing Deficits of Children with Reading Difficulties. Proc. EARLI Conference, Budapest, Hungary, 26-31 Aug. 2007.
* Taraszow, T., Wahl, J., Laouris, Y., Scheiter, K., & Gerjets, P. (2007, August). Using Dynamic Visualization with Written Explanation to By-Pass Information Processing Deficits of Children with Reading Difficulties. Proc. EARLI Conference, Budapest, Hungary, 26-31 Aug. 2007.
* Taraszow, T., & Laouris, Y. The development of the categorization ability in children age 8 to 12. Proc. EARLI Conference, Budapest, Hungary, 26-31 Aug. 2007.
* Taraszow, T., & Laouris, Y. (2007, August). The development of the categorization ability in children age 8 to 12. Proc. EARLI Conference, Budapest, Hungary, 26-31 Aug. 2007.
* Aristodemou, E., Taraszow, T., Laouris, Y., Papadopoulos, T., & Makris, P. (2008). Prediction of Reading Performance Using the MAPS (Mental Attributes Profiling System) Multimodal Interactive ICT Application. Proceedings of 7th European Conference on e-learning, Agia Napa, Cyprus.
* Aristodemou, E., Taraszow, T., Laouris, Y., Papadopoulos, T., & Makris, P. (2008, November). Prediction of Reading Performance Using the MAPS (Mental Attributes Profiling System) Multimodal Interactive ICT Application. Paper presented at the 7th European Conference on e-Learning (ECEL), Ayia Napa, Cyprus.
* Taraszow, T., Aristodemou, E., & Laouris, Y. (2008, May). E-Learning adapted to the learning profiles of young learners. Paper presented at the Conference Child and Youth Research in the 21st Century: A Critical Appraisal, Nicosia, Cyprus.
* Taraszow, T., Laouris,Y., Wahl, J., Scheiter, K., & Gerjets, P. (2008, November). A Matter of Modality: Do Multimedia Learning Environments really Support Students with Reading Difficulties? Paper presented at the 7th European Conference on e-Learning (ECEL), Ayia Napa, Cyprus.
 
==MAPS; part of many conferences==
MAPS has motivated the organization of a number of local and international conferences:
{{#categorytree:Dyslexia Conferences|mode="all" |hideroot=on|mode=pages|style=bold}}




Line 131: Line 190:
# [[Chris Singletton]], University of Hull
# [[Chris Singletton]], University of Hull
# [[Costas Apostolides]], [[Cyprus Dyslexia Association]]
# [[Costas Apostolides]], [[Cyprus Dyslexia Association]]
# [[Christiane El Haddad]], Psycholosgist
# [[Christiane El Haddad]], Psychologist
# Anna Basileiadou-Elissaiou]], [[Cyprus Dyslexia Association]]
# [[Joulietta Kalli-Laouri]], Child- and Adolescent Psychiatry Specialist
# [[Anna Basileiadou-Elissaiou]], [[Cyprus Dyslexia Association]]
# [[Petroula Petrou]]
# [[Petroula Petrou]]
# [[Mira Ioannou]]
# [[Mira Ioannou]]
Line 144: Line 204:
# [[Maria Mavromichali]]
# [[Maria Mavromichali]]


==Validation in Public Schools==
The following Elementary School teachers have collaborated with the [[Cyprus Neuroscience and Technology Institute]], the Ministry of Education and the University of Cyprus for the testing of MAPS in 30 public elementary schools between February and June 2003. The results of this validation have been reported at the 54th Annual International Dyslexia Association Annual Meeting.
* Angeliki Charalambous
* Margarita
* Elena Kokkinofta
* Tatiana Koufalidou
* Stelios Hadjiyiannis
* Panayiota Yianni
* Thoi Asvestas
* Maria Pieri
* Tzeni Digigoropoulou
* Lambros Stavrou
* Natalia Tsangarides
* Triada Kyriakou
* Dora Costa Spanashi
* Filanthi Christou Maria Pieri
* Loukia Cosma – Achilleos
* Christina Theophanous
* Maria Petinari
* Tarsi Hadjitheodoulou
* Christina Pieridou
* Elena Georgalidou
* Stelios Stylianou
* Anastasia Georgiadou
* Marina Constantinou
* Antonis Andreou
* Andri Michael
* [[Stella Ioannides]]




[[Category:Future Worlds Center Projects]]
[[Category:Future Worlds Center Projects]]
[[Category:RPF Projects]]
[[Category:RIF Projects]]
 
[[Category:Software Products]]
[[Category:Mobile Apps]]
[[Category:Ekkotek Products]]
Ninja, Ninla, Bots, Bureaucrats, recentchangescleanup, Administrators
4,970

edits

Navigation menu