Mental Attributes Profiling System: Difference between revisions

From Future Worlds Center Wiki
Jump to navigationJump to search
no edit summary
No edit summary
No edit summary
 
(31 intermediate revisions by 3 users not shown)
Line 2: Line 2:
       |acronym=MAPS             
       |acronym=MAPS             
       |logo=MAPS Blue Logo.jpg
       |logo=MAPS Blue Logo.jpg
       |project_title= Mental Attributes Profiling System 
       |project_title=[[Development of methodology and software for the diagnosis and treatement of learning difficulties with emphasis in dyslexia]] <br>[[Educational Hypermedia (Hibernation)]]<br>[[Dynamic Visualizations in Instruction]]<br>[[Unified e-Hoop approach to learning differences]]
       |contract_number=
       |contract_number=
       |funding_period=
       |funding_period=1999-2016
       |coordinator=
       |coordinator=[[Cyprus Neuroscience and Technology Institute]]
       |funding_agency=[[Research Promotion Foundation]]
       |funding_agency=[[Research Promotion Foundation]]<br>Leibniz Association, Germany<br>European Commission
       |total_cost=
       |total_cost=ca. 700K
       |partners=[[Cyprus Dyslexia Association]]
       |partners=[[Cyprus Dyslexia Association]]
       |website=
       |website=http://mentalmaps.org
}}     
}}     


   
   
The '''Mental Attributes Profiling System''' (MAPS) is a language independent screening test, i.e., relies on cognitive rather than language-based measures, capable of predicting children at risk (possible dyslexics) and equipping teachers with a profile of their mental abilities so as to design personalized remediation programs.
The '''Mental Attributes Profiling System''' (MAPS) is a language independent screening test, i.e., relies on cognitive rather than language-based measures, capable of predicting children at risk (possible dyslexics) and equipping teachers with a profile of their mental abilities so as to design personalized remediation programs.
MAPS has been developed as a research project within the [[Cyprus Neuroscience and Technology Institute]].
In 2017, the IP was transferred to a spin off company, [[GNOUS Labs Ltd.]].
        
        
__TOC__
__TOC__
Line 21: Line 24:


The following eight tests have been used for the multimodal evaluation reported here:</ul>
The following eight tests have been used for the multimodal evaluation reported here:</ul>


==Categorization==
==Categorization==
Line 33: Line 37:
# An animal (or fish) to be placed in one of three possible environments: sea, sky, or open fields;
# An animal (or fish) to be placed in one of three possible environments: sea, sky, or open fields;
# Objects usually found in a home environment, to be “placed” in one of three rooms of a house (office, kitchen, or bathroom), i.e., as in the example snapshot shown in the illustration. The software measured the time to response and whether the move was correct or false. For this study we used only the average response time (i.e., the cumulative response time of each test divided by 5: (ctt1+ctt2+...ctt5)/5 in Table 1), as it has been shown previously to have a higher correlation than the Correct/False values.
# Objects usually found in a home environment, to be “placed” in one of three rooms of a house (office, kitchen, or bathroom), i.e., as in the example snapshot shown in the illustration. The software measured the time to response and whether the move was correct or false. For this study we used only the average response time (i.e., the cumulative response time of each test divided by 5: (ctt1+ctt2+...ctt5)/5 in Table 1), as it has been shown previously to have a higher correlation than the Correct/False values.


==Orientation: “Left-Right Game” Lateral awareness Test==
==Orientation: “Left-Right Game” Lateral awareness Test==
[[File:Laterization.jpg|left|thumb|upright=1|alt=Logo.|Ability to orientate: Left/Right]]
[[File:Laterization.jpg|left|thumb|upright=1|alt=Logo.|Ability to orientate: Left/Right]]
This test provides two types of measures. One, it evaluates the child’s ability to make left-right discriminations on his own body. During the first part, the test shows a child “sitting” in the same orientation as the subject (i.e., the subject sees on the screen the back of the child) in front of two objects, one on his/her upper left and the other on his/her upper right visual field. The computer asks the subject to “grab” the left or right object using his/her left or right hand. What is measured is (1) the time taken to select an arm (by clicking at the shoulder), and (2) whether the selected arm was correct or wrong. The same procedure is repeated during the second part of the test, in which the orientation of the child on the screen is reversed, i.e., the child on the screen is facing the subject. The second type of measures are derived from Piaget’s (19) tests to evaluate awareness of right-left relations outside our own body, i.e., in the environment. To evaluate this ability, the test measures (1) the time taken by the subject to decide which object to “grab” and (2) the number of correct/wrong decisions.  
This test provides two types of measures. One, it evaluates the child’s ability to make left-right discriminations on his own body. During the first part, the test shows a child “sitting” in the same orientation as the subject (i.e., the subject sees on the screen the back of the child) in front of two objects, one on his/her upper left and the other on his/her upper right visual field. The computer asks the subject to “grab” the left or right object using his/her left or right hand. What is measured is (1) the time taken to select an arm (by clicking at the shoulder), and (2) whether the selected arm was correct or wrong. The same procedure is repeated during the second part of the test, in which the orientation of the child on the screen is reversed, i.e., the child on the screen is facing the subject. The second type of measures are derived from Piaget’s (19) tests to evaluate awareness of right-left relations outside our own body, i.e., in the environment. To evaluate this ability, the test measures (1) the time taken by the subject to decide which object to “grab” and (2) the number of correct/wrong decisions.  


==Navigation==
==Navigation==
Line 47: Line 55:


<br>
<br>




Line 60: Line 69:


<br>
<br>


==Visual memory==
==Visual memory==
Line 76: Line 83:


<br>
<br>




Line 88: Line 94:
* AMTc: Total number of correct responses
* AMTc: Total number of correct responses
* AMTcw: weighted number of correct digits (i.e., all correctly types digits even when the sequence was wrong) were used as measures of auditory digit span.
* AMTcw: weighted number of correct digits (i.e., all correctly types digits even when the sequence was wrong) were used as measures of auditory digit span.


==Auditory discrimination test==
==Auditory discrimination test==
Line 111: Line 118:
==Visual discrimination==
==Visual discrimination==
Modeled using the popular “Find the differences” game, this test presents three pictures very similar to one another and one that must be matched. The exercise is repeated four times with four different pictures. The total number of correct responses is used as an index of visual discrimination.  
Modeled using the popular “Find the differences” game, this test presents three pictures very similar to one another and one that must be matched. The exercise is repeated four times with four different pictures. The total number of correct responses is used as an index of visual discrimination.  
==Grants and Funding of MAPS==
{| class="wikitable"
! Grants
! Year
! Amount (€)
! Source
|-
| DYSLEXIA
| 1999-2001
| style="text-align:right;" | 44,803
| Research Promotion Foundation
|-
| [[Dynamic Visualizations in Instruction|Educational Hypermedia]]
| 2007
| style="text-align:right;" | 44,460
| Leibniz Association, Germany
|-
| Dynamic Visualizations in Instruction
| 2007-2009
| style="text-align:right;" | 51,300
| Leibniz Association, Germany
|-
| [[E-Hoop|Unified e-Hoop approach to learning differences]]
| 2012-2016
| style="text-align:right;" | 562,599<SUP>1</SUP>
| European Commission
|}
Note 1: Only 25% of this amount was exclusively used for R&D.


==Publications==
==Publications==
Line 117: Line 157:
* Haddad, C. and Laouris, Y. (2011). The ability of children with mild learning disabilities to encode emotions through facial expressions. In Towards Autonomous, adaptive and context-aware multimodal interfaces: Theoretical and Practical Issues. A. Esposito, Esposito, A.M., Martone, R., Müller, V.C., Scarpetta, G  (Eds.): COST 2102 Int. Training School 2010, Lecture Notes In Artificial Intelligence Springer-Verlag, Berlin, Heidelberg 6456, 387-402.
* Haddad, C. and Laouris, Y. (2011). The ability of children with mild learning disabilities to encode emotions through facial expressions. In Towards Autonomous, adaptive and context-aware multimodal interfaces: Theoretical and Practical Issues. A. Esposito, Esposito, A.M., Martone, R., Müller, V.C., Scarpetta, G  (Eds.): COST 2102 Int. Training School 2010, Lecture Notes In Artificial Intelligence Springer-Verlag, Berlin, Heidelberg 6456, 387-402.
* Laouris, Y., Makris, P., Papadopoulos, T. (2011). Validation of MAPS in 16 schools: Computer-based battery of 8 mental attributes tests (in preparation; draft available).
* Laouris, Y., Makris, P., Papadopoulos, T. (2011). Validation of MAPS in 16 schools: Computer-based battery of 8 mental attributes tests (in preparation; draft available).


==MAPS presented in International Conferences==
==MAPS presented in International Conferences==
Line 129: Line 170:
* Loredana Mihalca, L., Laouris, Y., Miclea, M. (2007). Children’s school readiness: predicting school competence in first grade. Proc. EARLI Conference, Budapest, Hungary, 26-31 Aug. 2007.  
* Loredana Mihalca, L., Laouris, Y., Miclea, M. (2007). Children’s school readiness: predicting school competence in first grade. Proc. EARLI Conference, Budapest, Hungary, 26-31 Aug. 2007.  
* Laouris, Y., & Makris, P. (2007). An interactive ICT demonstration of MAPS (Mental Attributes Profiling System). Proc. EARLI Conference, Budapest, Hungary, 26-31 Aug. 2007.
* Laouris, Y., & Makris, P. (2007). An interactive ICT demonstration of MAPS (Mental Attributes Profiling System). Proc. EARLI Conference, Budapest, Hungary, 26-31 Aug. 2007.
* Taraszow, T., Wahl, J., Laouris, Y., Scheiter, K., & Gerjets, P. (2007). Using Dynamic Visualization with Written Explanation to By-Pass Information Processing Deficits of Children with Reading Difficulties. Proc. EARLI Conference, Budapest, Hungary, 26-31 Aug. 2007.
* Taraszow, T., Wahl, J., Laouris, Y., Scheiter, K., & Gerjets, P. (2007, August). Using Dynamic Visualization with Written Explanation to By-Pass Information Processing Deficits of Children with Reading Difficulties. Proc. EARLI Conference, Budapest, Hungary, 26-31 Aug. 2007.
* Taraszow, T., & Laouris, Y. The development of the categorization ability in children age 8 to 12. Proc. EARLI Conference, Budapest, Hungary, 26-31 Aug. 2007.
* Taraszow, T., & Laouris, Y. (2007, August). The development of the categorization ability in children age 8 to 12. Proc. EARLI Conference, Budapest, Hungary, 26-31 Aug. 2007.
* Aristodemou, E., Taraszow, T., Laouris, Y., Papadopoulos, T., & Makris, P. (2008). Prediction of Reading Performance Using the MAPS (Mental Attributes Profiling System) Multimodal Interactive ICT Application. Proceedings of 7th European Conference on e-learning, Agia Napa, Cyprus.
* Aristodemou, E., Taraszow, T., Laouris, Y., Papadopoulos, T., & Makris, P. (2008, November). Prediction of Reading Performance Using the MAPS (Mental Attributes Profiling System) Multimodal Interactive ICT Application. Paper presented at the 7th European Conference on e-Learning (ECEL), Ayia Napa, Cyprus.
* Taraszow, T., Aristodemou, E., & Laouris, Y. (2008, May). E-Learning adapted to the learning profiles of young learners. Paper presented at the Conference Child and Youth Research in the 21st Century: A Critical Appraisal, Nicosia, Cyprus.
* Taraszow, T., Laouris,Y., Wahl, J., Scheiter, K., & Gerjets, P. (2008, November). A Matter of Modality: Do Multimedia Learning Environments really Support Students with Reading Difficulties? Paper presented at the 7th European Conference on e-Learning (ECEL), Ayia Napa, Cyprus.


==MAPS; part of many conferences==
==MAPS; part of many conferences==
MAPS has motivated the organization of a number of local and international conferences:
MAPS has motivated the organization of a number of local and international conferences:
{{#categorytree:Dyslexia Conferences|mode="all" |hideroot=on|mode=pages|style=bold}}
{{#categorytree:Dyslexia Conferences|mode="all" |hideroot=on|mode=pages|style=bold}}


==MAPS Associates==
==MAPS Associates==
Line 146: Line 190:
# [[Chris Singletton]], University of Hull
# [[Chris Singletton]], University of Hull
# [[Costas Apostolides]], [[Cyprus Dyslexia Association]]
# [[Costas Apostolides]], [[Cyprus Dyslexia Association]]
# [[Christiane El Haddad]], Psycholosgist
# [[Christiane El Haddad]], Psychologist
# Anna Basileiadou-Elissaiou]], [[Cyprus Dyslexia Association]]
# [[Joulietta Kalli-Laouri]], Child- and Adolescent Psychiatry Specialist
# [[Anna Basileiadou-Elissaiou]], [[Cyprus Dyslexia Association]]
# [[Petroula Petrou]]
# [[Petroula Petrou]]
# [[Mira Ioannou]]
# [[Mira Ioannou]]
Line 169: Line 214:
* Panayiota Yianni  
* Panayiota Yianni  
* Thoi Asvestas  
* Thoi Asvestas  
* Maria Pieri
* Tzeni Digigoropoulou  
* Tzeni Digigoropoulou  
* Lambros Stavrou
* Lambros Stavrou
Line 186: Line 232:
* Antonis Andreou
* Antonis Andreou
* Andri Michael
* Andri Michael
* Stella Ioannides
* [[Stella Ioannides]]




[[Category:Future Worlds Center Projects]]
[[Category:Future Worlds Center Projects]]
[[Category:RPF Projects]]
[[Category:RIF Projects]]
 
[[Category:Software Products]]
[[Category:Mobile Apps]]
[[Category:Ekkotek Products]]
Ninja, Ninla, Bots, Bureaucrats, recentchangescleanup, Administrators
4,970

edits

Navigation menu