1 / 60

NIH WORKSHOP: INFORMATICS FOR DATA AND RESOURCE DISCOVERY IN ADDICTION RESEARCH July 8, 2010

NIH WORKSHOP: INFORMATICS FOR DATA AND RESOURCE DISCOVERY IN ADDICTION RESEARCH July 8, 2010. Case Study 5 (NEMO) : Informatics tools to support theoretical and practical integration of human neuroscience data Gwen Frishkoff, Ph.D. Psychology & Neuroscience, Georgia State University

keith-hall
Download Presentation

NIH WORKSHOP: INFORMATICS FOR DATA AND RESOURCE DISCOVERY IN ADDICTION RESEARCH July 8, 2010

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. NIH WORKSHOP: INFORMATICS FOR DATA AND RESOURCE DISCOVERY IN ADDICTION RESEARCHJuly 8, 2010 Case Study 5 (NEMO): Informatics tools to support theoretical and practical integration of human neuroscience data Gwen Frishkoff, Ph.D. Psychology & Neuroscience, Georgia State University NeuroInformatics Center, University of Oregon http://nemo.nic.uoregon.edu

  2. Neuro–Informatics: Crossing the language divide

  3. What the computer scientist says… Should we write out the data to XML or RDF triples? And do you plan to use ontologyrules to do complex reasoning or just use SQL to query the data?

  4. What the neuroscientist hears… Blah blah blah blah blah…data… blah blah blah? And blah blah blah…the data?

  5. GOALS FOR THIS TUTORIAL • What is an ontology & what’s it for? • Why bother? (Case Study: Classification of EEG/ERP data) • What are some “best practices” in ontology design & implementation? • What is RDF & what’s it for? • How does RDF represent information? • How is it used to link data to ontologies? • How can ontology-based annotation be used to support classification of data?

  6. Case Study 5 (NEMO): Neural ElectroMagnetic Ontologies • The problem (pattern classification) • The methods & tools • ontologies • RDF database • Proof of concept (a worked example)

  7. Case Study 5 (NEMO): Neural ElectroMagnetic Ontologies • The challenge (pattern classification) • The methods & tools • ontologies • RDF database • Proof of concept (a worked example)

  8. 2-min Primer on EEG/ERP Methods EEGs (“brainwaves” or flunctuations in brain electrical potentials) are recorded by placing two or more electrodes on the scalp surface. ~5,000 ms 256-channel Geodesic Sensor Net

  9. Event-related potentials (ERP) • ERPs (“event-related potentials”) are the result of averaging across multiple segments of EEG, time-locking to an event of interest.

  10. ERP Patterns (“Components”) 120 ms ERP Patterns are characterized by 3 dimensions: TIME — peak latency, duration (WHEN in time) SPACE— scalp “topography” (WHERE on scalp) FUNCTION — sensitivity to experiment factors Donchin & Duncan-Johnson, 1977

  11. Brain Electrophysiology (EEG/ERP): The promise (Biomarkers of addiction?) • Tried and true method for noninvasive brain functional mapping • Millisecond temporal resolution • Direct measure neuronal activity • Portable and inexpensive • Recent innovations give new windows into rich, multi-dimensionalpatterns • More spatial info (high-density EEG) • More temporal & spectral info (JTF, etc.) • Multimodal integration & joint recordings of EEG and fMRI • Specificity of different patterns  beyond “reduction in P300” amplitude… 1 sec

  12. Brain Electrophysiology (EEG/ERP): The challenge • An embarrassment of riches • A wealth of data • A plethora of methods • A lack of integration • How to compare patterns across studies, labs? • How to do valid meta-analyses in ERP research? • A need for robust pattern classification • Bottom-up (data-driven) methods • Top-down (knowledge-driven) methods

  13. An embarrassment of riches

  14. A lack of standardization Hypothetical Database Query: Show me all the N400 patterns in data set X. 450 ms Peak latency 410 ms Will the “real” N400 please step forward? 330 ms 410 ms

  15. A Need for Integration Parietal N400 Putative “N400”-labeled patterns ≠ Frontal N400 ≠ Parietal P600

  16. Neural ElectroMagnetic Ontologies(NEMO) • The driving goal is to develop methods and tools to support cross-lab, cross-experiment integration of EEG and MEG data • We bring a set of methods & tools to bear to address this: • A set of formal (OWL) ontologies for representation of EEG/MEG and ERP/ERF data • A suite of tools for ontology-based annotation and analysis of EEG and ERP data • AnRDF database that stores annotated data from our NEMO ERP consortium and supports ERP pattern classification via SPARQL queries

  17. Case Study 5 (NEMO): Neural ElectroMagnetic Ontologies • The challenge (EEG pattern classification) • The methods & tools • ontologies • RDF database • Proof of concept (a worked example)

  18. What’s an ontology & what’s it for? “Highly semantically structured” What does this mean & what does it buy us?

  19. *NOTE: We can record pattern definitions from literature in ontology without committing to the truth of these records now and forever Science evolves… So do ontologies!! Maryann: “Avoid ontology wars…” Ontologies for high-level, explicit representation of domain knowledge  theoretical integration*

  20. Ontology design principles(based on OBO Foundry recommendations) • Factor the domain to generate modular (“orthogonal”) ontologies that can be reused, integrated for other projects • Reuse existing ontologies (esp. foundational concepts) to define basic (low-level) concepts • Validate definitions of high-level concepts in bottom-up (data-driven) as well as top-down (knowledge-driven) methods • Collaborate with a community of experts in collaborative design, testing of ontology-based tools for data representation and analysis

  21. Factoring the ERP domain TIME SPACE 1 sec FUNCTION  Modulation of pattern features (time, space, amplitude) under different experiment conditions

  22. Overview: NEMO Ontologies • NEMO core modules: • NEMO_spatial • NEMO_temporal • NEMO_functional • NEMO_ERP • NEMO_data • NEMO backend: • NEMO_relations • NEMO_imports • NEMO_deprecated • NEMO_annotation_properties

  23. ERP spatial subdomain TIME SPACE 1 sec FUNCTION  Modulation of ERP pattern features under different experiment conditions

  24. International 10-10 EEG Electrode Locations ITT electrode location Fz (medial frontal)

  25. Scalp surface “regions of interest” LEFT MEDIAL RIGHT FRONTAL TEMPORAL PARIETAL OCCIPITAL

  26. Reuse in dev’t of NEMO Spatial BFO (Basic Formal Ontology)  “UPPER ONTOLOGY” FMA (Foundational Model of Anatomy)  “MIDLEVEL ONTOLOGY”

  27. ERP temporal subdomain TIME SPACE 1 sec FUNCTION  Modulation of ERP pattern features under different experiment conditions

  28. Early (“exogenous”) vs. Late(“endogenous”) ERP patterns EARLY ~0-150 ms after event (e.g., stimulus onset) MID-LATENCY ~151-500 after event (e.g., stimulus onset) LATE 501 ms or more after event (e.g., stimulus onset)

  29. NEMO Temporal Ontology

  30. Collaboration in dev’t of NEMO ERP TIME SPACE 1 sec FUNCTION  Modulation of ERP pattern features under different experiment conditions

  31. NEMO Functional Ontology Jessica Turner BIRN (now part of Neurolex) http://brainmap.org/scribe/index.html CogPO Angela Laird BrainMap

  32. “Cognitive ontologies”  Formalization of experiment metadata

  33. CARMEN Project: Development of MINI Frank Gibson & colleagues

  34. Reconsistituting the ERP domain… TIME SPACE 1 sec FUNCTION  Modulation of ERP pattern features under different experiment conditions

  35. Validation through application of NEMO ontologies in modeling ERP data Frishkoff, Frank, et al., 2007

  36. Case Study 5 (NEMO): Neural ElectroMagnetic Ontologies • The challenge (EEG pattern classification) • The methods & tools • ontologies • RDF database • Proof of concept (a worked example)

  37. Ontologies for high-level, explicit representation of domain knowledge  theoretical integration • RDF to support principled mark-up of data for meta-analysis • practical integration

  38. NEMO International Language & Literacy Consortium John Connolly McMaster University Chuck Perfetti University of Pittsburgh Dennis Molfese University of Louisville Kerry Kiborn University of Glasgow Tim Curran University of Colorado Formed in 2007

  39. What is RDF and what is it for? RDF graph (data model)

  40. Annoting EEG/ERP data Pattern Labels Functional attributes Temporal attributes Spatial attributes = + + Concepts coded in OWL  NEMO ontology Data coded in RDF  NEMO database HOW? Robert M. Frank

  41. Annotating Data in RDF • Data Annotation • The process of marking up or “tagging” data with meaningful symbols; tags may come from ontology linked to a URI • URI (Uniform Resource Identifier) • A compact sequence of characters that identifies an abstract or physical resource (typically located on the Web) • RDF (Resource Description Framework) • RDF is a directed, labeled graph (data model) for representing information (typically on the Web) *See Glossary (http://www.seiservices.com/nida/1014080/ReadingRoom.aspx)

  42. Recall: The goal is to formulate pattern definitions, use them to classify data, and ultimately to revise them based on meta-analysis results Observed Pattern = “N400” iff • Event type is onset of meaningful stimulus (e.g., word) AND • Peak latency is between 300 and 500 ms AND • Scalp region of interest (ROI) is centroparietal AND • Polarity over ROI is negative(>0)

  43. The rule (just the temporal criterion)as it appears in Protégé Protégé rendering OWL/RDF rendering

  44. Typical tabular representation of summary ERP data ERP observation (pattern extracted from “raw” ERP data) Peak latency measurement

  45. The “RDF Triple” • In RDF form: <001> <type> <NEMO_0000093> • Subject – Predicate – Object • In natural language: • The data represented in row A is an instance of (“is a”) some ERP pattern. • That is, measurements (cells) are “about” ERP patterns (rows). • In graph form:

  46. RDF Triple #2 • In RDF form: <002> <type> <NEMO_0745000> • Subject – Predicate – Object • In natural language = • The data represented in cell Z (row A, column 1) is an instance of (“is a”) a peak latency temporal measurement (i.e., the time at which the pattern is of maximal amplitude)

  47. RDF Triple #3 • This graph represents an assertion, expressed in RDF = • <001> <is_peak_latency_measurement_of> <002> • The data represented in cell Z is a temporal property of the ERP pattern represented in row A

  48. Recall: Pattern definition is encoded in the ontology (not in RDF data rep!)

  49. This is the inference that we want to make

  50. Pattern classification is the goal

More Related