Let's talk about dark data — what it means and how to navigate it. Graphic byMiguel Tovar/University of Houston

Is it necessary to share ALL your data? Is transparency a good thing or does it make researchers “vulnerable,” as author Nathan Schneider suggests in the Chronicle of Higher Education article, “Why Researchers Shouldn’t Share All Their Data.”

Dark Data Defined

Dark data is defined as the universe of information an organization collects, processes and stores – oftentimes for compliance reasons. Dark data never makes it to the official publication part of the project. According to the Gartner Glossary, “storing and securing data typically incurs more expense (and sometimes greater risk) than value.”

This topic is reminiscent of the file drawer effect, a phenomenon which reflects the influence of the results of a study on whether or not the study is published. Negative results can be just as important as hypotheses that are proven.

Publication bias and the need to only publish positive research that supports the PI’s hypothesis, it can be argued, is not good science. According to an article in the Indian Journal of Anaesthesia, authors Priscilla Joys Nagarajan, et al., wrote: “It is speculated that every significant result in the published world has 19 non-significant counterparts in file drawers.” That’s one definition of dark data.

Total Transparency

But what to do with all your excess information that did not make it to publication, most likely because of various constraints? Should everything, meaning every little tidbit, be readily available to the research community?

Schneider doesn’t think it should be. In his article, he writes that he hides some findings in a paper notebook or behind a password, and he keeps interviews and transcripts offline altogether to protect his sources.

Open-source

Open-source software communities tend to regard total transparency as inherently good. What are the advantages of total transparency? You may make connections between projects that you wouldn’t have otherwise. You can easily reproduce a peer’s experiment. You can even become more meticulous in your note-taking and experimental methods since you know it’s not private information. Similarly, journalists will recognize this thought pattern as the recent, popular call to engage in “open journalism.” Essentially, an author’s entire writing and editing process can be recorded, step by step.

TMI

This trend has led researchers to open-source programs like Jupyter and GitHub. Open-source programs detail every change that occurs along a project’s timeline. Is unorganized, excessive amounts of unpublishable data really what transparency means? Or does it confuse those looking for meaningful research that is meticulously curated?

The Big Idea

And what about the “vulnerability” claim? Sharing every edit and every new direction taken opens a scientist up to scoffers and harassment, even. Dark data in industry even involves publishing salaries, which can feel unfair to underrepresented, marginalized populations.

In Model View Culture, Ellen Marie Dash wrote: “Let’s give safety and consent the absolute highest priority, with openness and transparency prioritized explicitly below those. This means digging deep, properly articulating in detail what problems you are trying to solve with openness and transparency, and handling them individually or in smaller groups.”

------

This article originally appeared on the University of Houston's The Big Idea. Sarah Hill, the author of this piece, is the communications manager for the UH Division of Research.

Ad Placement 300x100
Ad Placement 300x600

CultureMap Emails are Awesome

Houston organizations launch collaborative center to boost cancer outcomes

new to HOU

Rice University's new Synthesis X Center officially launched last month to bring together experts in cancer care and chemistry.

The center was born out of what started about seven years ago as informal meetings between Rice chemist Han Xiao's research group and others from the Baylor College of Medicine’s Dan L Duncan Comprehensive Cancer Center at the Baylor College of Medicine. The level of collaboration between the two teams has grown significantly over the years, and monthly meetings now draw about 100 participants from across disciplines, fields and Houston-based organizations, according to a statement from Rice.

Researchers at the new SynthX Center will aim to turn fundamental research into clinical applications and make precision adjustments to drug properties and molecules. It will focus on improving cancer outcomes by looking at an array of factors, including prevention and detection, immunotherapies, the use of artificial intelligence to speed drug discovery and development, and several other topics.

"At Rice, we are strong on the fundamental side of research in organic chemistry, chemical biology, bioengineering and nanomaterials,” Xiao says in the statement. “Starting at the laboratory bench, we can synthesize therapeutic molecules and proteins with atom-level precision, offering immense potential for real-world applications at the bedside ... But the clinicians and fundamental researchers don’t have a lot of time to talk and to exchange ideas, so SynthX wants to serve as the bridge and help make these connections.”

SynthX plans to issue its first merit-based seed grants to teams with representatives from Baylor and Rice this month.

With this recognition from Rice, the teams from Xiao's lab and the TMC will also be able to expand and formalize their programs. They will build upon annual retreats, in which investigators can share unpublished findings, and also plan to host a national conference, the first slated for this fall titled "Synthetic Innovations Towards a Cure for Cancer.”

“I am confident that the SynthX Center will be a great resource for both students and faculty who seek to translate discoveries from fundamental chemical research into medical applications that improve people’s lives,” Thomas Killian, dean of the Wiess School of Natural Sciences, says in the release.

Rice announced that it had invested in four other research centers along with SynthX last month. The other centers include the Center for Coastal Futures and Adaptive Resilience, the Center for Environmental Studies, the Center for Latin American and Latinx Studies and the Rice Center for Nanoscale Imaging Sciences.

Earlier this year, Rice also announced its first-ever recipients of its One Small Step Grant program, funded by its Office of Innovation. The program will provide funding to faculty working on "promising projects with commercial potential," according to the website.

Houston physicist scores $15.5M grant for high-energy nuclear physics research

FUTURE OF PHYSICS

A team of Rice University physicists has been awarded a prestigious grant from the Department of Energy's Office of Nuclear Physics for their work in high-energy nuclear physics and research into a new state of matter.

The five-year $15.5 million grant will go towards Rice physics and astronomy professor Wei Li's discoveries focused on the Compact Muon Solenoid (CMS), a large, general-purpose particle physics detector built on the Large Hadron Collider (LHC) at CERN, a European organization for nuclear research in France and Switzerland. The work is "poised to revolutionize our understanding of fundamental physics," according to a statement from Rice.

Li's team will work to develop an ultra-fast silicon timing detector, known as the endcap timing layer (ETL), that will provide upgrades to the CMS detector. The ETl is expected to have a time resolution of 30 picoseconds per particle, which will allow for more precise time-of-flight particle identification.

The Rice team is collaborating with others from MIT, Oak Ridge National Lab, the University of Illinois Chicago and University of Kansas. Photo via Rice.edu

This will also help boost the performance of the High-Luminosity Large Hadron Collider (HL-LHC), which is scheduled to launch at CERN in 2029, allowing it to operate at about 10 times the luminosity than originally planned. The ETL also has applications for other colliders apart from the LHC, including the DOE’s electron-ion collider at the Brookhaven National Laboratory in Long Island, New York.

“The ETL will enable breakthrough science in the area of heavy ion collisions, allowing us to delve into the properties of a remarkable new state of matter called the quark-gluon plasma,” Li explained in a statement. “This, in turn, offers invaluable insights into the strong nuclear force that binds particles at the core of matter.”

The ETL is also expected to aid in other areas of physics, including the search for the Higgs particle and understanding the makeup of dark matter.

Li is joined on this work by co-principal investigator Frank Geurts and researchers Nicole Lewis and Mike Matveev from Rice. The team is collaborating with others from MIT, Oak Ridge National Lab, the University of Illinois Chicago and University of Kansas.

Last year, fellow Rice physicist Qimiao Si, a theoretical quantum physicist, earned the prestigious Vannevar Bush Faculty Fellowship grant. The five-year fellowship, with up to $3 million in funding, will go towards his work to establish an unconventional approach to create and control topological states of matter, which plays an important role in materials research and quantum computing.

Meanwhile, the DOE recently tapped three Houston universities to compete in its annual startup competition focused on "high-potential energy technologies,” including one team from Rice.

------

This article originally ran on EnergyCapital.