The impact of computer science education in primary schools: Evidence from a randomized controlled trial in Iraq

PLOS One

View Article

With the growing digitization of society, there is a need to enhance computational thinking as an indispensable skill for modern daily life. Consequently, computer science education for children at early ages has become increasingly important. This study conducts a randomized controlled trial to examine the impact of the interventions using educational robotics as well as computer-aided mathematics drills (via a “math app”) on students’ performance in primary schools in Basra, Iraq. We provide several new empirical findings. First, the short-run impact of robotics-based learning on computational thinking is positive and statistically significant for girls, particularly poor performing girls, but not for boys. Second, the impact on computational thinking is augmented by introducing a math app, further improving computational thinking. Together, these two interventions also enhance general intelligence. Third, the positive impact was still evident more than three months after the interventions for girls who received both computer science and math education, suggesting their complementarity. Our results show that computer science education using educational robots in primary schools is effective in enhancing computational thinking and relevant skills.

A field guide to cultivating computational biology

PLOS Biology

View Article

Evolving in sync with the computation revolution over the past 30 years, computational biology has emerged as a mature scientific field. While the field has made major contributions toward improving scientific knowledge and human health, individual computational biology practitioners at various institutions often languish in career development. As optimistic biologists passionate about the future of our field, we propose solutions for both eager and reluctant individual scientists, institutions, publishers, funding agencies, and educators to fully embrace computational biology. We believe that in order to pave the way for the next generation of discoveries, we need to improve recognition for computational biologists and better align pathways of career success with pathways of scientific progress. With 10 outlined steps, we call on all adjacent fields to move away from the traditional individual, single-discipline investigator research model and embrace multidisciplinary, data-driven, team science. Do you want to attract computational biologists to your project or to your department? Despite the major contributions of computational biology, those attempting to bridge the interdisciplinary gap often languish in career advancement, publication, and grant review. Here, sixteen computational biologists around the globe present "A field guide to cultivating computational biology," focusing on solutions.

Factors shaping the gender wage gap among college-educated computer science workers

PLOS ONE

View Article

Encouraging women to pursue STEM employment is frequently touted as a means of reducing the gender wage gap. We examine whether the attributes of computer science workers–who account for nearly half of those working in STEM jobs–explain the persistent gender wage gap in computer science, using American Community Survey (ACS) data from 2009 to 2019. Our analysis focuses on working-age respondents between the ages of 22 and 60 who had a college degree and were employed full-time. We use ordinary least squares (OLS) regression of logged wages on observed characteristics, before turning to regression decomposition techniques to estimate what proportion of the gender wage gap would remain if men and women were equally rewarded for the same attributes–such as parenthood or marital status, degree field, or occupation. Women employed in computer science jobs earned about 86.6 cents for every dollar that men earned–a raw gender gap that is smaller than it is for the overall labor force (where it was 82 percent). Controlling for compositional effects (family attributes, degree field and occupation) narrows the gender wage gap, though women continue to earn 9.1 cents per dollar less than their male counterparts. But differential returns to family characteristics and human capital measures account for almost two-thirds of the gender wage gap in computer science jobs. Women working in computer science receive both a marriage and parenthood premium relative to unmarried or childless women, but these are significantly smaller than the bonus that married men and fathers receive over their childless and unmarried peers. Men also receive sizable wage premiums for having STEM degrees in computer science and engineering when they work in computer science jobs, advantages that do not accrue to women. Closing the gender wage gap in computer science requires treating women more like men, not just increasing their representation.

Role of machine and organizational structure in science

PLOS ONE

View Article

The progress of science increasingly relies on machine learning (ML) and machines work alongside humans in various domains of science. This study investigates the team structure of ML-related projects and analyzes the contribution of ML to scientific knowledge production under different team structure, drawing on bibliometric analyses of 25,000 scientific publications in various disciplines. Our regression analyses suggest that (1) interdisciplinary collaboration between domain scientists and computer scientists as well as the engagement of interdisciplinary individuals who have expertise in both domain and computer sciences are common in ML-related projects; (2) the engagement of interdisciplinary individuals seem more important in achieving high impact and novel discoveries, especially when a project employs computational and domain approaches interdependently; and (3) the contribution of ML and its implication to team structure depend on the depth of ML.

A First Attempt to Bring Computational Biology into Advanced High School Biology Classrooms

PLoS Computational Biology

View Article

Computer science has become ubiquitous in many areas of biological research, yet most high school and even college students are unaware of this. As a result, many college biology majors graduate without adequate computational skills for contemporary fields of biology. The absence of a computational element in secondary school biology classrooms is of growing concern to the computational biology community and biology teachers who would like to acquaint their students with updated approaches in the discipline. We present a first attempt to correct this absence by introducing a computational biology element to teach genetic evolution into advanced biology classes in two local high schools. Our primary goal was to show students how computation is used in biology and why a basic understanding of computation is necessary for research in many fields of biology. This curriculum is intended to be taught by a computational biologist who has worked with a high school advanced biology teacher to adapt the unit for his/her classroom, but a motivated high school teacher comfortable with mathematics and computing may be able to teach this alone. In this paper, we present our curriculum, which takes into consideration the constraints of the required curriculum, and discuss our experiences teaching it. We describe the successes and challenges we encountered while bringing this unit to high school students, discuss how we addressed these challenges, and make suggestions for future versions of this curriculum.We believe that our curriculum can be a valuable seed for further development of computational activities aimed at high school biology students. Further, our experiences may be of value to others teaching computational biology at this level. Our curriculum can be obtained at http://ecsite.cs.colorado.edu/?page_id=149#biology or by contacting the authors. Author Summary: We have designed and implemented a curriculum to teach basic computational biology to advanced high school students. The curriculum includes an introduction to the concept of algorithms, an overview of the Basic Local Alignment Search Tool (BLAST) algorithm used to compare DNA sequences, and methods for building phylogenetic trees. We taught this curriculum in advanced biology classes at two local high schools. As a result of this, we were able to give many students an appreciation of the role computers play in biology and an idea of why computational methods are needed in biological research. We found that while the high school students lacked the necessary background in math and computer science to be able to write their own algorithms, they were able to use existing algorithms, analyze them, and compare the results. We also encountered a number of challenges that could arise in other attempts to teach computational biology to students at this level, whether using our curriculum or another. We discuss each of these challenges and possible ways that they can be overcome.

Digital imaging and vision analysis in science project improves the self-efficacy and skill of undergraduate students in computational work

PLOS ONE

View Article

In many areas of science, the ability to use computers to process, analyze, and visualize large data sets has become essential. The mismatch between the ability to generate large data sets and the computing skill to analyze them is arguably the most striking within the life sciences. The Digital Image and Vision Applications in Science (DIVAS) project describes a scaffolded series of interventions implemented over the span of a year to build the coding and computing skill of undergraduate students majoring primarily in the natural sciences. The program is designed as a community of practice, providing support within a network of learners. The program focus, images as data, provides a compelling ‘hook’ for participating scholars. Scholars begin the program with a one-credit spring semester seminar where they are exposed to image analysis. The program continues in the summer with a one-week, intensive Python and image processing workshop. From there, scholars tackle image analysis problems using a pair programming approach and can finish the summer with independent research. Finally, scholars participate in a follow-up seminar the subsequent spring and help onramp the next cohort of incoming scholars. We observed promising growth in participant self-efficacy in computing that was maintained throughout the project as well as significant growth in key computational skills. DIVAS program funding was able to support seventeen DIVAS over three years, with 76% of DIVAS scholars identifying as women and 14% of scholars identifying as members of an underrepresented minority group. Most scholars (82%) entered the program as first year students, with 94% of DIVAS scholars retained for the duration of the program and 100% of scholars remaining a STEM major one year after completing the program. The outcomes of the DIVAS project support the efficacy of building computational skill through repeated exposure of scholars to relevant applications over an extended period within a community of practice.

Key theories and technologies and implementation mechanism of parallel computing for ternary optical computer

PLOS ONE

View Article

Ternary Optical Computer (TOC) is more advanced than traditional computer systems in parallel computing, which is characterized by huge amounts of repeated computations. However, the application of the TOC is still limited because of lack of key theories and technologies. In order to make the TOC applicable and advantageous, this paper systematically elaborates the key theories and technologies of parallel computing for the TOC through a programming platform, including reconfigurability and groupable usability of optical processor bits, parallel carry-free optical adder and the TOC’s application characteristics, communication file to express user’s needs and data organization method of the TOC. Finally, experiments are carried out to show the effectiveness of the present theories and technologies for parallel computing, as well as the feasibility of the implementation method of the programming platform. For a special instance, it is shown that the clock cycle on the TOC is only 0.26% of on a traditional computer, and the computing resource spent on the TOC is 25% of that on a traditional computer. Based on the study of the TOC in this paper, more complex parallel computing can be realized in the future.

Performance comparison of heuristic algorithms for task scheduling in IaaS cloud computing environment

PLOS ONE

View Article

Cloud computing infrastructure is suitable for meeting computational needs of large task sizes. Optimal scheduling of tasks in cloud computing environment has been proved to be an NP-complete problem, hence the need for the application of heuristic methods. Several heuristic algorithms have been developed and used in addressing this problem, but choosing the appropriate algorithm for solving task assignment problem of a particular nature is difficult since the methods are developed under different assumptions. Therefore, six rule based heuristic algorithms are implemented and used to schedule autonomous tasks in homogeneous and heterogeneous environments with the aim of comparing their performance in terms of cost, degree of imbalance, makespan and throughput. First Come First Serve (FCFS), Minimum Completion Time (MCT), Minimum Execution Time (MET), Max-min, Min-min and Sufferage are the heuristic algorithms considered for the performance comparison and analysis of task scheduling in cloud computing.

Women are underrepresented in computational biology: An analysis of the scholarly literature in biology, computer science and computational biology

PLOS Computational Biology

View Article

While women are generally underrepresented in STEM fields, there are noticeable differences between fields. For instance, the gender ratio in biology is more balanced than in computer science. We were interested in how this difference is reflected in the interdisciplinary field of computational/quantitative biology. To this end, we examined the proportion of female authors in publications from the PubMed and arXiv databases. There are fewer female authors on research papers in computational biology, as compared to biology in general. This is true across authorship position, year, and journal impact factor. A comparison with arXiv shows that quantitative biology papers have a higher ratio of female authors than computer science papers, placing computational biology in between its two parent fields in terms of gender representation. Both in biology and in computational biology, a female last author increases the probability of other authors on the paper being female, pointing to a potential role of female PIs in influencing the gender balance. Author summary: There are fewer women than men working in Science, Technology, Engineering and Mathematics (STEM). However, some fields within STEM are more gender-balanced than others. For instance, biology has a relatively high proportion of women, whereas there are few women in computer science. But what about computational biology? As an interdisciplinary STEM field, would its gender balance be close to one of its “parent” fields, or in between the two? To investigate this question, we examined authorship data from databases of scholarly publications in biology, computational biology, and computer science. We found that computational biology lies in between computer science and biology, as far as female representation goes. This is independent of other factors, e.g. year of publication. This suggests that computational biology might provide an environment that is more conducive to female participation that other areas of computer science. Across all three fields, we also found that if the last author on a publication—usually the person leading the study—is a women, then there will also be more women in other authorship positions. This suggests that having women in leadership positions might be beneficial for overall gender balance, though our data do not allow us to uncover the underlying mechanism.

Computation harvesting from nature dynamics for predicting wind speed and direction

PLOS ONE

View Article

Natural phenomena generate complex dynamics because of nonlinear interactions among their components. The dynamics can be exploited as a kind of computational resource. For example, in the framework of natural computation, various natural phenomena such as quantum mechanics and cellular dynamics are used to realize general purpose calculations or logical operations. In recent years, simple collection of such nature dynamics has become possible in a sensor-rich society. For example, images of plant movement that have been captured indirectly by a surveillance camera can be regarded as sensor outputs reflecting the state of the wind striking the plant. Herein, based on ideas of physical reservoir computing, we present a methodology for wind speed and direction estimation from naturally occurring sensors in movies. Then we demonstrate its effectiveness through experimentation. Specifically using the proposed methodology, we investigate the computational capability of the nature dynamics, revealing its high robustness and generalization performance for computation.

Assessment of problem solving ability in novice programmers

PLOS ONE

View Article

Problem Solving (PS) skills allow students to handle problems within an educational context. PS is a core competence of Computer Science education and affects programming success. In this vein, this paper aims to investigate PS ability performance in primary school pupils of a computer course, implemented according to the Neo-Piagetian theory of cognitive development. The study included 945 Slovenian pupils, ranging from fourth to sixth grade. The effects of gender, age and consecutive years of attending the course were examined on pupils’ PS ability at the pre-operational and concrete operational stages. Pupils completed a survey questionnaire with four types of tasks (a series of statements, if-statements, loops and variables) at both stages. The analysis revealed three findings: the performance of PS ability in all tasks was, at the pre-operational stage, associated positively with performance at the concrete operational stage; there were no gender differences in PS performance at both stages, and both the grade and consecutive year of taking the computer course had an effect on PS ability performance at both stages. Those in the lowest grade and those taking the course for the first year reported lower performances than their older counterparts. These findings may help curriculum designers across the world develop efficient approaches to teaching computer courses.

Energy-Harvesting Reinforcement Learning-based Offloading Decision Algorithm for Mobile Edge Computing Networks (EHRL)

PLOS One

View Article

Mobile Edge Computing (MEC) is a computational paradigm that brings resources closer to the network edge to provide fast and efficient computing services for Mobile Devices (MDs). However, MDs are often constrained by limited energy and computational resources, which are insufficient to handle the high number of tasks. The problems of limited energy resources and the low computing capability of wireless nodes have led to the emergence of Wireless Power Transfer (WPT) and Energy Harvesting (EH) as a potential solution where electrical energy is transmitted wirelessly and then harvested by MDs and converted into power. This paper considers a wireless-powered MEC network employing a binary offloading policy, in which the computation tasks of MDs are either executed locally or fully offloaded to an edge server (ES). The objective is to optimize binary offloading decisions under dynamic wireless channel conditions and energy harvesting constraints. Hence, an Energy-Harvesting Reinforcement Learning-based Offloading Decision Algorithm (EHRL) is proposed. EHRL integrates Reinforcement Learning (RL) with Deep Neural Networks (DNNs) to dynamically optimize binary offloading decisions, which in turn obviates the requirement for manually labeled training data and thus avoids the need for solving complex optimization problems repeatedly. To enhance the offloading decision-making process, the algorithm incorporates the Newton-Raphson method for fast and efficient optimization of the computation rate under energy constraints. Simultaneously, the DNN is trained using the Nadam optimizer (Nesterov-accelerated Adaptive Moment Estimation), which combines the benefits of Adam and Nesterov momentum, offering improved convergence speed and training stability. The proposed algorithm addresses the dual challenges of limited energy availability in MDs and the need for efficient task offloading to minimize latency and maximize computational performance. Numerical results validate the superiority of the proposed approach, demonstrating significant gains in computation performance and time efficiency compared to conventional techniques, making real-time and optimal offloading design truly viable even in a fast-fading environment.

An implementation framework to improve the transparency and reproducibility of computational models of infectious diseases

PLOS Computational Biology

View Article

Computational models of infectious diseases have become valuable tools for research and the public health response against epidemic threats. The reproducibility of computational models has been limited, undermining the scientific process and possibly trust in modeling results and related response strategies, such as vaccination. We translated published reproducibility guidelines from a wide range of scientific disciplines into an implementation framework for improving reproducibility of infectious disease computational models. The framework comprises 22 elements that should be described, grouped into 6 categories: computational environment, analytical software, model description, model implementation, data, and experimental protocol. The framework can be used by scientific communities to develop actionable tools for sharing computational models in a reproducible way.

Wrangling distributed computing for high-throughput environmental science: An introduction to HTCondor

PLOS Computational Biology

View Article

Biologists and environmental scientists now routinely solve computational problems that were unimaginable a generation ago. Examples include processing geospatial data, analyzing -omics data, and running large-scale simulations. Conventional desktop computing cannot handle these tasks when they are large, and high-performance computing is not always available nor the most appropriate solution for all computationally intense problems. High-throughput computing (HTC) is one method for handling computationally intense research. In contrast to high-performance computing, which uses a single "supercomputer," HTC can distribute tasks over many computers (e.g., idle desktop computers, dedicated servers, or cloud-based resources). HTC facilities exist at many academic and government institutes and are relatively easy to create from commodity hardware. Additionally, consortia such as Open Science Grid facilitate HTC, and commercial entities sell cloud-based solutions for researchers who lack HTC at their institution. We provide an introduction to HTC for biologists and environmental scientists. Our examples from biology and the environmental sciences use HTCondor, an open source HTC system. Author summary: Computational biology often requires processing large amounts of data, running many simulations, or other computationally intensive tasks. In this hybrid primer/tutorial, we describe how high-throughput computing (HTC) can be used to solve these problems. First, we present an overview of high-throughput computing. Second, we describe how to break jobs down so that they can run with HTC. Third, we describe how to use HTCondor software as a method for HTC. Fourth, we describe how HTCondor may be applied to other situations and a series of online tutorials.

Computational thinking in university students: The role of fluid intelligence and visuospatial ability

PLOS ONE

View Article

Computational thinking (CT) is a set of problem-solving skills with high relevance in education and work contexts. The present paper explores the role of key cognitive factors underlying CT performance in non-programming university students. We collected data from 97 non-programming adults in higher education in a supervised setting. Fluid intelligence, crystallized intelligence, and visuospatial ability were assessed using computerized adaptive tests; CT was measured using the Computational Thinking test. The direct and indirect effects of gender and visuospatial ability through fluid intelligence on CT were tested in a serial multiple mediator model. Fluid intelligence predicted CT when controlling for the effects of gender, age, and visuospatial ability, while crystallized intelligence did not predict CT. Men had a small advantage in CT performance when holding the effects of cognitive abilities constant. Despite its large correlation with gender and CT, visuospatial ability did not directly influence CT performance. Overall, we found that programming-naive computational thinkers draw on their reasoning ability that does not rely on previously acquired knowledge to solve CT problems. Visuospatial ability and CT were spuriously associated. Drawing on the process overlap theory we propose that tests of fluid intelligence and CT sample an overlapping set of underlying visuospatial processes.

Computational Biology in Cuba: An Opportunity to Promote Science in a Developing Country

PLoS Computational Biology

View Article

What do computer scientists tweet? Analyzing the link-sharing practice on Twitter

PLOS ONE

View Article

Twitter communication has permeated every sphere of society. To highlight and share small pieces of information with possibly vast audiences or small circles of the interested has some value in almost any aspect of social life. But what is the value exactly for a scientific field? We perform a comprehensive study of computer scientists using Twitter and their tweeting behavior concerning the sharing of web links. Discerning the domains, hosts and individual web pages being tweeted and the differences between computer scientists and a Twitter sample enables us to look in depth at the Twitter-based information sharing practices of a scientific community. Additionally, we aim at providing a deeper understanding of the role and impact of altmetrics in computer science and give a glance at the publications mentioned on Twitter that are most relevant for the computer science community. Our results show a link sharing culture that concentrates more heavily on public and professional quality information than the Twitter sample does. The results also show a broad variety in linked sources and especially in linked publications with some publications clearly related to community-specific interests of computer scientists, while others with a strong relation to attention mechanisms in social media. This refers to the observation that Twitter is a hybrid form of social media between an information service and a social network service. Overall the computer scientists’ style of usage seems to be more on the information-oriented side and to some degree also on professional usage. Therefore, altmetrics are of considerable use in analyzing computer science.

All biology is computational biology

PLOS Biology

View Article

Here, I argue that computational thinking and techniques are so central to the quest of understanding life that today all biology is computational biology. Computational biology brings order into our understanding of life, it makes biological concepts rigorous and testable, and it provides a reference map that holds together individual insights. The next modern synthesis in biology will be driven by mathematical, statistical, and computational methods being absorbed into mainstream biological training, turning biology into a quantitative science.

Application-aware deadline constraint job scheduling mechanism on large-scale computational grid

PLOS ONE

View Article

Recently, computational Grids have proven to be a good solution for processing large-scale, computation intensive problems. However, the heterogeneity, dynamics of resources and diversity of applications requirements have always been important factors affecting their performance. In response to these challenges, this work first builds a Grid job scheduling architecture that can dynamically monitor Grid computing center resources and make corresponding scheduling decisions. Second, a Grid job model is proposed to describe the application requirements. Third, this paper studies the characteristics of commercial interconnection networks used in Grids and forecast job transmission time. Fourth, this paper proposes an application-aware job scheduling mechanism (AJSM) that includes periodic scheduling flow and a heuristic application-aware deadline constraint job scheduling algorithm. The rigorous performance evaluation results clearly demonstrate that the proposed application-aware job scheduling mechanism can successful schedule more Grid jobs than the existing algorithms. For successful scheduled jobs, our proposed AJSM method is the best algorithm for job average processing time and makespan.

A security technology of power relay using edge computing

PLOS ONE

View Article

The purposes are to find the techniques suitable for the safety relay protection of intelligent substations and discuss the applicability of edge computing in relay protection. Regarding relay protection in intelligent substations, edge computing and optimized simulated annealing algorithm (OSAA) are combined innovatively to form an edge computing strategy. On this basis, an edge computing model is proposed based on relay fault traveling waves. Under different computing shunt tasks, OSAA can converge after about 1,100 iterations, and its computing time is relatively short. As the global optimal time delay reaches 0.5295, the corresponding computing time is 456.27s, apparently better than the linear search method. The proposed model can reduce the computing time significantly, playing an active role in the safe shunting of power relays. The simulation also finds that the voltage and current waveforms corresponding to the fault state of Phase A are consistent with the actual situations. To sum up, this model provides a reference for improving and optimizing intelligent substation relay protection.

Securing the future of research computing in the biosciences

PLOS Computational Biology

View Article

Author summary: Improvements in technology often drive scientific discovery. Therefore, research requires sustained investment in the latest equipment and training for the researchers who are going to use it. Prioritising and administering infrastructure investment is challenging because future needs are difficult to predict. In the past, highly computationally demanding research was associated primarily with particle physics and astronomy experiments. However, as biology becomes more quantitative and bioscientists generate more and more data, their computational requirements may ultimately exceed those of physical scientists. Computation has always been central to bioinformatics, but now imaging experiments have rapidly growing data processing and storage requirements. There is also an urgent need for new modelling and simulation tools to provide insight and understanding of these biophysical experiments. Bioscience communities must work together to provide the software and skills training needed in their areas. Research-active institutions need to recognise that computation is now vital in many more areas of discovery and create an environment where it can be embraced. The public must also become aware of both the power and limitations of computing, particularly with respect to their health and personal data.