id
stringlengths
6
15
question_type
stringclasses
1 value
question
stringlengths
15
683
choices
listlengths
4
4
answer
stringclasses
5 values
explanation
stringclasses
481 values
prompt
stringlengths
1.75k
10.9k
sciq-5905
multiple_choice
Sodium is an example of what sort of thing that cannot exist by itself in nature?
[ "aspect", "element", "compound", "concept" ]
B
Relavent Documents: Document 0::: Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas. Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below: During adiabatic expansion of an ideal gas, its temperatureincreases decreases stays the same Impossible to tell/need more information The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well. Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in Document 1::: Sodium hydride is the chemical compound with the empirical formula NaH. This alkali metal hydride is primarily used as a strong yet combustible base in organic synthesis. NaH is a saline (salt-like) hydride, composed of Na+ and H− ions, in contrast to molecular hydrides such as borane, methane, ammonia, and water. It is an ionic material that is insoluble in all solvents (other than molten Na), consistent with the fact that H− ions do not exist in solution. Because of the insolubility of NaH, all reactions involving NaH occur at the surface of the solid. Basic properties and structure NaH is produced by the direct reaction of hydrogen and liquid sodium. Pure NaH is colorless, although samples generally appear grey. NaH is around 40% denser than Na (0.968 g/cm3). NaH, like LiH, KH, RbH, and CsH, adopts the NaCl crystal structure. In this motif, each Na+ ion is surrounded by six H− centers in an octahedral geometry. The ionic radii of H− (146 pm in NaH) and F− (133 pm) are comparable, as judged by the Na−H and Na−F distances. "Inverse sodium hydride" A very unusual situation occurs in a compound dubbed "inverse sodium hydride", which contains H+ and Na− ions. Na− is an alkalide, and this compound differs from ordinary sodium hydride in having a much higher energy content due to the net displacement of two electrons from hydrogen to sodium. A derivative of this "inverse sodium hydride" arises in the presence of the base [36]adamanzane. This molecule irreversibly encapsulates the H+ and shields it from interaction with the alkalide Na−. Theoretical work has suggested that even an unprotected protonated tertiary amine complexed with the sodium alkalide might be metastable under certain solvent conditions, though the barrier to reaction would be small and finding a suitable solvent might be difficult. Applications in organic synthesis As a strong base NaH is a base of wide scope and utility in organic chemistry. As a superbase, it is capable of deprotonating a ra Document 2::: A nonmetal is a chemical element that mostly lacks metallic properties. Seventeen elements are generally considered nonmetals, though some authors recognize more or fewer depending on the properties considered most representative of metallic or nonmetallic character. Some borderline elements further complicate the situation. Nonmetals tend to have low density and high electronegativity (the ability of an atom in a molecule to attract electrons to itself). They range from colorless gases like hydrogen to shiny solids like the graphite form of carbon. Nonmetals are often poor conductors of heat and electricity, and when solid tend to be brittle or crumbly. In contrast, metals are good conductors and most are pliable. While compounds of metals tend to be basic, those of nonmetals tend to be acidic. The two lightest nonmetals, hydrogen and helium, together make up about 98% of the observable ordinary matter in the universe by mass. Five nonmetallic elements—hydrogen, carbon, nitrogen, oxygen, and silicon—make up the overwhelming majority of the Earth's crust, atmosphere, oceans and biosphere. The distinct properties of nonmetallic elements allow for specific uses that metals often cannot achieve. Elements like hydrogen, oxygen, carbon, and nitrogen are essential building blocks for life itself. Moreover, nonmetallic elements are integral to industries such as electronics, energy storage, agriculture, and chemical production. Most nonmetallic elements were not identified until the 18th and 19th centuries. While a distinction between metals and other minerals had existed since antiquity, a basic classification of chemical elements as metallic or nonmetallic emerged only in the late 18th century. Since then nigh on two dozen properties have been suggested as single criteria for distinguishing nonmetals from metals. Definition and applicable elements Properties mentioned hereafter refer to the elements in their most stable forms in ambient conditions unless otherwise Document 3::: The nitrite ion has the chemical formula . Nitrite (mostly sodium nitrite) is widely used throughout chemical and pharmaceutical industries. The nitrite anion is a pervasive intermediate in the nitrogen cycle in nature. The name nitrite also refers to organic compounds having the –ONO group, which are esters of nitrous acid. Production Sodium nitrite is made industrially by passing a mixture of nitrogen oxides into aqueous sodium hydroxide or sodium carbonate solution: The product is purified by recrystallization. Alkali metal nitrites are thermally stable up to and beyond their melting point (441 °C for KNO2). Ammonium nitrite can be made from dinitrogen trioxide, N2O3, which is formally the anhydride of nitrous acid: 2 NH3 + H2O + N2O3 → 2 NH4NO2 Structure The nitrite ion has a symmetrical structure (C2v symmetry), with both N–O bonds having equal length and a bond angle of about 115°. In valence bond theory, it is described as a resonance hybrid with equal contributions from two canonical forms that are mirror images of each other. In molecular orbital theory, there is a sigma bond between each oxygen atom and the nitrogen atom, and a delocalized pi bond made from the p orbitals on nitrogen and oxygen atoms which is perpendicular to the plane of the molecule. The negative charge of the ion is equally distributed on the two oxygen atoms. Both nitrogen and oxygen atoms carry a lone pair of electrons. Therefore, the nitrite ion is a Lewis base. In the gas phase it exists predominantly as a trans-planar molecule. Reactions Acid-base properties Nitrite is the conjugate base of the weak acid nitrous acid: HNO2 H+ + ; pKa ≈ 3.3 at 18 °C Nitrous acid is also highly volatile, tending to disproportionate: 3 HNO2 (aq) H3O+ + + 2 NO This reaction is slow at 0 °C. Addition of acid to a solution of a nitrite in the presence of a reducing agent, such as iron(II), is a way to make nitric oxide (NO) in the laboratory. Oxidation and reduction The formal oxidation sta Document 4::: With Sn2+ ions, N2O is formed: 2 HNO2 + 6 HCl + 2 SnCl2 → 2 SnCl4 + N2O + 3 H2O + 2 KCl With SO2 gas, NH2OH is formed: 2 HNO2 + 6 H2O + 4 SO2 → 3 H2SO4 + K2SO4 + 2 NH2OH With Zn in alkali solution, NH3 is formed: 5 H2O + KNO2 + 3 Zn → NH3 + KOH + 3 Zn(OH)2 With , both HN3 The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses. Sodium is an example of what sort of thing that cannot exist by itself in nature? A. aspect B. element C. compound D. concept Answer:
sciq-5111
multiple_choice
Toward the late stages of pregnancy, a drop in progesterone and stretching forces from the fetus lead to increasing uterine irritability and prompt labor. contractions serve to dilate the cervix and expel this?
[ "uteris", "blood", "tissues", "newborn" ]
D
Relavent Documents: Document 0::: The human reproductive system includes the male reproductive system which functions to produce and deposit sperm; and the female reproductive system which functions to produce egg cells, and to protect and nourish the fetus until birth. Humans have a high level of sexual differentiation. In addition to differences in nearly every reproductive organ, there are numerous differences in typical secondary sex characteristics. Human reproduction usually involves internal fertilization by sexual intercourse. In this process, the male inserts his penis into the female's vagina and ejaculates semen, which contains sperm. A small proportion of the sperm pass through the cervix into the uterus, and then into the fallopian tubes for fertilization of the ovum. Only one sperm is required to fertilize the ovum. Upon successful fertilization, the fertilized ovum, or zygote, travels out of the fallopian tube and into the uterus, where it implants in the uterine wall. This marks the beginning of gestation, better known as pregnancy, which continues for around nine months as the fetus develops. When the fetus has developed to a certain point, pregnancy is concluded with childbirth, involving labor. During labor, the muscles of the uterus contract and the cervix dilates over the course of hours, and the baby passes out of the vagina. Human infants are completely dependent on their caregivers, and require high levels of parental care. Infants rely on their caregivers for comfort, cleanliness, and food. Food may be provided by breastfeeding or formula feeding. Structure Female The human female reproductive system is a series of organs primarily located inside the body and around the pelvic region of a female that contribute towards the reproductive process. The human female reproductive system contains three main parts: the vulva, which leads to the vagina, the vaginal opening, to the uterus; the uterus, which holds the developing fetus; and the ovaries, which produce the female's o Document 1::: Prenatal perception is the study of the extent of somatosensory and other types of perception during pregnancy. In practical terms, this means the study of fetuses; none of the accepted indicators of perception are present in embryos. Studies in the field inform the abortion debate, along with certain related pieces of legislation in countries affected by that debate. As of 2022, there is no scientific consensus on whether a fetus can feel pain. Prenatal hearing Numerous studies have found evidence indicating a fetus's ability to respond to auditory stimuli. The earliest fetal response to a sound stimulus has been observed at 16 weeks' gestational age, while the auditory system is fully functional at 25–29 weeks' gestation. At 33–41 weeks' gestation, the fetus is able to distinguish its mother's voice from others. Prenatal pain The hypothesis that human fetuses are capable of perceiving pain in the first trimester has little support, although fetuses at 14 weeks may respond to touch. A multidisciplinary systematic review from 2005 found limited evidence that thalamocortical pathways begin to function "around 29 to 30 weeks' gestational age", only after which a fetus is capable of feeling pain. In March 2010, the Royal College of Obstetricians and Gynecologists submitted a report, concluding that "Current research shows that the sensory structures are not developed or specialized enough to respond to pain in a fetus of less than 24 weeks", The report specifically identified the anterior cingulate as the area of the cerebral cortex responsible for pain processing. The anterior cingulate is part of the cerebral cortex, which begins to develop in the fetus at week 26. A co-author of that report revisited the evidence in 2020, specifically the functionality of the thalamic projections into the cortical subplate, and posited "an immediate and unreflective pain experience...from as early as 12 weeks." There is a consensus among developmental neurobiologists that the Document 2::: The obstetrical dilemma is a hypothesis to explain why humans often require assistance from other humans during childbirth to avoid complications, whereas most non-human primates give birth unassisted with relatively little difficulty. This occurs due to the tight fit of the fetal head to the maternal birth canal, which is additionally convoluted, meaning the head and therefore body of the infant must rotate during childbirth in order to fit, unlike in other, non-upright walking mammals. Consequently, there is a usually high incidence of cephalopelvic disproportion and obstructed labor in humans. The obstetrical dilemma claims that this difference is due to the biological trade-off imposed by two opposing evolutionary pressures in the development of the human pelvis: smaller birth canals in the mothers, and larger brains, and therefore skulls in the babies. Proponents believe bipedal locomotion (the ability to walk upright) decreased the size of the bony parts of the birth canal. They also believe that as hominids' and humans' skull and brain sizes increased over the millennia, that women needed wider hips to give birth, that these wider hips made women inherently less able to walk or run than men, and that babies had to be born earlier to fit through the birth canal, resulting in the so-called fourth trimester period for newborns (being born when the baby seems less developed than in other animals). Recent evidence has suggested bipedal locomotion is only a part of the strong evolutionary pressure constraining the expansion of the maternal birth canal. In addition to bipedal locomotion, the reduced strength of the pelvic floor due to a wider maternal pelvis also leads to fitness detriments in the mother pressuring the birth canal to remain relatively narrow. This idea was widely accepted when first published in 1960, but has since been criticized by other scientists. History The term, obstetrical dilemma, was coined in 1960, by Sherwood Larned Washburn, a pr Document 3::: Early pregnancy loss is a medical term that when referring to humans can variously be used to mean: Death of an embryo or fetus during the first trimester. This can happen by implantation failure, miscarriage, embryo resorption, early fetal resorption or vanishing twin syndrome. Death of an embryo or fetus before 20 weeks gestation, as in all pregnancy loss before it becomes considered stillbirth. Causes of early pregnancy loss Pregnancy loss, in many cases, occurs for unknown reasons, often involving random chromosome issues during conception. Miscarriage is not caused by everyday activities like working, exercising, or having sex. Even falls or blows are rarely to blame. Research on the effects of alcohol, tobacco, and caffeine on miscarriage is inconclusive, so it's not something you could have prevented. It's crucial not to blame yourself for a miscarriage, as it's not the result of anything you did or didn't do. Symptoms of early pregnancy loss The most prevalent indication of pregnancy loss is vaginal bleeding. In the later stages of pregnancy, a woman experiencing a stillbirth may cease to sense fetal movements. However, it's important to note that each type of pregnancy loss presents distinct symptoms, so it's essential to consult your healthcare provider for a proper diagnosis. See also Pregnancy with abortive outcome Document 4::: Montevideo units are a method of measuring uterine performance during labor. They were created in 1949 by two physicians, Roberto Caldeyro-Barcia and Hermogenes Alvarez, from Montevideo, Uruguay. They are exactly equal to 1 mmHg within 10 minutes. A standard adequate measurement is 200; this is generally equivalent to 27 kPa of combined pressure change within 10 minutes. Units are directly equal to pressure change in mmHg summed over a ten-minute window. It is calculated by internally (not externally) measuring peak uterine pressure amplitude (in mmHg), subtracting the resting tone of the contraction, and adding up the numbers in a 10-minute period. Uterine pressure is generally measured through an intrauterine pressure catheter. Montevideo units can be more simply calculated by summing the individual contraction intensities in a ten-minute period, a process which should arrive at a result identical to the original method of calculation. Generally, above 200 MVUs is considered necessary for adequate labor during the active phase. Example If, for instance: Peak uterine pressure amplitudes were 50 mmHg during the 10 minute period of measurement 3 contractions occurred subtract the resting tone from the peak intensity of the contraction add the 3 contractions together to get the MVUs Montevideo units are calculated by obtaining the peak uterine pressure amplitude and subtracting the resting tone. Then adding up those numbers generated by each contraction within a 10-minute window. For example, five contractions occurred, producing peak pressures of 55, 50, 45, 65, and 50 mm Hg, respectively. The resting tone of the contractions is 10. 55-10 = 45 50-10 = 40 45-10 = 35 65-10 = 55 50-10 = 40 45+40+35+55+40 = 215 MVUs The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses. Toward the late stages of pregnancy, a drop in progesterone and stretching forces from the fetus lead to increasing uterine irritability and prompt labor. contractions serve to dilate the cervix and expel this? A. uteris B. blood C. tissues D. newborn Answer:
sciq-8006
multiple_choice
Bony fish can live in either of what type of aquatic biome?
[ "estuaries or saltwater", "oceans or estuaries", "coral reefs or freshwater", "saltwater or freshwater" ]
D
Relavent Documents: Document 0::: A Geographic Information System is a tool for mapping and analyzing data. The ability to layer many features onto the same map and select or unselect as needed allows for a multitude of views and ease of interpreting data. More important, this allows for in depth scientific analysis and problem solving. Ichthyology involves many areas of study related to fishes and their habitat. The natural habitat is water, but fish are dependent upon many other factors. Water quality, type, food, cover, sediment are essential for the life cycle of any given fish. Being able to map the presence of certain species with layers of these features provides invaluable insight into species requirements. GIS is an essential tool that allows immediate visualization of all data present and to accurately interpret impacts of habitat degradation or species success. GIS GIS is useful when data is specific to a location. It is used to classify, analyze and understand data relationships based on the location and then drawing conclusions from the data. Data capture can occur in the field on small, handheld GPS devices, and then imported and compared to an existing map. This freedom of movement between field and computer is critical to streamlining data collection in field endeavors and generating more accurate data sets. Ichthyology Ichthyology requires an understanding species geographic requirements. Fish require different abiotic environments or sediments for successful completion of biological life cycle based on species. Serious examinations of species should always include habitat because habitat differences create changes in population. Sediment could thereby be mapped and changes in sediment could easily be verified using previous records while simultaneously showing changes in resident fish populations. Various factors relating to the fish life cycle, such as food sources, migration patterns, changes in spawning grounds, could all be more accurately explained and documente Document 1::: Fisheries science is the academic discipline of managing and understanding fisheries. It is a multidisciplinary science, which draws on the disciplines of limnology, oceanography, freshwater biology, marine biology, meteorology, conservation, ecology, population dynamics, economics, statistics, decision analysis, management, and many others in an attempt to provide an integrated picture of fisheries. In some cases new disciplines have emerged, as in the case of bioeconomics and fisheries law. Because fisheries science is such an all-encompassing field, fisheries scientists often use methods from a broad array of academic disciplines. Over the most recent several decades, there have been declines in fish stocks (populations) in many regions along with increasing concern about the impact of intensive fishing on marine and freshwater biodiversity. Fisheries science is typically taught in a university setting, and can be the focus of an undergraduate, master's or Ph.D. program. Some universities offer fully integrated programs in fisheries science. Graduates of university fisheries programs typically find employment as scientists, fisheries managers of both recreational and commercial fisheries, researchers, aquaculturists, educators, environmental consultants and planners, conservation officers, and many others. Fisheries research Because fisheries take place in a diverse set of aquatic environments (i.e., high seas, coastal areas, large and small rivers, and lakes of all sizes), research requires different sampling equipment, tools, and techniques. For example, studying trout populations inhabiting mountain lakes requires a very different set of sampling tools than, say, studying salmon in the high seas. Ocean fisheries research vessels (FRVs) often require platforms which are capable of towing different types of fishing nets, collecting plankton or water samples from a range of depths, and carrying acoustic fish-finding equipment. Fisheries research vessels a Document 2::: Aquatic science is the study of the various bodies of water that make up our planet including oceanic and freshwater environments. Aquatic scientists study the movement of water, the chemistry of water, aquatic organisms, aquatic ecosystems, the movement of materials in and out of aquatic ecosystems, and the use of water by humans, among other things. Aquatic scientists examine current processes as well as historic processes, and the water bodies that they study can range from tiny areas measured in millimeters to full oceans. Moreover, aquatic scientists work in Interdisciplinary groups. For example, a physical oceanographer might work with a biological oceanographer to understand how physical processes, such as tropical cyclones or rip currents, affect organisms in the Atlantic Ocean. Chemists and biologists, on the other hand, might work together to see how the chemical makeup of a certain body of water affects the plants and animals that reside there. Aquatic scientists can work to tackle global problems such as global oceanic change and local problems, such as trying to understand why a drinking water supply in a certain area is polluted. There are two main fields of study that fall within the field of aquatic science. These fields of study include oceanography and limnology. Oceanography Oceanography refers to the study of the physical, chemical, and biological characteristics of oceanic environments. Oceanographers study the history, current condition, and future of the planet's oceans. They also study marine life and ecosystems, ocean circulation, plate tectonics, the geology of the seafloor, and the chemical and physical properties of the ocean. Oceanography is interdisciplinary. For example, there are biological oceanographers and marine biologists. These scientists specialize in marine organisms. They study how these organisms develop, their relationship with one another, and how they interact and adapt to their environment. Biological oceanographers Document 3::: AquaMaps is a collaborative project with the aim of producing computer-generated (and ultimately, expert reviewed) predicted global distribution maps for marine species on a 0.5 x 0.5 degree grid of the oceans based on data available through online species databases such as FishBase and SeaLifeBase and species occurrence records from OBIS or GBIF and using an environmental envelope model (see niche modelling) in conjunction with expert input. The underlying model represents a modified version of the relative environmental suitability (RES) model developed by Kristin Kaschner to generate global predictions of marine mammal occurrences. According to the AquaMaps website in August 2013, the project held standardized distribution maps for over 17,300 species of fishes, marine mammals and invertebrates. The project is also expanding to incorporate freshwater species, with more than 600 biodiversity maps for freshwater fishes of the Americas available as at November 2009. AquaMaps predictions have been validated successfully for a number of species using independent data sets and the model was shown to perform equally well or better than other standard species distribution models, when faced with the currently existing suboptimal input data sets. In addition to displaying individual maps per species, AquaMaps provides tools to generate species richness maps by higher taxon, plus a spatial search for all species overlapping a specified grid square. There is also the facility to create custom maps for any species via the web by modifying the input parameters and re-running the map generating algorithm in real time, and a variety of other tools including the investigation of effects of climate change on species distributions (see relevant section of the AquaMaps search page). Coordination The project is coordinated by Dr Rainer Froese of IFM-GEOMAR and involves contributions from other research institutes including the Evolutionary Biology and Ecology Lab, Albert-Ludwigs Document 4::: Integrated multi-trophic aquaculture (IMTA) provides the byproducts, including waste, from one aquatic species as inputs (fertilizers, food) for another. Farmers combine fed aquaculture (e.g., fish, shrimp) with inorganic extractive (e.g., seaweed) and organic extractive (e.g., shellfish) aquaculture to create balanced systems for environment remediation (biomitigation), economic stability (improved output, lower cost, product diversification and risk reduction) and social acceptability (better management practices). Selecting appropriate species and sizing the various populations to provide necessary ecosystem functions allows the biological and chemical processes involved to achieve a stable balance, mutually benefiting the organisms and improving ecosystem health. Ideally, the co-cultured species each yield valuable commercial "crops". IMTA can synergistically increase total output, even if some of the crops yield less than they would, short-term, in a monoculture. Terminology and related approaches "Integrated" refers to intensive and synergistic cultivation, using water-borne nutrient and energy transfer. "Multi-trophic" means that the various species occupy different trophic levels, i.e., different (but adjacent) links in the food chain. IMTA is a specialized form of the age-old practice of aquatic polyculture, which was the co-culture of various species, often without regard to trophic level. In this broader case, the organisms may share biological and chemical processes that may be minimally complementary, potentially leading to reduced production of both species due to competition for the same food resource. However, some traditional systems such as polyculture of carps in China employ species that occupy multiple niches within the same pond, or the culture of fish that is integrated with a terrestrial agricultural species, can be considered forms of IMTA. The more general term "Integrated Aquaculture" is used to describe the integration of monocult The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses. Bony fish can live in either of what type of aquatic biome? A. estuaries or saltwater B. oceans or estuaries C. coral reefs or freshwater D. saltwater or freshwater Answer:
sciq-5705
multiple_choice
What advantage equals the distance of effort divided by the distance the object moves and is also equal to the output force divided by the input force?
[ "atomic", "mechanical", "potential", "constant" ]
B
Relavent Documents: Document 0::: In physics, work is the energy transferred to or from an object via the application of force along a displacement. In its simplest form, for a constant force aligned with the direction of motion, the work equals the product of the force strength and the distance traveled. A force is said to do positive work if when applied it has a component in the direction of the displacement of the point of application. A force does negative work if it has a component opposite to the direction of the displacement at the point of application of the force. For example, when a ball is held above the ground and then dropped, the work done by the gravitational force on the ball as it falls is positive, and is equal to the weight of the ball (a force) multiplied by the distance to the ground (a displacement). If the ball is thrown upwards, the work done by the gravitational force is negative, and is equal to the weight multiplied by the displacement in the upwards direction. Both force and displacement are vectors. The work done is given by the dot product of the two vectors. When the force is constant and the angle between the force and the displacement is also constant, then the work done is given by: Work is a scalar quantity, so it has only magnitude and no direction. Work transfers energy from one place to another, or one form to another. The SI unit of work is the joule (J), the same unit as for energy. History The ancient Greek understanding of physics was limited to the statics of simple machines (the balance of forces), and did not include dynamics or the concept of work. During the Renaissance the dynamics of the Mechanical Powers, as the simple machines were called, began to be studied from the standpoint of how far they could lift a load, in addition to the force they could apply, leading eventually to the new concept of mechanical work. The complete dynamic theory of simple machines was worked out by Italian scientist Galileo Galilei in 1600 in Le Meccaniche (On Me Document 1::: Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas. Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below: During adiabatic expansion of an ideal gas, its temperatureincreases decreases stays the same Impossible to tell/need more information The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well. Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in Document 2::: Engineering mathematics is a branch of applied mathematics concerning mathematical methods and techniques that are typically used in engineering and industry. Along with fields like engineering physics and engineering geology, both of which may belong in the wider category engineering science, engineering mathematics is an interdisciplinary subject motivated by engineers' needs both for practical, theoretical and other considerations outside their specialization, and to deal with constraints to be effective in their work. Description Historically, engineering mathematics consisted mostly of applied analysis, most notably: differential equations; real and complex analysis (including vector and tensor analysis); approximation theory (broadly construed, to include asymptotic, variational, and perturbative methods, representations, numerical analysis); Fourier analysis; potential theory; as well as linear algebra and applied probability, outside of analysis. These areas of mathematics were intimately tied to the development of Newtonian physics, and the mathematical physics of that period. This history also left a legacy: until the early 20th century subjects such as classical mechanics were often taught in applied mathematics departments at American universities, and fluid mechanics may still be taught in (applied) mathematics as well as engineering departments. The success of modern numerical computer methods and software has led to the emergence of computational mathematics, computational science, and computational engineering (the last two are sometimes lumped together and abbreviated as CS&E), which occasionally use high-performance computing for the simulation of phenomena and the solution of problems in the sciences and engineering. These are often considered interdisciplinary fields, but are also of interest to engineering mathematics. Specialized branches include engineering optimization and engineering statistics. Engineering mathematics in tertiary educ Document 3::: Computer science and engineering (CSE) is an academic program at many universities which comprises computer science classes (e.g. data structures and algorithms) and computer engineering classes (e.g computer architecture). There is no clear division in computing between science and engineering, just like in the field of materials science and engineering. CSE is also a term often used in Europe to translate the name of engineering informatics academic programs. It is offered in both undergraduate as well postgraduate with specializations. Academic courses Academic programs vary between colleges, but typically include a combination of topics in computer science, computer engineering, and electrical engineering. Undergraduate courses usually include programming, algorithms and data structures, computer architecture, operating systems, computer networks, parallel computing, embedded systems, algorithms design, circuit analysis and electronics, digital logic and processor design, computer graphics, scientific computing, software engineering, database systems, digital signal processing, virtualization, computer simulations and games programming. CSE programs also include core subjects of theoretical computer science such as theory of computation, numerical methods, machine learning, programming theory and paradigms. Modern academic programs also cover emerging computing fields like image processing, data science, robotics, bio-inspired computing, computational biology, autonomic computing and artificial intelligence. Most CSE programs require introductory mathematical knowledge, hence the first year of study is dominated by mathematical courses, primarily discrete mathematics, mathematical analysis, linear algebra, probability, and statistics, as well as the basics of electrical and electronic engineering, physics, and electromagnetism. Example universities with CSE majors and departments APJ Abdul Kalam Technological University American International University-B Document 4::: The Force Concept Inventory is a test measuring mastery of concepts commonly taught in a first semester of physics developed by Hestenes, Halloun, Wells, and Swackhamer (1985). It was the first such "concept inventory" and several others have been developed since for a variety of topics. The FCI was designed to assess student understanding of the Newtonian concepts of force. Hestenes (1998) found that while "nearly 80% of the [students completing introductory college physics courses] could state Newton's Third Law at the beginning of the course, FCI data showed that less than 15% of them fully understood it at the end". These results have been replicated in a number of studies involving students at a range of institutions (see sources section below), and have led to greater recognition in the physics education research community of the importance of students' "active engagement" with the materials to be mastered. The 1995 version has 30 five-way multiple choice questions. Example question (question 4): Gender differences The FCI shows a gender difference in favor of males that has been the subject of some research in regard to gender equity in education. Men score on average about 10% higher. The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses. What advantage equals the distance of effort divided by the distance the object moves and is also equal to the output force divided by the input force? A. atomic B. mechanical C. potential D. constant Answer:
sciq-7402
multiple_choice
What do plants lose the most of through the stomata?
[ "heat", "water", "chlorophyll", "energy" ]
B
Relavent Documents: Document 0::: Stomatal conductance, usually measured in mmol m−2 s−1 by a porometer, estimates the rate of gas exchange (i.e., carbon dioxide uptake) and transpiration (i.e., water loss as water vapor) through the leaf stomata as determined by the degree of stomatal aperture (and therefore the physical resistances to the movement of gases between the air and the interior of the leaf). The stomatal conductance, or its inverse, stomatal resistance, is under the direct biological control of the leaf through its guard cells, which surround the stomatal pore. The turgor pressure and osmotic potential of guard cells are directly related to the stomatal conductance. Stomatal conductance is a function of stomatal density, stomatal aperture, and stomatal size. Stomatal conductance is integral to leaf level calculations of transpiration. Multiple studies have shown a direct correlation between the use of herbicides and changes in physiological and biochemical growth processes in plants, particularly non-target plants, resulting in a reduction in stomatal conductance and turgor pressure in leaves. Relation to stomatal opening For mechanism, see: Stomatal opening and closingStomatal conductance is a function of the density, size and degree of opening of the stomata; with more open stomata allowing greater conductance, and consequently indicating that photosynthesis and transpiration rates are potentially higher. Therefore, stomatal opening and closing has a direct relationship to stomatal conductance. Light-dependent stomatal opening Light-dependent stomatal opening occurs in many species and under many different conditions. Light is a major stimulus involved in stomatal conductance, and has two key elements that are involved in the process: 1) the stomatal response to blue light, and 2) photosynthesis in the chloroplast of the guard cell. In C3 and C4 plants, the stomata open when there is an increase in light, and they close when there is a decrease in light. In CAM plants, however, Document 1::: In botany, a stoma (from Greek στόμα, "mouth", : stomata), also called a stomate (: stomates), is a pore found in the epidermis of leaves, stems, and other organs, that controls the rate of gas exchange. The pore is bordered by a pair of specialized parenchyma cells known as guard cells that regulate the size of the stomatal opening. The term is usually used collectively to refer to the entire stomatal complex, consisting of the paired guard cells and the pore itself, which is referred to as the stomatal aperture. Air, containing oxygen, which is used in respiration, and carbon dioxide, which is used in photosynthesis, passes through stomata by gaseous diffusion. Water vapour diffuses through the stomata into the atmosphere as part of a process called transpiration. Stomata are present in the sporophyte generation of all land plant groups except liverworts. In vascular plants the number, size and distribution of stomata varies widely. Dicotyledons usually have more stomata on the lower surface of the leaves than the upper surface. Monocotyledons such as onion, oat and maize may have about the same number of stomata on both leaf surfaces. In plants with floating leaves, stomata may be found only on the upper epidermis and submerged leaves may lack stomata entirely. Most tree species have stomata only on the lower leaf surface. Leaves with stomata on both the upper and lower leaf surfaces are called amphistomatous leaves; leaves with stomata only on the lower surface are hypostomatous, and leaves with stomata only on the upper surface are epistomatous or hyperstomatous. Size varies across species, with end-to-end lengths ranging from 10 to 80 µm and width ranging from a few to 50 µm. Function CO2 gain and water loss Carbon dioxide, a key reactant in photosynthesis, is present in the atmosphere at a concentration of about 400 ppm. Most plants require the stomata to be open during daytime. The air spaces in the leaf are saturated with water vapour, which exits the Document 2::: Plant Physiology is a monthly peer-reviewed scientific journal that covers research on physiology, biochemistry, cellular and molecular biology, genetics, biophysics, and environmental biology of plants. The journal has been published since 1926 by the American Society of Plant Biologists. The current editor-in-chief is Yunde Zhao (University of California San Diego. According to the Journal Citation Reports, the journal has a 2021 impact factor of 8.005. Document 3::: Excretion is a process in which metabolic waste is eliminated from an organism. In vertebrates this is primarily carried out by the lungs, kidneys, and skin. This is in contrast with secretion, where the substance may have specific tasks after leaving the cell. Excretion is an essential process in all forms of life. For example, in mammals, urine is expelled through the urethra, which is part of the excretory system. In unicellular organisms, waste products are discharged directly through the surface of the cell. During life activities such as cellular respiration, several chemical reactions take place in the body. These are known as metabolism. These chemical reactions produce waste products such as carbon dioxide, water, salts, urea and uric acid. Accumulation of these wastes beyond a level inside the body is harmful to the body. The excretory organs remove these wastes. This process of removal of metabolic waste from the body is known as excretion. Green plants excrete carbon dioxide and water as respiratory products. In green plants, the carbon dioxide released during respiration gets used during photosynthesis. Oxygen is a by product generated during photosynthesis, and exits through stomata, root cell walls, and other routes. Plants can get rid of excess water by transpiration and guttation. It has been shown that the leaf acts as an 'excretophore' and, in addition to being a primary organ of photosynthesis, is also used as a method of excreting toxic wastes via diffusion. Other waste materials that are exuded by some plants — resin, saps, latex, etc. are forced from the interior of the plant by hydrostatic pressures inside the plant and by absorptive forces of plant cells. These latter processes do not need added energy, they act passively. However, during the pre-abscission phase, the metabolic levels of a leaf are high. Plants also excrete some waste substances into the soil around them. In animals, the main excretory products are carbon dioxide, ammoni Document 4::: Maintenance respiration (or maintenance energy) refers to metabolism occurring in an organism that is needed to maintain that organism in a healthy, living state. Maintenance respiration contrasts with growth respiration, which is responsible for the synthesis of new structures in growth, nutrient uptake, nitrogen (N) reduction and phloem loading, whereas maintenance respiration is associated with protein and membrane turnover and maintenance of ion concentrations and gradients. In plants Maintenance respiration in plants refers to the amount of cellular respiration, measured by the carbon dioxide (CO2) released or oxygen (O2) consumed, during the generation of usable energy (mainly ATP, NADPH, and NADH) and metabolic intermediates used for (i) resynthesis of compounds that undergo renewal (turnover) in the normal process of metabolism (examples are enzymatic proteins, ribonucleic acids, and membrane lipids); (ii) maintenance of chemical gradients of ions and metabolites across cellular membranes that are necessary for cellular integrity and plant health; and (iii) operation of metabolic processes involved in physiological adjustment (i.e., acclimation) to a change in the plant's environment. The metabolic costs of the repair of injury from biotic or abiotic stress may also be considered a part of maintenance respiration. Maintenance respiration is essential for biological health and growth of plants. It is estimated that about half of the respiration carried out by terrestrial plants during their lifetime is for the support of maintenance processes. Because typically more than half of global terrestrial plant photosynthesis (or gross primary production) is used for plant respiration, more than one quarter of global terrestrial plant photosynthesis is presumably consumed in maintenance respiration. Maintenance respiration is a key component of most physiologically based mathematical models of plant growth, including models of crop growth and yield and models of The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses. What do plants lose the most of through the stomata? A. heat B. water C. chlorophyll D. energy Answer:
sciq-8373
multiple_choice
What can be calculated by solving the average speed formula?
[ "mass or time", "accelaration", "distance or mass", "distance or time" ]
D
Relavent Documents: Document 0::: Advanced Placement (AP) Physics C: Mechanics (also known as AP Mechanics) is an introductory physics course administered by the College Board as part of its Advanced Placement program. It is intended to proxy a one-semester calculus-based university course in mechanics. The content of Physics C: Mechanics overlaps with that of AP Physics 1, but Physics 1 is algebra-based, while Physics C is calculus-based. Physics C: Mechanics may be combined with its electricity and magnetism counterpart to form a year-long course that prepares for both exams. Course content Intended to be equivalent to an introductory college course in mechanics for physics or engineering majors, the course modules are: Kinematics Newton's laws of motion Work, energy and power Systems of particles and linear momentum Circular motion and rotation Oscillations and gravitation. Methods of calculus are used wherever appropriate in formulating physical principles and in applying them to physical problems. Therefore, students should have completed or be concurrently enrolled in a Calculus I class. This course is often compared to AP Physics 1: Algebra Based for its similar course material involving kinematics, work, motion, forces, rotation, and oscillations. However, AP Physics 1: Algebra Based lacks concepts found in Calculus I, like derivatives or integrals. This course may be combined with AP Physics C: Electricity and Magnetism to make a unified Physics C course that prepares for both exams. AP test The course culminates in an optional exam for which high-performing students may receive some credit towards their college coursework, depending on the institution. Registration The AP examination for AP Physics C: Mechanics is separate from the AP examination for AP Physics C: Electricity and Magnetism. Before 2006, test-takers paid only once and were given the choice of taking either one or two parts of the Physics C test. Format The exam is typically administered on a Monday aftern Document 1::: Advanced Placement (AP) Physics 1 is a year-long introductory physics course administered by the College Board as part of its Advanced Placement program. It is intended to proxy a one-semester algebra-based university course in mechanics. Along with AP Physics 2, the first AP Physics 1 exam was administered in 2015. In its first five years, AP Physics 1 covered forces and motion, conservation laws, waves, and electricity. As of 2021, AP Physics 1 includes mechanics topics only. History The heavily computational AP Physics B course served for four decades as the College Board's algebra-based offering. As part of the College Board's redesign of science courses, AP Physics B was discontinued; therefore, AP Physics 1 and 2 were created with guidance from the National Research Council and the National Science Foundation. The course covers material of a first-semester university undergraduate physics course offered at American universities that use best practices of physics pedagogy. The first AP Physics 1 classes had begun in the 2014–2015 school year, with the first AP exams administered in May 2015. Curriculum AP Physics 1 is an algebra-based, introductory college-level physics course that includes mechanics topics such as motion, force, momentum, energy, harmonic motion, and rotation; The College Board published a curriculum framework that includes seven big ideas on which the AP Physics 1 and 2 courses are based, along with "enduring understandings" students are expected to acquire within each of the big ideas.: Questions for the exam are constructed with direct reference to items in the curriculum framework. Student understanding of each topic is tested with reference to multiple skills—that is, questions require students to use quantitative, semi-quantitative, qualitative, and experimental reasoning in each content area. Exam Science Practices Assessed Multiple Choice and Free Response Sections of the AP® Physics 1 exam are also assessed on scientific prac Document 2::: Advanced Placement (AP) Physics C: Electricity and Magnetism (also known as AP Physics C: E&M or AP E&M) is an introductory physics course administered by the College Board as part of its Advanced Placement program. It is intended to proxy a second-semester calculus-based university course in electricity and magnetism. The content of Physics C: E&M overlaps with that of AP Physics 2, but Physics 2 is algebra-based and covers other topics outside of electromagnetism, while Physics C is calculus-based and only covers electromagnetism. Physics C: E&M may be combined with its mechanics counterpart to form a year-long course that prepares for both exams. Course content E&M is equivalent to an introductory college course in electricity and magnetism for physics or engineering majors. The course modules are: Electrostatics Conductors, capacitors, and dielectrics Electric circuits Magnetic fields Electromagnetism. Methods of calculus are used wherever appropriate in formulating physical principles and in applying them to physical problems. Therefore, students should have completed or be concurrently enrolled in a calculus class. AP test The course culminates in an optional exam for which high-performing students may receive some credit towards their college coursework, depending on the institution. Registration The AP examination for AP Physics C: Electricity and Magnetism is separate from the AP examination for AP Physics C: Mechanics. Before 2006, test-takers paid only once and were given the choice of taking either one or two parts of the Physics C test. Format The exam is typically administered on a Monday afternoon in May. The exam is configured in two categories: a 35-question multiple choice section and a 3-question free response section. Test takers are allowed to use an approved calculator during the entire exam. The test is weighted such that each section is worth half of the final score. This and AP Physics C: Mechanics are the shortest AP exams, with Document 3::: The SAT Subject Test in Biology was the name of a one-hour multiple choice test given on biology by the College Board. A student chose whether to take the test depending upon college entrance requirements for the schools in which the student is planning to apply. Until 1994, the SAT Subject Tests were known as Achievement Tests; and from 1995 until January 2005, they were known as SAT IIs. Of all SAT subject tests, the Biology E/M test was the only SAT II that allowed the test taker a choice between the ecological or molecular tests. A set of 60 questions was taken by all test takers for Biology and a choice of 20 questions was allowed between either the E or M tests. This test was graded on a scale between 200 and 800. The average for Molecular is 630 while Ecological is 591. On January 19 2021, the College Board discontinued all SAT Subject tests, including the SAT Subject Test in Biology E/M. This was effective immediately in the United States, and the tests were to be phased out by the following summer for international students. This was done as a response to changes in college admissions due to the impact of the COVID-19 pandemic on education. Format This test had 80 multiple-choice questions that were to be answered in one hour. All questions had five answer choices. Students received one point for each correct answer, lost ¼ of a point for each incorrect answer, and received 0 points for questions left blank. The student's score was based entirely on his or her performance in answering the multiple-choice questions. The questions covered a broad range of topics in general biology. There were more specific questions related respectively on ecological concepts (such as population studies and general Ecology) on the E test and molecular concepts such as DNA structure, translation, and biochemistry on the M test. Preparation The College Board suggested a year-long course in biology at the college preparatory level, as well as a one-year course in algebra, a Document 4::: Advanced Placement (AP) Statistics (also known as AP Stats) is a college-level high school statistics course offered in the United States through the College Board's Advanced Placement program. This course is equivalent to a one semester, non-calculus-based introductory college statistics course and is normally offered to sophomores, juniors and seniors in high school. One of the College Board's more recent additions, the AP Statistics exam was first administered in May 1996 to supplement the AP program's math offerings, which had previously consisted of only AP Calculus AB and BC. In the United States, enrollment in AP Statistics classes has increased at a higher rate than in any other AP class. Students may receive college credit or upper-level college course placement upon passing the three-hour exam ordinarily administered in May. The exam consists of a multiple-choice section and a free-response section that are both 90 minutes long. Each section is weighted equally in determining the students' composite scores. History The Advanced Placement program has offered students the opportunity to pursue college-level courses while in high school. Along with the Educational Testing Service, the College Board administered the first AP Statistics exam in May 1997. The course was first taught to students in the 1996-1997 academic year. Prior to that, the only mathematics courses offered in the AP program included AP Calculus AB and BC. Students who didn't have a strong background in college-level math, however, found the AP Calculus program inaccessible and sometimes declined to take a math course in their senior year. Since the number of students required to take statistics in college is almost as large as the number of students required to take calculus, the College Board decided to add an introductory statistics course to the AP program. Since the prerequisites for such a program doesn't require mathematical concepts beyond those typically taught in a second-year al The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses. What can be calculated by solving the average speed formula? A. mass or time B. accelaration C. distance or mass D. distance or time Answer:
sciq-7780
multiple_choice
How is population density expressed?
[ "arable land per capita", "per arable land", "number per household", "number per area" ]
D
Relavent Documents: Document 0::: Outline of demography contains human demography and population related important concepts and high-level aggregated lists compiled in the useful categories. The subheadings have been grouped by the following 4 categories: Meta (lit. "highest" level) units, such as the universal important concepts related to demographics and places. Macro (lit. "high" level) units where the "whole world" is the smallest unit of measurement, such as the aggregated summary demographics at global level. For example, United Nations. Meso (lit. "middle" or "intermediate" level) units where the smallest unit of measurement cover more than one nation and more than one continent but not all the nations or continents. For example, summary list at continental level, e.g. Eurasia and Latin America or Middle East which cover two or more continents. Other examples include the intercontinental organisations e.g. the Commonwealth of Nations or the organisation of Arab states. Micro (lit. "lower" or "smaller") level units where country is the smallest unit of measurement, such as the "globally aggregated lists" by the "individual countries" . Please do not add sections on the items that are the nano (lit. "minor" or "tiny") level units as per the context described above, e.g. list of things within a city must be kept out. Meta or important concepts Global human population World population Demographics of the world Fertility and intelligence Human geography Geographic mobility Globalization Human migration List of lists on linguistics Impact of human population Human impact on the environment Biological dispersal Carrying capacity Doomsday argument Environmental migrant Human overpopulation Malthusian catastrophe List of countries by carbon dioxide emissions List of countries by carbon dioxide emissions per capita List of countries by greenhouse gas emissions List of countries by greenhouse gas emissions per capita Overconsumption Overexploitation Population eco Document 1::: Demography (), also known as Demographics, is the statistical study of populations, especially human beings. Demographic analysis examines and measures the dimensions and dynamics of populations; it can cover whole societies or groups defined by criteria such as education, nationality, religion, and ethnicity. Educational institutions usually treat demography as a field of sociology, though there are a number of independent demography departments. These methods have primarily been developed to study human populations, but are extended to a variety of areas where researchers want to know how populations of social actors can change across time through processes of birth, death, and migration. In the context of human biological populations, demographic analysis uses administrative records to develop an independent estimate of the population. Demographic analysis estimates are often considered a reliable standard for judging the accuracy of the census information gathered at any time. In the labor force, demographic analysis is used to estimate sizes and flows of populations of workers; in population ecology the focus is on the birth, death, migration and immigration of individuals in a population of living organisms, alternatively, in social human sciences could involve movement of firms and institutional forms. Demographic analysis is used in a wide variety of contexts. For example, it is often used in business plans, to describe the population connected to the geographic location of the business. Demographic analysis is usually abbreviated as DA. For the 2010 U.S. Census, The U.S. Census Bureau has expanded its DA categories. Also as part of the 2010 U.S. Census, DA now also includes comparative analysis between independent housing estimates, and census address lists at different key time points. Patient demographics form the core of the data for any medical institution, such as patient and emergency contact information and patient medical record data. They allo Document 2::: Population density (in agriculture: standing stock or plant density) is a measurement of population per unit land area. It is mostly applied to humans, but sometimes to other living organisms too. It is a key geographical term. Biological population densities Population density is population divided by total land area, sometimes including seas and oceans, as appropriate. Low densities may cause an extinction vortex and further reduce fertility. This is called the Allee effect after the scientist who identified it. Examples of the causes of reduced fertility in low population densities are: Increased problems with locating sexual mates Increased inbreeding ===Human densities=== Population density is the number of people per unit of area, usually transcribed as "per square kilometer" or square mile, and which may include or exclude, for example, areas of water or glaciers. Commonly this is calculated for a county, city, country, another territory or the entire world. The world's population is around 8,000,000,000 and the Earth's total area (including land and water) is . Therefore, from this very crude type of calculation, the worldwide human population density is approximately 8,000,000,000 ÷ 510,000,000 = . However, if only the Earth's land area of is taken into account, then human population density is . This includes all continental and island land area, including Antarctica. However, if Antarctica is excluded, then population density rises to over . The European Commission's Joint Research Centre (JRC) has developed a suite of (open and free) data and tools named the Global Human Settlement Layer (GHSL) to improve the science for policy support to the European Commission Directorate Generals and Services and as support to the United Nations system. Several of the most densely populated territories in the world are city-states, microstates and urban dependencies. In fact, 95% of the world's population is concentrated on just 10% of the world's land. Document 3::: Female education in STEM refers to child and adult female representation in the educational fields of science, technology, engineering, and mathematics (STEM). In 2017, 33% of students in STEM fields were women. The organization UNESCO has stated that this gender disparity is due to discrimination, biases, social norms and expectations that influence the quality of education women receive and the subjects they study. UNESCO also believes that having more women in STEM fields is desirable because it would help bring about sustainable development. Current status of girls and women in STEM education Overall trends in STEM education Gender differences in STEM education participation are already visible in early childhood care and education in science- and math-related play, and become more pronounced at higher levels of education. Girls appear to lose interest in STEM subjects with age, particularly between early and late adolescence. This decreased interest affects participation in advanced studies at the secondary level and in higher education. Female students represent 35% of all students enrolled in STEM-related fields of study at this level globally. Differences are also observed by disciplines, with female enrollment lowest in engineering, manufacturing and construction, natural science, mathematics and statistics and ICT fields. Significant regional and country differences in female representation in STEM studies can be observed, though, suggesting the presence of contextual factors affecting girls’ and women's engagement in these fields. Women leave STEM disciplines in disproportionate numbers during their higher education studies, in their transition to the world of work and even in their career cycle. Learning achievement in STEM education Data on gender differences in learning achievement present a complex picture, depending on what is measured (subject, knowledge acquisition against knowledge application), the level of education/age of students, and Document 4::: Homoarchy is "the relation of elements to one another when they are rigidly ranked one way only, and thus possess no (or not more than very limited) potential for being unranked or ranked in another or a number of different ways at least without cardinal reshaping of the whole socio-political order." Homoarchy and Heterarchy This notion is coupled with the one of heterarchy, defined by Crumley as "the relation of elements to one another when they are unranked or when they possess the potential for being ranked in a number of different ways". Note that heterarchy is not the opposite of any hierarchy all together, but is rather the opposite of "homoarchy". Homoarchy and Hierarchy Homoarchy must not be identified with hierarchy (as well as heterarchy must not be confused with egalitarianism in the proper meaning of the word). In any society both “vertical” and “horizontal” social links may be observed. More so: sometimes it seems too difficult to designate a society as “homoarchic” or “heterarchic” even at the most general level of analysis, like in the cases of the late-ancient Germans and early-medieval “Barbarian kingdoms” in which one can observe the monarchy and quite rigid social hierarchy combined with (at least at the beginning) democratic institutions and procedures (like selection of the king), not less significant for the whole socio-political system's operation. So, it does look like it is impossible to measure degrees of homoarchy and heterarchy in a society with mathematical exactness, for example, in per cent. A purely quantitative approach is also inapplicable here: the presence of, say five hierarchies in a society as an entity does not make it more heterarchic and less homoarchic in comparison with a society with four hierarchies if in the former there is and in the latter there is no one dominant hierarchy. The pathway to evaluation of a society as heterarchic or homoarchic (in either absolute or relative categories) goes through an analysis of i The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses. How is population density expressed? A. arable land per capita B. per arable land C. number per household D. number per area Answer:
sciq-10592
multiple_choice
What part of the cell is surrounded by a membrane and holds the genetic information for the cell?
[ "nucleus", "lower", "upper", "cell wall" ]
A
Relavent Documents: Document 0::: Cellular components are the complex biomolecules and structures of which cells, and thus living organisms, are composed. Cells are the structural and functional units of life. The smallest organisms are single cells, while the largest organisms are assemblages of trillions of cells. DNA, double stranded macromolecule that carries the hereditary information of the cell and found in all living cells; each cell carries chromosome(s) having a distinctive DNA sequence. Examples include macromolecules such as proteins and nucleic acids, biomolecular complexes such as a ribosome, and structures such as membranes, and organelles. While the majority of cellular components are located within the cell itself, some may exist in extracellular areas of an organism. Cellular components may also be called biological matter or biological material. Most biological matter has the characteristics of soft matter, being governed by relatively small energies. All known life is made of biological matter. To be differentiated from other theoretical or fictional life forms, such life may be called carbon-based, cellular, organic, biological, or even simply living – as some definitions of life exclude hypothetical types of biochemistry. See also Cell (biology) Cell biology Biomolecule Organelle Tissue (biology) External links https://web.archive.org/web/20130918033010/http://bioserv.fiu.edu/~walterm/FallSpring/review1_fall05_chap_cell3.htm Document 1::: Cell physiology is the biological study of the activities that take place in a cell to keep it alive. The term physiology refers to normal functions in a living organism. Animal cells, plant cells and microorganism cells show similarities in their functions even though they vary in structure. General characteristics There are two types of cells: prokaryotes and eukaryotes. Prokaryotes were the first of the two to develop and do not have a self-contained nucleus. Their mechanisms are simpler than later-evolved eukaryotes, which contain a nucleus that envelops the cell's DNA and some organelles. Prokaryotes Prokaryotes have DNA located in an area called the nucleoid, which is not separated from other parts of the cell by a membrane. There are two domains of prokaryotes: bacteria and archaea. Prokaryotes have fewer organelles than eukaryotes. Both have plasma membranes and ribosomes (structures that synthesize proteins and float free in cytoplasm). Two unique characteristics of prokaryotes are fimbriae (finger-like projections on the surface of a cell) and flagella (threadlike structures that aid movement). Eukaryotes Eukaryotes have a nucleus where DNA is contained. They are usually larger than prokaryotes and contain many more organelles. The nucleus, the feature of a eukaryote that distinguishes it from a prokaryote, contains a nuclear envelope, nucleolus and chromatin. In cytoplasm, endoplasmic reticulum (ER) synthesizes membranes and performs other metabolic activities. There are two types, rough ER (containing ribosomes) and smooth ER (lacking ribosomes). The Golgi apparatus consists of multiple membranous sacs, responsible for manufacturing and shipping out materials such as proteins. Lysosomes are structures that use enzymes to break down substances through phagocytosis, a process that comprises endocytosis and exocytosis. In the mitochondria, metabolic processes such as cellular respiration occur. The cytoskeleton is made of fibers that support the str Document 2::: This lecture, named in memory of Keith R. Porter, is presented to an eminent cell biologist each year at the ASCB Annual Meeting. The ASCB Program Committee and the ASCB President recommend the Porter Lecturer to the Porter Endowment each year. Lecturers Source: ASCB See also List of biology awards Document 3::: H2.00.04.4.01001: Lymphoid tissue H2.00.05.0.00001: Muscle tissue H2.00.05.1.00001: Smooth muscle tissue H2.00.05.2.00001: Striated muscle tissue H2.00.06.0.00001: Nerve tissue H2.00.06.1.00001: Neuron H2.00.06.2.00001: Synapse H2.00.06.2.00001: Neuroglia h3.01: Bones h3.02: Joints h3.03: Muscles h3.04: Alimentary system h3.05: Respiratory system h3.06: Urinary system h3.07: Genital system h3.08: Document 4::: The cell is the basic structural and functional unit of all forms of life. Every cell consists of cytoplasm enclosed within a membrane, and contains many macromolecules such as proteins, DNA and RNA, as well as many small molecules of nutrients and metabolites. The term comes from the Latin word meaning 'small room'. Cells can acquire specified function and carry out various tasks within the cell such as replication, DNA repair, protein synthesis, and motility. Cells are capable of specialization and mobility within the cell. Most plant and animal cells are only visible under a light microscope, with dimensions between 1 and 100 micrometres. Electron microscopy gives a much higher resolution showing greatly detailed cell structure. Organisms can be classified as unicellular (consisting of a single cell such as bacteria) or multicellular (including plants and animals). Most unicellular organisms are classed as microorganisms. The study of cells and how they work has led to many other studies in related areas of biology, including: discovery of DNA, cancer systems biology, aging and developmental biology. Cell biology is the study of cells, which were discovered by Robert Hooke in 1665, who named them for their resemblance to cells inhabited by Christian monks in a monastery. Cell theory, first developed in 1839 by Matthias Jakob Schleiden and Theodor Schwann, states that all organisms are composed of one or more cells, that cells are the fundamental unit of structure and function in all living organisms, and that all cells come from pre-existing cells. Cells emerged on Earth about 4 billion years ago. Discovery With continual improvements made to microscopes over time, magnification technology became advanced enough to discover cells. This discovery is largely attributed to Robert Hooke, and began the scientific study of cells, known as cell biology. When observing a piece of cork under the scope, he was able to see pores. This was shocking at the time as i The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses. What part of the cell is surrounded by a membrane and holds the genetic information for the cell? A. nucleus B. lower C. upper D. cell wall Answer:
sciq-10807
multiple_choice
While spermatogenesis produces four haploid sperm cells, oogenesis produces one mature what?
[ "ovum", "egg", "datum", "semen" ]
A
Relavent Documents: Document 0::: Spermatogenesis is the process by which haploid spermatozoa develop from germ cells in the seminiferous tubules of the testis. This process starts with the mitotic division of the stem cells located close to the basement membrane of the tubules. These cells are called spermatogonial stem cells. The mitotic division of these produces two types of cells. Type A cells replenish the stem cells, and type B cells differentiate into primary spermatocytes. The primary spermatocyte divides meiotically (Meiosis I) into two secondary spermatocytes; each secondary spermatocyte divides into two equal haploid spermatids by Meiosis II. The spermatids are transformed into spermatozoa (sperm) by the process of spermiogenesis. These develop into mature spermatozoa, also known as sperm cells. Thus, the primary spermatocyte gives rise to two cells, the secondary spermatocytes, and the two secondary spermatocytes by their subdivision produce four spermatozoa and four haploid cells. Spermatozoa are the mature male gametes in many sexually reproducing organisms. Thus, spermatogenesis is the male version of gametogenesis, of which the female equivalent is oogenesis. In mammals it occurs in the seminiferous tubules of the male testes in a stepwise fashion. Spermatogenesis is highly dependent upon optimal conditions for the process to occur correctly, and is essential for sexual reproduction. DNA methylation and histone modification have been implicated in the regulation of this process. It starts during puberty and usually continues uninterrupted until death, although a slight decrease can be discerned in the quantity of produced sperm with increase in age (see Male infertility). Spermatogenesis starts in the bottom part of seminiferous tubes and, progressively, cells go deeper into tubes and moving along it until mature spermatozoa reaches the lumen, where mature spermatozoa are deposited. The division happens asynchronically; if the tube is cut transversally one could observe different Document 1::: Spermatidogenesis is the creation of spermatids from secondary spermatocytes during spermatogenesis. Secondary spermatocytes produced earlier rapidly enter meiosis II and divide to produce haploid spermatids. The brevity of this stage means that secondary spermatocytes are rarely seen in histological preparations. Mouse stem cells were grown into cells resembling spermatids in 2016. These spermatids, when injected into mouse eggs, were able to produced pups. Document 2::: Sperm (: sperm or sperms) is the male reproductive cell, or gamete, in anisogamous forms of sexual reproduction (forms in which there is a larger, female reproductive cell and a smaller, male one). Animals produce motile sperm with a tail known as a flagellum, which are known as spermatozoa, while some red algae and fungi produce non-motile sperm cells, known as spermatia. Flowering plants contain non-motile sperm inside pollen, while some more basal plants like ferns and some gymnosperms have motile sperm. Sperm cells form during the process known as spermatogenesis, which in amniotes (reptiles and mammals) takes place in the seminiferous tubules of the testes. This process involves the production of several successive sperm cell precursors, starting with spermatogonia, which differentiate into spermatocytes. The spermatocytes then undergo meiosis, reducing their chromosome number by half, which produces spermatids. The spermatids then mature and, in animals, construct a tail, or flagellum, which gives rise to the mature, motile sperm cell. This whole process occurs constantly and takes around 3 months from start to finish. Sperm cells cannot divide and have a limited lifespan, but after fusion with egg cells during fertilization, a new organism begins developing, starting as a totipotent zygote. The human sperm cell is haploid, so that its 23 chromosomes can join the 23 chromosomes of the female egg to form a diploid cell with 46 paired chromosomes. In mammals, sperm is stored in the epididymis and is released from the penis during ejaculation in a fluid known as semen. The word sperm is derived from the Greek word σπέρμα, sperma, meaning "seed". Evolution It is generally accepted that isogamy is the ancestor to sperm and eggs. However, there are no fossil records for the evolution of sperm and eggs from isogamy leading there to be a strong emphasis on mathematical models to understand the evolution of sperm. A widespread hypothesis states that sperm evolve Document 3::: The spermatid is the haploid male gametid that results from division of secondary spermatocytes. As a result of meiosis, each spermatid contains only half of the genetic material present in the original primary spermatocyte. Spermatids are connected by cytoplasmic material and have superfluous cytoplasmic material around their nuclei. When formed, early round spermatids must undergo further maturational events to develop into spermatozoa, a process termed spermiogenesis (also termed spermeteliosis). The spermatids begin to grow a living thread, develop a thickened mid-piece where the mitochondria become localised, and form an acrosome. Spermatid DNA also undergoes packaging, becoming highly condensed. The DNA is packaged firstly with specific nuclear basic proteins, which are subsequently replaced with protamines during spermatid elongation. The resultant tightly packed chromatin is transcriptionally inactive. In 2016 scientists at Nanjing Medical University claimed they had produced cells resembling mouse spermatids artificially from stem cells. They injected these spermatids into mouse eggs and produced pups. DNA repair As postmeiotic germ cells develop to mature sperm they progressively lose the ability to repair DNA damage that may then accumulate and be transmitted to the zygote and ultimately the embryo. In particular, the repair of DNA double-strand breaks by the non-homologous end joining pathway, although present in round spermatids, appears to be lost as they develop into elongated spermatids. Additional images See also List of distinct cell types in the adult human body Document 4::: Gonocytes are the precursors of spermatogonia that differentiate in the testis from primordial germ cells around week 7 of embryonic development and exist up until the postnatal period, when they become spermatogonia. Despite some uses of the term to refer to the precursors of oogonia, it was generally restricted to male germ cells. Germ cells operate as vehicles of inheritance by transferring genetic and epigenetic information from one generation to the next. Male fertility is centered around continual spermatogonia which is dependent upon a high stem cell population. Thus, the function and quality of a differentiated sperm cell is dependent upon the capacity of its originating spermatogonial stem cell (SSC). Gonocytes represent the germ cells undergoing the successive, short-term and migratory stages of development. This occurs between the time they inhabit the forming gonads on the genital ridge to the time they migrate to the basement membrane of the seminiferous cords. Gonocyte development consists of several phases of cell proliferation, differentiation, migration and apoptosis. The abnormal development of gonocytes leads to fertility-related diseases. They are also identified as prespermatogonia, prospermatogonia and primitive germ cells, although gonocyte is most common. History Gonocytes are described as large and spherical, with a prominent nucleus and two nucleoli. The term, gonocyte, was created in 1957 by Canadian scientists Yves Clermont and Bernard Perey. They considered it essential to study the origin of spermatogonia and carried out a study on rats to investigate this. In 1987, Clermont referred to gonocytes as the cells that differentiate into type A spermatogonia, which differentiate into type B spermatogonia and spermatocytes. Very few studies used gonocytes to also refer to the female germ cells in the ovarium primordium. The specification of gonocytes to be confined to male germ cells occurred after foundational differences between the me The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses. While spermatogenesis produces four haploid sperm cells, oogenesis produces one mature what? A. ovum B. egg C. datum D. semen Answer:
sciq-2358
multiple_choice
What type of plant is the dominant group of land plants?
[ "grass plants", "thermal plants", "vascular plants", "oxygen plants" ]
C
Relavent Documents: Document 0::: The following is a list of vascular plants, bryophytes and lichens which are constant species in one or more community of the British National Vegetation Classification system. Vascular plants Grasses Sedges and rushes Trees Other dicotyledons Other monocotyledons Ferns Clubmosses Bryophytes Mosses Liverworts Lichens British National Vegetation Classification Lists of biota of the United Kingdom British National Vegetation Classification, constant Document 1::: Plant ecology is a subdiscipline of ecology that studies the distribution and abundance of plants, the effects of environmental factors upon the abundance of plants, and the interactions among plants and between plants and other organisms. Examples of these are the distribution of temperate deciduous forests in North America, the effects of drought or flooding upon plant survival, and competition among desert plants for water, or effects of herds of grazing animals upon the composition of grasslands. A global overview of the Earth's major vegetation types is provided by O.W. Archibold. He recognizes 11 major vegetation types: tropical forests, tropical savannas, arid regions (deserts), Mediterranean ecosystems, temperate forest ecosystems, temperate grasslands, coniferous forests, tundra (both polar and high mountain), terrestrial wetlands, freshwater ecosystems and coastal/marine systems. This breadth of topics shows the complexity of plant ecology, since it includes plants from floating single-celled algae up to large canopy forming trees. One feature that defines plants is photosynthesis. Photosynthesis is the process of a chemical reactions to create glucose and oxygen, which is vital for plant life. One of the most important aspects of plant ecology is the role plants have played in creating the oxygenated atmosphere of earth, an event that occurred some 2 billion years ago. It can be dated by the deposition of banded iron formations, distinctive sedimentary rocks with large amounts of iron oxide. At the same time, plants began removing carbon dioxide from the atmosphere, thereby initiating the process of controlling Earth's climate. A long term trend of the Earth has been toward increasing oxygen and decreasing carbon dioxide, and many other events in the Earth's history, like the first movement of life onto land, are likely tied to this sequence of events. One of the early classic books on plant ecology was written by J.E. Weaver and F.E. Clements. It Document 2::: Plant functional types (PFTs) refers to a grouping or classification system often used by ecologists and climatologists to classify plant species based on their similar functions and performances in an ecosystem. It is a way to simplify the complexity of plant diversity and behaviour in ecological models by grouping plants into categories that share common functional characteristics. This simplification helps researchers model vegetation dynmaics which can be used in land use studies and climate models. PFTs provide a finer level of modeling than biomes, which represent gross areas such as desert, savannah, deciduous forest. In creating models with PFTs, areas as small as 1 km2 are modeled by defining the predominant plant type for that area, interpreted from satellite data or other means. For each plant functional type, a number of key parameters are defined, such as fecundity, competitiveness, resorption (rate at which plant decays and returns nutrients to the soil after death), etc. The value of each parameter is determined or inferred from observable characteristics such as plant height, leaf area, etc. Plant Functional Type (PFT) models have some limitations and problems. For example, it is difficult for climatologists and ecologists to determine which minimal set of plant characteristics best model the actual responses of the biosphere in response to climate changes. Furthermore, by oversimplifying species to a few key traits, researchers may not capture the full diversity and variability of plant species within a given ecosystem or represent rare or unique species. As such, researchers are developing more sophisticated models, such as trait-based models, to address these problems. See also Ecotone Document 3::: Land cover is the physical material at the surface of Earth. Land covers include grass, asphalt, trees, bare ground, water, etc. Earth cover is the expression used by ecologist Frederick Edward Clements that has its closest modern equivalent being vegetation. The expression continues to be used by the United States Bureau of Land Management. There are two primary methods for capturing information on land cover: field survey, and analysis of remotely sensed imagery. Land change models can be built from these types of data to assess changes in land cover over time. One of the major land cover issues (as with all natural resource inventories) is that every survey defines similarly named categories in different ways. For instance, there are many definitions of "forest"—sometimes within the same organisation—that may or may not incorporate a number of different forest features (e.g., stand height, canopy cover, strip width, inclusion of grasses, and rates of growth for timber production). Areas without trees may be classified as forest cover "if the intention is to re-plant" (UK and Ireland), while areas with many trees may not be labelled as forest "if the trees are not growing fast enough" (Norway and Finland). Distinction from "land use" "Land cover" is distinct from "land use", despite the two terms often being used interchangeably. Land use is a description of how people utilize the land and of socio-economic activity. Urban and agricultural land uses are two of the most commonly known land use classes. At any one point or place, there may be multiple and alternate land uses, the specification of which may have a political dimension. The origins of the "land cover/land use" couplet and the implications of their confusion are discussed in Fisher et al. (2005). Types Following table is Land Cover statistics by Food and Agriculture Organization (FAO) with 14 classes. Mapping Land cover change detection using remote sensing and geospatial data provides baselin Document 4::: Ecological classification or ecological typology is the classification of land or water into geographical units that represent variation in one or more ecological features. Traditional approaches focus on geology, topography, biogeography, soils, vegetation, climate conditions, living species, habitats, water resources, and sometimes also anthropic factors. Most approaches pursue the cartographical delineation or regionalisation of distinct areas for mapping and planning. Approaches to classifications Different approaches to ecological classifications have been developed in terrestrial, freshwater and marine disciplines. Traditionally these approaches have focused on biotic components (vegetation classification), abiotic components (environmental approaches) or implied ecological and evolutionary processes (biogeographical approaches). Ecosystem classifications are specific kinds of ecological classifications that consider all four elements of the definition of ecosystems: a biotic component, an abiotic complex, the interactions between and within them, and the physical space they occupy (ecotope). Vegetation classification Vegetation is often used to classify terrestrial ecological units. Vegetation classification can be based on vegetation structure and floristic composition. Classifications based entirely on vegetation structure overlap with land cover mapping categories. Many schemes of vegetation classification are in use by the land, resource and environmental management agencies of different national and state jurisdictions. The International Vegetation Classification (IVC or EcoVeg) has been recently proposed but has not been yet widely adopted. Vegetation classifications have limited use in aquatic systems, since only a handful of freshwater or marine habitats are dominated by plants (e.g. kelp forests or seagrass meadows). Also, some extreme terrestrial environments, like subterranean or cryogenic ecosystems, are not properly described in vegetation c The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses. What type of plant is the dominant group of land plants? A. grass plants B. thermal plants C. vascular plants D. oxygen plants Answer:
sciq-3988
multiple_choice
What is composed of a calcium cation and a nitrate anion?
[ "dynamite", "calcium carbonate", "nitrous oxide", "calcium nitrate" ]
D
Relavent Documents: Document 0::: The nitrite ion has the chemical formula . Nitrite (mostly sodium nitrite) is widely used throughout chemical and pharmaceutical industries. The nitrite anion is a pervasive intermediate in the nitrogen cycle in nature. The name nitrite also refers to organic compounds having the –ONO group, which are esters of nitrous acid. Production Sodium nitrite is made industrially by passing a mixture of nitrogen oxides into aqueous sodium hydroxide or sodium carbonate solution: The product is purified by recrystallization. Alkali metal nitrites are thermally stable up to and beyond their melting point (441 °C for KNO2). Ammonium nitrite can be made from dinitrogen trioxide, N2O3, which is formally the anhydride of nitrous acid: 2 NH3 + H2O + N2O3 → 2 NH4NO2 Structure The nitrite ion has a symmetrical structure (C2v symmetry), with both N–O bonds having equal length and a bond angle of about 115°. In valence bond theory, it is described as a resonance hybrid with equal contributions from two canonical forms that are mirror images of each other. In molecular orbital theory, there is a sigma bond between each oxygen atom and the nitrogen atom, and a delocalized pi bond made from the p orbitals on nitrogen and oxygen atoms which is perpendicular to the plane of the molecule. The negative charge of the ion is equally distributed on the two oxygen atoms. Both nitrogen and oxygen atoms carry a lone pair of electrons. Therefore, the nitrite ion is a Lewis base. In the gas phase it exists predominantly as a trans-planar molecule. Reactions Acid-base properties Nitrite is the conjugate base of the weak acid nitrous acid: HNO2 H+ + ; pKa ≈ 3.3 at 18 °C Nitrous acid is also highly volatile, tending to disproportionate: 3 HNO2 (aq) H3O+ + + 2 NO This reaction is slow at 0 °C. Addition of acid to a solution of a nitrite in the presence of a reducing agent, such as iron(II), is a way to make nitric oxide (NO) in the laboratory. Oxidation and reduction The formal oxidation sta Document 1::: With Sn2+ ions, N2O is formed: 2 HNO2 + 6 HCl + 2 SnCl2 → 2 SnCl4 + N2O + 3 H2O + 2 KCl With SO2 gas, NH2OH is formed: 2 HNO2 + 6 H2O + 4 SO2 → 3 H2SO4 + K2SO4 + 2 NH2OH With Zn in alkali solution, NH3 is formed: 5 H2O + KNO2 + 3 Zn → NH3 + KOH + 3 Zn(OH)2 With , both HN3 Document 2::: Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas. Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below: During adiabatic expansion of an ideal gas, its temperatureincreases decreases stays the same Impossible to tell/need more information The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well. Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in Document 3::: {|class="wikitable" |- ! colspan="2" | Common monatomic anions |- | hydride | H− |- | fluoride | F− |- | chloride | Cl− |- | bromide | Br− |- | iodide | I− |- | oxide | O2− |- | sulfide | S2− |- | nitride | N3− |- | phosphide | P3− |- |} Document 4::: [()2]x+ [(Xn−)x/n · y]x-, where Xn− is the intercalating anion (or anions). Most commonly, = Ca2+, Mg2+, Mn2+, Fe2+, Co2+, Ni2+, Cu2+ or Zn2+, and is another trivalent cation, possibly of the same element. Fixed-composition phases have been shown to exist over the rang The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses. What is composed of a calcium cation and a nitrate anion? A. dynamite B. calcium carbonate C. nitrous oxide D. calcium nitrate Answer:
sciq-6576
multiple_choice
Embryos usually don't survive if they have extra or too few of these genetic units?
[ "genes", "phenotypes", "ribosomes", "chromosomes" ]
D
Relavent Documents: Document 0::: Embryomics is the identification, characterization and study of the diverse cell types which arise during embryogenesis, especially as this relates to the location and developmental history of cells in the embryo. Cell type may be determined according to several criteria: location in the developing embryo, gene expression as indicated by protein and nucleic acid markers and surface antigens, and also position on the embryogenic tree. Embryome There are many cell markers useful in distinguishing, classifying, separating and purifying the numerous cell types present at any given time in a developing organism. These cell markers consist of select RNAs and proteins present inside, and surface antigens present on the surface of, the cells making up the embryo. For any given cell type, these RNA and protein markers reflect the genes characteristically active in that cell type. The catalog of all these cell types and their characteristic markers is known as the organism's embryome. The word is a portmanteau of embryo and genome. “Embryome” may also refer to the totality of the physical cell markers themselves. Embryogenesis As an embryo develops from a fertilized egg, the single egg cell splits into many cells, which grow in number and migrate to the appropriate locations inside the embryo at appropriate times during development. As the embryo's cells grow in number and migrate, they also differentiate into an increasing number of different cell types, ultimately turning into the stable, specialized cell types characteristic of the adult organism. Each of the cells in an embryo contains the same genome, characteristic of the species, but the level of activity of each of the many thousands of genes that make up the complete genome varies with, and determines, a particular cell's type (e.g. neuron, bone cell, skin cell, muscle cell, etc.). During embryo development (embryogenesis), many cell types are present which are not present in the adult organism. These temporary c Document 1::: Dysgenesis is an abnormal organ development during embryonic growth and development. As opposed to agenesis, which refers to the complete failure of an organ to develop, dysgenesis usually implies disordered development or malformation and in some cases represents the milder end of a spectrum of abnormalities. Dysgenesis occurs during fetal development immediately after conception. Classification One of the first organs that is affected is the brain, this is known as cerebral dysgenesis. Dysplasia is a form of dysgenesis in adults that alters the size and shape of their cells that lead to abnormal development. One of the most common forms of dysgenesis is within the gonads. Examples: Gonadal dysgenesis Adrenal dysgenesis Thyroid dysgenesis Anterior segment dysgenesis Document 2::: In the field of developmental biology, regional differentiation is the process by which different areas are identified in the development of the early embryo. The process by which the cells become specified differs between organisms. Cell fate determination In terms of developmental commitment, a cell can either be specified or it can be determined. Specification is the first stage in differentiation. A cell that is specified can have its commitment reversed while the determined state is irreversible. There are two main types of specification: autonomous and conditional. A cell specified autonomously will develop into a specific fate based upon cytoplasmic determinants with no regard to the environment the cell is in. A cell specified conditionally will develop into a specific fate based upon other surrounding cells or morphogen gradients. Another type of specification is syncytial specification, characteristic of most insect classes. Specification in sea urchins uses both autonomous and conditional mechanisms to determine the anterior/posterior axis. The anterior/posterior axis lies along the animal/vegetal axis set up during cleavage. The micromeres induce the nearby tissue to become endoderm while the animal cells are specified to become ectoderm. The animal cells are not determined because the micromeres can induce the animal cells to also take on mesodermal and endodermal fates. It was observed that β-catenin was present in the nuclei at the vegetal pole of the blastula. Through a series of experiments, one study confirmed the role of β-catenin in the cell-autonomous specification of vegetal cell fates and the micromeres inducing ability. Treatments of lithium chloride sufficient to vegetalize the embryo resulted in increases in nuclearly localized b-catenin. Reduction of expression of β-catenin in the nucleus correlated with loss of vegetal cell fates. Transplants of micromeres lacking nuclear accumulation of β-catenin were unable to induce a second axis. Document 3::: According to the principle of nuclear equivalence, the nuclei of essentially all differentiated adult cells of an individual are genetically (though not necessarily metabolically) identical to one another and to the nucleus of the zygote from which they descended. This means that virtually all somatic cells in an adult have the same genes. However, different cells express different subsets of these genes. The evidence for nuclear equivalence comes from cases in which differentiated cells or their nuclei have been found to retain the potential of directing the development of the entire organism. Such cells or nuclei are said to exhibit totipotency. Document 4::: An organism () is any biological living system that functions as an individual life form. All organisms are composed of cells. The idea of organism is based on the concept of minimal functional unit of life. Three traits have been proposed to play the main role in qualification as an organism: noncompartmentability – structure that cannot be divided without its functionality loss, individuality – the entity has simultaneous holding of genetic uniqueness, genetic homogeneity and autonomy, distinctness – genetic information has to maintain open-system (a cell). Organisms include multicellular animals, plants, and fungi; or unicellular microorganisms such as protists, bacteria, and archaea. All types of organisms are capable of reproduction, growth and development, maintenance, and some degree of response to stimuli. Most multicellular organisms differentiate into specialized tissues and organs during their development. In 2016, a set of 355 genes from the last universal common ancestor (LUCA) of all organisms from Earth was identified. Etymology The term "organism" (from Greek ὀργανισμός, organismos, from ὄργανον, organon, i.e. "instrument, implement, tool, organ of sense or apprehension") first appeared in the English language in 1703 and took on its current definition by 1834 (Oxford English Dictionary). It is directly related to the term "organization". There is a long tradition of defining organisms as self-organizing beings, going back at least to Immanuel Kant's 1790 Critique of Judgment. Definitions An organism may be defined as an assembly of molecules functioning as a more or less stable whole that exhibits the properties of life. Dictionary definitions can be broad, using phrases such as "any living structure, such as a plant, animal, fungus or bacterium, capable of growth and reproduction". Many definitions exclude viruses and possible synthetic non-organic life forms, as viruses are dependent on the biochemical machinery of a host cell for repr The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses. Embryos usually don't survive if they have extra or too few of these genetic units? A. genes B. phenotypes C. ribosomes D. chromosomes Answer:
sciq-7141
multiple_choice
What is the term for individuals of a species that live in a particular habitat?
[ "family", "population", "biosphere", "ecosystem" ]
B
Relavent Documents: Document 0::: In ecology, habitat refers to the array of resources, physical and biotic factors that are present in an area, such as to support the survival and reproduction of a particular species. A species habitat can be seen as the physical manifestation of its ecological niche. Thus "habitat" is a species-specific term, fundamentally different from concepts such as environment or vegetation assemblages, for which the term "habitat-type" is more appropriate. The physical factors may include (for example): soil, moisture, range of temperature, and light intensity. Biotic factors include the availability of food and the presence or absence of predators. Every species has particular habitat requirements, with habitat generalist species able to thrive in a wide array of environmental conditions while habitat specialist species requiring a very limited set of factors to survive. The habitat of a species is not necessarily found in a geographical area, it can be the interior of a stem, a rotten log, a rock or a clump of moss; a parasitic organism has as its habitat the body of its host, part of the host's body (such as the digestive tract), or a single cell within the host's body. Habitat types are environmental categorizations of different environments based on the characteristics of a given geographical area, particularly vegetation and climate. Thus habitat types do not refer to a single species but to multiple species living in the same area. For example, terrestrial habitat types include forest, steppe, grassland, semi-arid or desert. Fresh-water habitat types include marshes, streams, rivers, lakes, and ponds; marine habitat types include salt marshes, the coast, the intertidal zone, estuaries, reefs, bays, the open sea, the sea bed, deep water and submarine vents. Habitat types may change over time. Causes of change may include a violent event (such as the eruption of a volcano, an earthquake, a tsunami, a wildfire or a change in oceanic currents); or change may occur mo Document 1::: This glossary of ecology is a list of definitions of terms and concepts in ecology and related fields. For more specific definitions from other glossaries related to ecology, see Glossary of biology, Glossary of evolutionary biology, and Glossary of environmental science. A B C D E F G H I J K L M N O P Q R S T U V W X Y Z See also Outline of ecology History of ecology Document 2::: Ecological units, comprise concepts such as population, community, and ecosystem as the basic units, which are at the basis of ecological theory and research, as well as a focus point of many conservation strategies. The concept of ecological units continues to suffer from inconsistencies and confusion over its terminology. Analyses of the existing concepts used in describing ecological units have determined that they differ in respects to four major criteria: The questions as to whether they are defined statistically or via a network of interactions, If their boundaries are drawn by topographical or process-related criteria, How high the required internal relationships are, And if they are perceived as "real" entities or abstractions by an observer. A population is considered to be the smallest ecological unit, consisting of a group of individuals that belong to the same species. A community would be the next classification, referring to all of the population present in an area at a specific time, followed by an ecosystem, referring to the community and it's interactions with its physical environment. An ecosystem is the most commonly used ecological unit and can be universally defined by two common traits: The unit is often defined in terms of a natural border (maritime boundary, watersheds, etc.) Abiotic components and organisms within the unit are considered to be interlinked. See also Biogeographic realm Ecoregion Ecotope Holobiont Functional ecology Behavior settings Regional geology Document 3::: Population ecology is a sub-field of ecology that deals with the dynamics of species populations and how these populations interact with the environment, such as birth and death rates, and by immigration and emigration. The discipline is important in conservation biology, especially in the development of population viability analysis which makes it possible to predict the long-term probability of a species persisting in a given patch of habitat. Although population ecology is a subfield of biology, it provides interesting problems for mathematicians and statisticians who work in population dynamics. History In the 1940s ecology was divided into autecology—the study of individual species in relation to the environment—and synecology—the study of groups of species in relation to the environment. The term autecology (from Ancient Greek: αὐτο, aúto, "self"; οίκος, oíkos, "household"; and λόγος, lógos, "knowledge"), refers to roughly the same field of study as concepts such as life cycles and behaviour as adaptations to the environment by individual organisms. Eugene Odum, writing in 1953, considered that synecology should be divided into population ecology, community ecology and ecosystem ecology, renaming autecology as 'species ecology' (Odum regarded "autecology" as an archaic term), thus that there were four subdivisions of ecology. Terminology A population is defined as a group of interacting organisms of the same species. A demographic structure of a population is how populations are often quantified. The total number of individuals in a population is defined as a population size, and how dense these individuals are is defined as population density. There is also a population’s geographic range, which has limits that a species can tolerate (such as temperature). Population size can be influenced by the per capita population growth rate (rate at which the population size changes per individual in the population.) Births, deaths, emigration, and immigration rates Document 4::: A biome () is a biogeographical unit consisting of a biological community that has formed in response to the physical environment in which they are found and a shared regional climate. Biomes may span more than one continent. Biome is a broader term than habitat and can comprise a variety of habitats. While a biome can cover small areas, a microbiome is a mix of organisms that coexist in a defined space on a much smaller scale. For example, the human microbiome is the collection of bacteria, viruses, and other microorganisms that are present on or in a human body. A biota is the total collection of organisms of a geographic region or a time period, from local geographic scales and instantaneous temporal scales all the way up to whole-planet and whole-timescale spatiotemporal scales. The biotas of the Earth make up the biosphere. Etymology The term was suggested in 1916 by Clements, originally as a synonym for biotic community of Möbius (1877). Later, it gained its current definition, based on earlier concepts of phytophysiognomy, formation and vegetation (used in opposition to flora), with the inclusion of the animal element and the exclusion of the taxonomic element of species composition. In 1935, Tansley added the climatic and soil aspects to the idea, calling it ecosystem. The International Biological Program (1964–74) projects popularized the concept of biome. However, in some contexts, the term biome is used in a different manner. In German literature, particularly in the Walter terminology, the term is used similarly as biotope (a concrete geographical unit), while the biome definition used in this article is used as an international, non-regional, terminology—irrespectively of the continent in which an area is present, it takes the same biome name—and corresponds to his "zonobiome", "orobiome" and "pedobiome" (biomes determined by climate zone, altitude or soil). In Brazilian literature, the term "biome" is sometimes used as synonym of biogeographic pr The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses. What is the term for individuals of a species that live in a particular habitat? A. family B. population C. biosphere D. ecosystem Answer:
sciq-10039
multiple_choice
What is formed when different gases are confined to the same container?
[ "linear mixture", "homogeneous mixture", "particulate mixture", "compact mixture" ]
B
Relavent Documents: Document 0::: Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas. Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below: During adiabatic expansion of an ideal gas, its temperatureincreases decreases stays the same Impossible to tell/need more information The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well. Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in Document 1::: The Gas composition of any gas can be characterised by listing the pure substances it contains, and stating for each substance its proportion of the gas mixture's molecule count.Nitrogen 78.084 Oxygen 20.9476 Argon Ar 0.934 Carbon Dioxide 0.0314 Gas composition of air To give a familiar example, air has a composition of: Standard Dry Air is the agreed-upon gas composition for air from which all water vapour has been removed. There are various standards bodies which publish documents that define a dry air gas composition. Each standard provides a list of constituent concentrations, a gas density at standard conditions and a molar mass. It is extremely unlikely that the actual composition of any specific sample of air will completely agree with any definition for standard dry air. While the various definitions for standard dry air all attempt to provide realistic information about the constituents of air, the definitions are important in and of themselves because they establish a standard which can be cited in legal contracts and publications documenting measurement calculation methodologies or equations of state. The standards below are two examples of commonly used and cited publications that provide a composition for standard dry air: ISO TR 29922-2017 provides a definition for standard dry air which specifies an air molar mass of 28,965 46 ± 0,000 17 kg·kmol-1. GPA 2145:2009 is published by the Gas Processors Association. It provides a molar mass for air of 28.9625 g/mol, and provides a composition for standard dry air as a footnote. Document 2::: In physics, a dynamical system is said to be mixing if the phase space of the system becomes strongly intertwined, according to at least one of several mathematical definitions. For example, a measure-preserving transformation T is said to be strong mixing if whenever A and B are any measurable sets and μ is the associated measure. Other definitions are possible, including weak mixing and topological mixing. The mathematical definition of mixing is meant to capture the notion of physical mixing. A canonical example is the Cuba libre: suppose one is adding rum (the set A) to a glass of cola. After stirring the glass, the bottom half of the glass (the set B) will contain rum, and it will be in equal proportion as it is elsewhere in the glass. The mixing is uniform: no matter which region B one looks at, some of A will be in that region. A far more detailed, but still informal description of mixing can be found in the article on mixing (mathematics). Every mixing transformation is ergodic, but there are ergodic transformations which are not mixing. Physical mixing The mixing of gases or liquids is a complex physical process, governed by a convective diffusion equation that may involve non-Fickian diffusion as in spinodal decomposition. The convective portion of the governing equation contains fluid motion terms that are governed by the Navier–Stokes equations. When fluid properties such as viscosity depend on composition, the governing equations may be coupled. There may also be temperature effects. It is not clear that fluid mixing processes are mixing in the mathematical sense. Small rigid objects (such as rocks) are sometimes mixed in a rotating drum or tumbler. The 1969 Selective Service draft lottery was carried out by mixing plastic capsules which contained a slip of paper (marked with a day of the year). See also Miscibility Document 3::: In chemistry, a mixture is a material made up of two or more different chemical substances which are not chemically bonded. A mixture is the physical combination of two or more substances in which the identities are retained and are mixed in the form of solutions, suspensions and colloids. Mixtures are one product of mechanically blending or mixing chemical substances such as elements and compounds, without chemical bonding or other chemical change, so that each ingredient substance retains its own chemical properties and makeup. Despite the fact that there are no chemical changes to its constituents, the physical properties of a mixture, such as its melting point, may differ from those of the components. Some mixtures can be separated into their components by using physical (mechanical or thermal) means. Azeotropes are one kind of mixture that usually poses considerable difficulties regarding the separation processes required to obtain their constituents (physical or chemical processes or, even a blend of them). Characteristics of mixtures All mixtures can be characterized as being separable by mechanical means (e.g. purification, distillation, electrolysis, chromatography, heat, filtration, gravitational sorting, centrifugation). Mixtures differ from chemical compounds in the following ways: the substances in a mixture can be separated using physical methods such as filtration, freezing, and distillation. there is little or no energy change when a mixture forms (see Enthalpy of mixing). The substances in a mixture keep its separate properties. In the example of sand and water, neither one of the two substances changed in any way when they are mixed. Although the sand is in the water it still keeps the same properties that it had when it was outside the water. mixtures have variable compositions, while compounds have a fixed, definite formula. when mixed, individual substances keep their properties in a mixture, while if they form a compound their properties Document 4::: In mathematical psychology and education theory, a knowledge space is a combinatorial structure used to formulate mathematical models describing the progression of a human learner. Knowledge spaces were introduced in 1985 by Jean-Paul Doignon and Jean-Claude Falmagne, and remain in extensive use in the education theory. Modern applications include two computerized tutoring systems, ALEKS and the defunct RATH. Formally, a knowledge space assumes that a domain of knowledge is a collection of concepts or skills, each of which must be eventually mastered. Not all concepts are interchangeable; some require other concepts as prerequisites. Conversely, competency at one skill may ease the acquisition of another through similarity. A knowledge space marks out which collections of skills are feasible: they can be learned without mastering any other skills. Under reasonable assumptions, the collection of feasible competencies forms the mathematical structure known as an antimatroid. Researchers and educators usually explore the structure of a discipline's knowledge space as a latent class model. Motivation Knowledge Space Theory attempts to address shortcomings of standardized testing when used in educational psychometry. Common tests, such as the SAT and ACT, compress a student's knowledge into a very small range of ordinal ranks, in the process effacing the conceptual dependencies between questions. Consequently, the tests cannot distinguish between true understanding and guesses, nor can they identify a student's particular weaknesses, only the general proportion of skills mastered. The goal of knowledge space theory is to provide a language by which exams can communicate What the student can do and What the student is ready to learn. Model structure Knowledge Space Theory-based models presume that an educational subject can be modeled as a finite set of concepts, skills, or topics. Each feasible state of knowledge about is then a subset of ; the set of The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses. What is formed when different gases are confined to the same container? A. linear mixture B. homogeneous mixture C. particulate mixture D. compact mixture Answer:
sciq-7763
multiple_choice
What type of resource is limited in supply and cannot be replaced except over millions of years?
[ "capital resources", "human resources", "natural resource", "renewable energy" ]
C
Relavent Documents: Document 0::: Resource refers to all the materials available in our environment which are technologically accessible, economically feasible and culturally sustainable and help us to satisfy our needs and wants. Resources can broadly be classified upon their availability — they are classified into renewable and non-renewable resources. They can also be classified as actual and potential on the basis of the level of development and use, on the basis of origin they can be classified as biotic and abiotic, and on the basis of their distribution, as ubiquitous and localised (private, community-owned, national and international resources). An item becomes a resource with time and developing technology. The benefits of resource utilization may include increased wealth, proper functioning of a system, or enhanced well-being. From a human perspective, a natural resource is anything obtained from the environment to satisfy human needs and wants. From a broader biological or ecological perspective, a resource satisfies the needs of a living organism (see biological resource). The concept of resources has been developed across many established areas of work, in economics, biology and ecology, computer science, management, and human resources for example - linked to the concepts of competition, sustainability, conservation, and stewardship. In application within human society, commercial or non-commercial factors require resource allocation through resource management. The concept of a resource can also be tied to the direction of leadership over resources, this can include the things leaders have responsibility for over the human resources, with management, help, support or direction such as in charge of a professional group, technical experts, innovative leaders, archiving expertise, academic management, association management, business management, healthcare management, military management, public administration, spiritual leadership and social networking administrator. individuals exp Document 1::: A non-renewable resource (also called a finite resource) is a natural resource that cannot be readily replaced by natural means at a pace quick enough to keep up with consumption. An example is carbon-based fossil fuels. The original organic matter, with the aid of heat and pressure, becomes a fuel such as oil or gas. Earth minerals and metal ores, fossil fuels (coal, petroleum, natural gas) and groundwater in certain aquifers are all considered non-renewable resources, though individual elements are always conserved (except in nuclear reactions, nuclear decay or atmospheric escape). Conversely, resources such as timber (when harvested sustainably) and wind (used to power energy conversion systems) are considered renewable resources, largely because their localized replenishment can occur within time frames meaningful to humans as well. Earth minerals and metal ores Earth minerals and metal ores are examples of non-renewable resources. The metals themselves are present in vast amounts in Earth's crust, and their extraction by humans only occurs where they are concentrated by natural geological processes (such as heat, pressure, organic activity, weathering and other processes) enough to become economically viable to extract. These processes generally take from tens of thousands to millions of years, through plate tectonics, tectonic subsidence and crustal recycling. The localized deposits of metal ores near the surface which can be extracted economically by humans are non-renewable in human time-frames. There are certain rare earth minerals and elements that are more scarce and exhaustible than others. These are in high demand in manufacturing, particularly for the electronics industry. Fossil fuels Natural resources such as coal, petroleum(crude oil) and natural gas take thousands of years to form naturally and cannot be replaced as fast as they are being consumed. Eventually it is considered that fossil-based resources will become too costly to harvest and Document 2::: The Ultimate Resource is a 1981 book written by Julian Lincoln Simon challenging the notion that humanity was running out of natural resources. It was revised in 1996 as The Ultimate Resource 2. Overview The overarching thesis on why there is no resource crisis is that as a particular resource becomes more scarce, its price rises. This price rise creates an incentive for people to discover more of the resource, ration and recycle it, and eventually, develop substitutes. The "ultimate resource" is not any particular physical object but the capacity for humans to invent and adapt. Scarcity The work opens with an explanation of scarcity, noting its relation to price; high prices denote relative scarcity and low prices indicate abundance. Simon usually measures prices in wage-adjusted terms, since this is a measure of how much labor is required to purchase a fixed amount of a particular resource. Since prices for most raw materials (e.g., copper) have fallen between 1800 and 1990 (adjusting for wages and adjusting for inflation), Simon argues that this indicates that those materials have become less scarce. Forecasting Simon makes a distinction between "engineering” and "economic" forecasting. Engineering forecasting consists of estimating the amount of known physical amount of resources, extrapolates the rate of use from current use and subtracts one from the other. Simon argues that these simple analyses are often wrong. While focusing only on proven resources is helpful in a business context, it is not appropriate for economy-wide forecasting. There exist undiscovered sources, sources not yet economically feasible to extract, sources not yet technologically feasible to extract, and ignored resources that could prove useful but are not yet worth trying to discover. To counter the problems of engineering forecasting, Simon proposes economic forecasting, which proceeds in three steps in order to capture, in part, the unknowns the engineering method leaves out (p 27) Document 3::: Energy quality is a measure of the ease with which a form of energy can be converted to useful work or to another form of energy: i.e. its content of thermodynamic free energy. A high quality form of energy has a high content of thermodynamic free energy, and therefore a high proportion of it can be converted to work; whereas with low quality forms of energy, only a small proportion can be converted to work, and the remainder is dissipated as heat. The concept of energy quality is also used in ecology, where it is used to track the flow of energy between different trophic levels in a food chain and in thermoeconomics, where it is used as a measure of economic output per unit of energy. Methods of evaluating energy quality often involve developing a ranking of energy qualities in hierarchical order. Examples: Industrialization, Biology The consideration of energy quality was a fundamental driver of industrialization from the 18th through 20th centuries. Consider for example the industrialization of New England in the 18th century. This refers to the construction of textile mills containing power looms for weaving cloth. The simplest, most economical and straightforward source of energy was provided by water wheels, extracting energy from a millpond behind a dam on a local creek. If another nearby landowner also decided to build a mill on the same creek, the construction of their dam would lower the overall hydraulic head to power the existing waterwheel, thus hurting power generation and efficiency. This eventually became an issue endemic to the entire region, reducing the overall profitability of older mills as newer ones were built. The search for higher quality energy was a major impetus throughout the 19th and 20th centuries. For example, burning coal to make steam to generate mechanical energy would not have been imaginable in the 18th century; by the end of the 19th century, the use of water wheels was long outmoded. Similarly, the quality of energy from elec Document 4::: Resource consumption is about the consumption of non-renewable, or less often, renewable resources. Specifically, it may refer to: water consumption energy consumption electric energy consumption world energy consumption natural gas consumption/gas depletion oil consumption/oil depletion logging/deforestation fishing/overfishing land use/land loss or resource depletion and general exploitation and associated environmental degradation Measures of resource consumption are resource intensity and resource efficiency. Industrialization and globalized markets have increased the tendency for overconsumption of resources. The resource consumption rate of a nation does not usually correspond with the primary resource availability, this is called resource curse. Unsustainable consumption by the steadily growing human population may lead to resource depletion and a shrinking of the earth's carrying capacity. See also Ecological footprint Jevons paradox Natural resource management Scarcity Uneconomic growth The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses. What type of resource is limited in supply and cannot be replaced except over millions of years? A. capital resources B. human resources C. natural resource D. renewable energy Answer:
sciq-2071
multiple_choice
What kind of air does a continental polar air mass hold?
[ "warm dry air", "cold dry air", "frigid wet air", "cool moist air" ]
B
Relavent Documents: Document 0::: The Hadley cell, also known as the Hadley circulation, is a global-scale tropical atmospheric circulation that features air rising near the equator, flowing poleward near the tropopause at a height of above the Earth's surface, cooling and descending in the subtropics at around 25 degrees latitude, and then returning equatorward near the surface. It is a thermally direct circulation within the troposphere that emerges due to differences in insolation and heating between the tropics and the subtropics. On a yearly average, the circulation is characterized by a circulation cell on each side of the equator. The Southern Hemisphere Hadley cell is slightly stronger on average than its northern counterpart, extending slightly beyond the equator into the Northern Hemisphere. During the summer and winter months, the Hadley circulation is dominated by a single, cross-equatorial cell with air rising in the summer hemisphere and sinking in the winter hemisphere. Analogous circulations may occur in extraterrestrial atmospheres, such as on Venus and Mars. Global climate is greatly influenced by the structure and behavior of the Hadley circulation. The prevailing trade winds are a manifestation of the lower branches of the Hadley circulation, converging air and moisture in the tropics to form the Intertropical Convergence Zone (ITCZ) where the Earth's heaviest rains are located. Shifts in the ITCZ associated with the seasonal variability of the Hadley circulation cause monsoons. The sinking branches of the Hadley cells give rise to the oceanic subtropical ridges and suppress rainfall; many of the Earth's deserts and arid regions are located in the subtropics coincident with the position of the sinking branches. The Hadley circulation is also a key mechanism for the meridional transport of heat, angular momentum, and moisture, contributing to the subtropical jet stream, the moist tropics, and maintaining a global thermal equilibrium. The Hadley circulation is named after George Document 1::: The Mars Regional Atmospheric Modeling System (MRAMS) is a computer program that simulates the circulations of the Martian atmosphere at regional and local scales. MRAMS, developed by Scot Rafkin and Timothy Michaels, is derived from the Regional Atmospheric Modeling System (RAMS) developed by William R. Cotton and Roger A. Pielke to study atmospheric circulations on the Earth. Key features of MRAMS include a non-hydrostatic, fully compressible dynamics, explicit bin dust, water, and carbon dioxide ice atmospheric physics model, and a fully prognostic regolith model that includes carbon dioxide deposition and sublimation. Several Mars exploration projects, including the Mars Exploration Rovers, the Phoenix Scout Mission, and the Mars Science Laboratory have used MRAMS to study a variety of atmospheric circulations. The MRAMS operates at the mesoscale and microscale, modeling and simulating the Martian atmosphere. The smaller scale modeling of the MRAMS gives it higher resolution data points and models over complex terrain and topography. It is able to identify topography driven flows like katabatic and anabatic winds through valleys and mountains that produce changes in atmospheric circulation. Structure Dynamic Core The dynamic core's role is to solve fluid mechanic equations related to atmospheric dynamics. The equations in the dynamic core of the MRAMS are based on primitive grid-volume Reynolds-averaged equations. The related equations are meant to solve for momentum, thermodynamics, tracers, and conservation of mass. The MRAMS dynamical core integrates equations for momentum, thermodynamics (atmosphere-surface heat exchange), tracers, and conservation of mass. Parameterizations The MRAMS dynamical core was developed from RAMS and has been changed excessively to account for the large difference in atmospheres between Mars and Earth. Some MRAMS models parameterize numerous features including dust and dust lifting, cloud microphysics, radiative transfer, an Document 2::: Upper-atmospheric models are simulations of the Earth's atmosphere between 20 and 100 km (65,000 and 328,000 feet) that comprises the stratosphere, mesosphere, and the lower thermosphere. Whereas most climate models simulate a region of the Earth's atmosphere from the surface to the stratopause, there also exist numerical models which simulate the wind, temperature and composition of the Earth's tenuous upper atmosphere, from the mesosphere to the exosphere, including the ionosphere. This region is affected strongly by the 11 year Solar cycle through variations in solar UV/EUV/Xray radiation and solar wind leading to high latitude particle precipitation and aurora. It has been proposed that these phenomena may have an effect on the lower atmosphere, and should therefore be included in simulations of climate change. For this reason there has been a drive in recent years to create whole atmosphere models to investigate whether or not this is the case. Jet stream perturbation model A jet stream perturbation model is employed by Weather Logistics UK, which simulates the diversion of the air streams in the upper atmosphere. North Atlantic air flow modelling is simulated by combining a monthly jet stream climatology input calculated at 20 to 30°W, with different blocking high patterns. The jet stream input is generated by thermal wind balance calculations at 316mbars (6 to 9 km aloft) in the mid-latitude range from 40 to 60°N. Long term blocking patterns are determined by the weather forecaster, who identifies the likely position and strength of North Atlantic Highs from synoptic charts, the North Atlantic Oscillation (NAO) and El Niño-Southern Oscillation (ENSO) patterns. The model is based on the knowledge that low pressure systems at the surface are steered by the fast ribbons (jet streams) of air in the upper atmosphere. The jet stream - blocking interaction model simulation examines the sea surface temperature field using data from NOAA tracked along the ocean on a Document 3::: Polar ecology is the relationship between plants and animals in a polar environment. Polar environments are in the Arctic and Antarctic regions. Arctic regions are in the Northern Hemisphere, and it contains land and the islands that surrounds it. Antarctica is in the Southern Hemisphere and it also contains the land mass, surrounding islands and the ocean. Polar regions also contain the subantarctic and subarctic zone which separate the polar regions from the temperate regions. Antarctica and the Arctic lie in the polar circles. The polar circles are imaginary lines shown on maps to be the areas that receives less sunlight due to less radiation. These areas either receive sunlight (midnight sun) or shade (polar night) 24 hours a day because of the earth's tilt. Plants and animals in the polar regions are able to withstand living in harsh weather conditions but are facing environmental threats that limit their survival. Climate Polar climates are cold, windy and dry. Because of the lack of precipitation and low temperatures the Arctic and Antarctic are considered the world's largest deserts or Polar deserts. Much of the radiation from the sun that is received is reflected off the snow making the polar regions cold. When the radiation is reflected, the heat is also reflected. The polar regions reflect 89-90% of the sun radiation that the earth receives. And because Antarctica is closer to the sun at perihelion, it receives 7% more radiation than the Arctic. Also in the polar region, the atmosphere is thin. Because of this the UV radiation that gets to the atmosphere can cause fast sun tanning and snow blindness. Polar regions are dry areas; there is very little precipitation due to the cold air. There are some times when the humidity may be high but the water vapor present in the air may be low. Wind is also strong in the polar region. Wind carries snow creating blizzard like conditions. Winds may also move small organisms or vegetation if it is present. The wind Document 4::: A reference atmospheric model describes how the ideal gas properties (namely: pressure, temperature, density, and molecular weight) of an atmosphere change, primarily as a function of altitude, and sometimes also as a function of latitude, day of year, etc. A static atmospheric model has a more limited domain, excluding time. A standard atmosphere is defined by the World Meteorological Organization as "a hypothetical vertical distribution of atmospheric temperature, pressure and density which, by international agreement, is roughly representative of year-round, midlatitude conditions." Typical usages are as a basis for pressure altimeter calibrations, aircraft performance calculations, aircraft and rocket design, ballistic tables, and meteorological diagrams." For example, the U.S. Standard Atmosphere derives the values for air temperature, pressure, and mass density, as a function of altitude above sea level. Other static atmospheric models may have other outputs, or depend on inputs besides altitude. Basic assumptions The gas which comprises an atmosphere is usually assumed to be an ideal gas, which is to say: Where ρ is mass density, M is average molecular weight, P is pressure, T is temperature, and R is the ideal gas constant. The gas is held in place by so-called "hydrostatic" forces. That is to say, for a particular layer of gas at some altitude: the downward (towards the planet) force of its weight, the downward force exerted by pressure in the layer above it, and the upward force exerted by pressure in the layer below, all sum to zero. Mathematically this is: Finally, these variables describing the system do not change with time; i.e. it is a static system. g_0, gravitational acceleration is used here as a constant, with same value as standard gravity (average acceleration due to gravity on the surface of the Earth or other big body). For the basis of simplicity it doesn't vary with latitude, altitude or location. The variation due to all these The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses. What kind of air does a continental polar air mass hold? A. warm dry air B. cold dry air C. frigid wet air D. cool moist air Answer:
sciq-9430
multiple_choice
What describes how much matter is in a certain amount of space?
[ "viscosity", "mass", "volume", "density" ]
D
Relavent Documents: Document 0::: In chemistry and related fields, the molar volume, symbol Vm, or of a substance is the ratio of the volume occupied by a substance to the amount of substance, usually given at a given temperature and pressure. It is equal to the molar mass (M) divided by the mass density (ρ): The molar volume has the SI unit of cubic metres per mole (m3/mol), although it is more typical to use the units cubic decimetres per mole (dm3/mol) for gases, and cubic centimetres per mole (cm3/mol) for liquids and solids. Definition The molar volume of a substance i is defined as its molar mass divided by its density ρi0: For an ideal mixture containing N components, the molar volume of the mixture is the weighted sum of the molar volumes of its individual components. For a real mixture the molar volume cannot be calculated without knowing the density: There are many liquid–liquid mixtures, for instance mixing pure ethanol and pure water, which may experience contraction or expansion upon mixing. This effect is represented by the quantity excess volume of the mixture, an example of excess property. Relation to specific volume Molar volume is related to specific volume by the product with molar mass. This follows from above where the specific volume is the reciprocal of the density of a substance: Ideal gases For ideal gases, the molar volume is given by the ideal gas equation; this is a good approximation for many common gases at standard temperature and pressure. The ideal gas equation can be rearranged to give an expression for the molar volume of an ideal gas: Hence, for a given temperature and pressure, the molar volume is the same for all ideal gases and is based on the gas constant: R = , or about . The molar volume of an ideal gas at 100 kPa (1 bar) is at 0 °C, at 25 °C. The molar volume of an ideal gas at 1 atmosphere of pressure is at 0 °C, at 25 °C. Crystalline solids For crystalline solids, the molar volume can be measured by X-ray crystallography. The unit cell Document 1::: Vapour density is the density of a vapour in relation to that of hydrogen. It may be defined as mass of a certain volume of a substance divided by mass of same volume of hydrogen. vapour density = mass of n molecules of gas / mass of n molecules of hydrogen gas . vapour density = molar mass of gas / molar mass of H2 vapour density = molar mass of gas / 2.016 vapour density = × molar mass (and thus: molar mass = ~2 × vapour density) For example, vapour density of mixture of NO2 and N2O4 is 38.3. Vapour density is a dimensionless quantity. Alternative definition In many web sources, particularly in relation to safety considerations at commercial and industrial facilities in the U.S., vapour density is defined with respect to air, not hydrogen. Air is given a vapour density of one. For this use, air has a molecular weight of 28.97 atomic mass units, and all other gas and vapour molecular weights are divided by this number to derive their vapour density. For example, acetone has a vapour density of 2 in relation to air. That means acetone vapour is twice as heavy as air. This can be seen by dividing the molecular weight of Acetone, 58.1 by that of air, 28.97, which equals 2. With this definition, the vapour density would indicate whether a gas is denser (greater than one) or less dense (less than one) than air. The density has implications for container storage and personnel safety—if a container can release a dense gas, its vapour could sink and, if flammable, collect until it is at a concentration sufficient for ignition. Even if not flammable, it could collect in the lower floor or level of a confined space and displace air, possibly presenting an asphyxiation hazard to individuals entering the lower part of that space. See also Relative density (also known as specific gravity) Victor Meyer apparatus Document 2::: Absolute molar mass is a process used to determine the characteristics of molecules. History The first absolute measurements of molecular weights (i.e. made without reference to standards) were based on fundamental physical characteristics and their relation to the molar mass. The most useful of these were membrane osmometry and sedimentation. Another absolute instrumental approach was also possible with the development of light scattering theory by Albert Einstein, Chandrasekhara Venkata Raman, Peter Debye, Bruno H. Zimm, and others. The problem with measurements made using membrane osmometry and sedimentation was that they only characterized the bulk properties of the polymer sample. Moreover, the measurements were excessively time consuming and prone to operator error. In order to gain information about a polydisperse mixture of molar masses, a method for separating the different sizes was developed. This was achieved by the advent of size exclusion chromatography (SEC). SEC is based on the fact that the pores in the packing material of chromatography columns could be made small enough for molecules to become temporarily lodged in their interstitial spaces. As the sample makes its way through a column the smaller molecules spend more time traveling in these void spaces than the larger ones, which have fewer places to "wander". The result is that a sample is separated according to its hydrodynamic volume . As a consequence, the big molecules come out first, and then the small ones follow in the eluent. By choosing a suitable column packing material it is possible to define the resolution of the system. Columns can also be combined in series to increase resolution or the range of sizes studied. The next step is to convert the time at which the samples eluted into a measurement of molar mass. This is possible because if the molar mass of a standard were known, the time at which this standard eluted should be equal to a specific molar mass. Using multiple Document 3::: Volume solid is the volume of paint after it has dried. This is different than the weight solid. Paint may contain solvent, resin, pigments, and additives. Many paints do not contain any solvent. After applying the paint, the solid portion will be left on the substrate. Volume solid is the term that indicates the solid proportion of the paint on a volume basis. For example, if the paint is applied in a wet film at a 100 μm thickness and the volume solid of paint is 50%, then the dry film thickness (DFT) will be 50 μm as 50% of the wet paint has evaporated. Suppose the volume solid is 100%, and the wet film thickness is also 100 μm. Then after complete drying of the paint, the DFT will be 100 μm because no solvent will be evaporated. This is an important concept when using paint industrially to calculate the cost of painting. It can be said that it is the real volume of paint. Here is the formula by which one can calculate the volume solid of paint, (Total sum by volume of each solid ingredient in paint x 100%)/ Total sum by volume of each ingredient in paint. A simple method that anyone can do to determine volume solids empirically is to apply paint to a steel surface with an application knife and measure the wet film thickness. Then cure the paint and measure the dry film thickness. The percentage of dry to wet represents the percentage of volume solids. In earlier days, the volume solid was measured by a disc method but now a sophisticated instrument is also available which takes only a drop of paint to check the volume solid. Understanding 'volume solids' allows knowing the true cost of different coatings and how much paint is used to perform its function. Generally, more expensive paints have a higher volume of solids and provide better coverage. Document 4::: Physical or chemical properties of materials and systems can often be categorized as being either intensive or extensive, according to how the property changes when the size (or extent) of the system changes. The terms "intensive and extensive quantities" were introduced into physics by German mathematician Georg Helm in 1898, and by American physicist and chemist Richard C. Tolman in 1917. According to International Union of Pure and Applied Chemistry (IUPAC), an intensive property or intensive quantity is one whose magnitude is independent of the size of the system. An intensive property is not necessarily homogeneously distributed in space; it can vary from place to place in a body of matter and radiation. Examples of intensive properties include temperature, T; refractive index, n; density, ρ; and hardness, η. By contrast, an extensive property or extensive quantity is one whose magnitude is additive for subsystems. Examples include mass, volume and entropy. Not all properties of matter fall into these two categories. For example, the square root of the volume is neither intensive nor extensive. If a system is doubled in size by juxtaposing a second identical system, the value of an intensive property equals the value for each subsystem and the value of an extensive property is twice the value for each subsystem. However the property √V is instead multiplied by √2 . Intensive properties An intensive property is a physical quantity whose value does not depend on the amount of substance which was measured. The most obvious intensive quantities are ratios of extensive quantities. In a homogeneous system divided into two halves, all its extensive properties, in particular its volume and its mass, are divided into two halves. All its intensive properties, such as the mass per volume (mass density) or volume per mass (specific volume), must remain the same in each half. The temperature of a system in thermal equilibrium is the same as the temperature of any part The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses. What describes how much matter is in a certain amount of space? A. viscosity B. mass C. volume D. density Answer:
sciq-10730
multiple_choice
What type of organisms do carnivores feed on?
[ "decomposers", "eggs", "plants", "animals" ]
D
Relavent Documents: Document 0::: Consumer–resource interactions are the core motif of ecological food chains or food webs, and are an umbrella term for a variety of more specialized types of biological species interactions including prey-predator (see predation), host-parasite (see parasitism), plant-herbivore and victim-exploiter systems. These kinds of interactions have been studied and modeled by population ecologists for nearly a century. Species at the bottom of the food chain, such as algae and other autotrophs, consume non-biological resources, such as minerals and nutrients of various kinds, and they derive their energy from light (photons) or chemical sources. Species higher up in the food chain survive by consuming other species and can be classified by what they eat and how they obtain or find their food. Classification of consumer types The standard categorization Various terms have arisen to define consumers by what they eat, such as meat-eating carnivores, fish-eating piscivores, insect-eating insectivores, plant-eating herbivores, seed-eating granivores, and fruit-eating frugivores and omnivores are meat eaters and plant eaters. An extensive classification of consumer categories based on a list of feeding behaviors exists. The Getz categorization Another way of categorizing consumers, proposed by South African American ecologist Wayne Getz, is based on a biomass transformation web (BTW) formulation that organizes resources into five components: live and dead animal, live and dead plant, and particulate (i.e. broken down plant and animal) matter. It also distinguishes between consumers that gather their resources by moving across landscapes from those that mine their resources by becoming sessile once they have located a stock of resources large enough for them to feed on during completion of a full life history stage. In Getz's scheme, words for miners are of Greek etymology and words for gatherers are of Latin etymology. Thus a bestivore, such as a cat, preys on live animal Document 1::: A herbivore is an animal anatomically and physiologically adapted to eating plant material, for example foliage or marine algae, for the main component of its diet. As a result of their plant diet, herbivorous animals typically have mouthparts adapted to rasping or grinding. Horses and other herbivores have wide flat teeth that are adapted to grinding grass, tree bark, and other tough plant material. A large percentage of herbivores have mutualistic gut flora that help them digest plant matter, which is more difficult to digest than animal prey. This flora is made up of cellulose-digesting protozoans or bacteria. Etymology Herbivore is the anglicized form of a modern Latin coinage, herbivora, cited in Charles Lyell's 1830 Principles of Geology. Richard Owen employed the anglicized term in an 1854 work on fossil teeth and skeletons. Herbivora is derived from Latin herba 'small plant, herb' and vora, from vorare 'to eat, devour'. Definition and related terms Herbivory is a form of consumption in which an organism principally eats autotrophs such as plants, algae and photosynthesizing bacteria. More generally, organisms that feed on autotrophs in general are known as primary consumers. Herbivory is usually limited to animals that eat plants. Insect herbivory can cause a variety of physical and metabolic alterations in the way the host plant interacts with itself and other surrounding biotic factors. Fungi, bacteria, and protists that feed on living plants are usually termed plant pathogens (plant diseases), while fungi and microbes that feed on dead plants are described as saprotrophs. Flowering plants that obtain nutrition from other living plants are usually termed parasitic plants. There is, however, no single exclusive and definitive ecological classification of consumption patterns; each textbook has its own variations on the theme. Evolution of herbivory The understanding of herbivory in geological time comes from three sources: fossilized plants, which may Document 2::: Feeding is the process by which organisms, typically animals, obtain food. Terminology often uses either the suffixes -vore, -vory, or -vorous from Latin vorare, meaning "to devour", or -phage, -phagy, or -phagous from Greek φαγεῖν (), meaning "to eat". Evolutionary history The evolution of feeding is varied with some feeding strategies evolving several times in independent lineages. In terrestrial vertebrates, the earliest forms were large amphibious piscivores 400 million years ago. While amphibians continued to feed on fish and later insects, reptiles began exploring two new food types, other tetrapods (carnivory), and later, plants (herbivory). Carnivory was a natural transition from insectivory for medium and large tetrapods, requiring minimal adaptation (in contrast, a complex set of adaptations was necessary for feeding on highly fibrous plant materials). Evolutionary adaptations The specialization of organisms towards specific food sources is one of the major causes of evolution of form and function, such as: mouth parts and teeth, such as in whales, vampire bats, leeches, mosquitos, predatory animals such as felines and fishes, etc. distinct forms of beaks in birds, such as in hawks, woodpeckers, pelicans, hummingbirds, parrots, kingfishers, etc. specialized claws and other appendages, for apprehending or killing (including fingers in primates) changes in body colour for facilitating camouflage, disguise, setting up traps for preys, etc. changes in the digestive system, such as the system of stomachs of herbivores, commensalism and symbiosis Classification By mode of ingestion There are many modes of feeding that animals exhibit, including: Filter feeding: obtaining nutrients from particles suspended in water Deposit feeding: obtaining nutrients from particles suspended in soil Fluid feeding: obtaining nutrients by consuming other organisms' fluids Bulk feeding: obtaining nutrients by eating all of an organism. Ram feeding and suction feeding: in Document 3::: The trophic level of an organism is the position it occupies in a food web. A food chain is a succession of organisms that eat other organisms and may, in turn, be eaten themselves. The trophic level of an organism is the number of steps it is from the start of the chain. A food web starts at trophic level 1 with primary producers such as plants, can move to herbivores at level 2, carnivores at level 3 or higher, and typically finish with apex predators at level 4 or 5. The path along the chain can form either a one-way flow or a food "web". Ecological communities with higher biodiversity form more complex trophic paths. The word trophic derives from the Greek τροφή (trophē) referring to food or nourishment. History The concept of trophic level was developed by Raymond Lindeman (1942), based on the terminology of August Thienemann (1926): "producers", "consumers", and "reducers" (modified to "decomposers" by Lindeman). Overview The three basic ways in which organisms get food are as producers, consumers, and decomposers. Producers (autotrophs) are typically plants or algae. Plants and algae do not usually eat other organisms, but pull nutrients from the soil or the ocean and manufacture their own food using photosynthesis. For this reason, they are called primary producers. In this way, it is energy from the sun that usually powers the base of the food chain. An exception occurs in deep-sea hydrothermal ecosystems, where there is no sunlight. Here primary producers manufacture food through a process called chemosynthesis. Consumers (heterotrophs) are species that cannot manufacture their own food and need to consume other organisms. Animals that eat primary producers (like plants) are called herbivores. Animals that eat other animals are called carnivores, and animals that eat both plants and other animals are called omnivores. Decomposers (detritivores) break down dead plant and animal material and wastes and release it again as energy and nutrients into Document 4::: A graminivore is a herbivorous animal that feeds primarily on grass, specifically "true" grasses, plants of the family Poaceae (also known as Graminae). Graminivory is a form of grazing. These herbivorous animals have digestive systems that are adapted to digest large amounts of cellulose, which is abundant in fibrous plant matter and more difficult to break down for many other animals. As such, they have specialized enzymes to aid in digestion and in some cases symbiotic bacteria that live in their digestive track and "assist" with the digestive process through fermentation as the matter travels through the intestines. Horses, cattle, geese, guinea pigs, hippopotamuses, capybara and giant pandas are examples of vertebrate graminivores. Some carnivorous vertebrates, such as dogs and cats, are known to eat grass occasionally. Grass consumption in dogs can be a way to rid their intestinal tract of parasites that may be threatening to the carnivore's health. Various invertebrates also have graminivorous diets. Many grasshoppers, such as individuals from the family Acrididae, have diets consisting primarily of plants from the family Poaceae. Although humans are not graminivores, we do get much of our nutrition from a type of grass called cereal, and especially from the fruit of that grass which is called grain. Graminivores generally exhibit a preference on which species of grass they choose to consume. For example, according to a study done on North American bison feeding on shortgrass plains in north-eastern Colorado, the cattle consumed a total of thirty-six different species of plant. Of that thirty-six, five grass species were favoured and consumed the most pervasively. The average consumption of these five species comprised about 80% of their diet. A few of these species include Aristida longiseta, Muhlenbergia species, and Bouteloua gracilis. The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses. What type of organisms do carnivores feed on? A. decomposers B. eggs C. plants D. animals Answer:
sciq-7497
multiple_choice
Juvenile hormone modulates the activity of what?
[ "ecdysteroid", "growth", "estrogen", "reproduction" ]
A
Relavent Documents: Document 0::: The following is a list of hormones found in Homo sapiens. Spelling is not uniform for many hormones. For example, current North American and international usage uses estrogen and gonadotropin, while British usage retains the Greek digraph in oestrogen and favours the earlier spelling gonadotrophin. Hormone listing Steroid Document 1::: Hormonal imprinting (HI) is a phenomenon which takes place at the first encounter between a hormone and its developing receptor in the critical periods of life (in unicellulars during the whole life) and determines the later signal transduction capacity of the cell. The most important period in mammals is the perinatal one, however this system can be imprinted at weaning, at puberty and in case of continuously dividing cells during the whole life. Faulty imprinting is caused by drugs, environmental pollutants and other hormone-like molecules present in excess at the critical periods with lifelong receptorial, morphological, biochemical and behavioral consequences. HI is transmitted to the hundreds of progeny generations in unicellulars and (as proved) to a few generations also in mammals. Document 2::: In molecular biology, the crustacean neurohormone family of proteins is a family of neuropeptides expressed by arthropods. The family includes the following types of neurohormones: Crustacean hyperglycaemic hormone (CHH). CHH is primarily involved in blood sugar regulation, but also plays a role in the control of moulting and reproduction. Moult-inhibiting hormone (MIH). MIH inhibits Y-organs where moulting hormone (ecdysteroid) is secreted. A moulting cycle is initiated when MIH secretion diminishes or stops. Gonad-inhibiting hormone (GIH), also known as vitellogenesis-inhibiting hormone (VIH) because of its role in inhibiting vitellogenesis in female animals. Mandibular organ-inhibiting hormone (MOIH). MOIH represses the synthesis of methyl farnesoate, the precursor of insect juvenile hormone III in the mandibular organ. Ion transport peptide (ITP) from locust. ITP stimulates salt and water reabsorption and inhibits acid secretion in the ileum of the locust. Caenorhabditis elegans uncharacterised protein ZC168.2. These neurohormones are peptides of 70 to 80 amino acid residues which are processed from larger precursors. They contain six conserved cysteines that are involved in disulfide bonds. Document 3::: Maria Iandolo New is a professor of Pediatrics, Genomics and Genetics at Icahn School of Medicine at Mount Sinai in New York City. She is an expert in congenital adrenal hyperplasia (CAH), a genetic condition affecting the adrenal gland that can affect sexual development. Medical education New received her undergraduate degree from Cornell University in Ithaca, New York, in 1950, and her M. D. from the Perelman School of Medicine at the University of Pennsylvania in Philadelphia, in 1954. She completed an internship in medicine at Bellevue Hospital in New York, followed by a residency in pediatrics at the New York Hospital. From 1957 to 1958 she studied renal functioning under a fellowship from the National Institutes of Health (NIH). She was a research pediatrician to the Diabetic Study Group of the Comprehensive Care Teaching Program at the New York Hospital-Cornell Medical Center from 1958 to 1961, and had a second NIH fellowship under Ralph E. Peterson from 1961 to 1964, to study specific steroid hormone production during infancy, childhood and adolescence. Career In 1964, New was appointed Chief of Pediatric Endocrinology at Cornell University Medical College, a position she held for 40 years. In 1978, she was named Harold and Percy Uris Professor of Pediatric Endocrinology and Metabolism. In 1980, New was appointed Chairman of the Department of Pediatrics at Cornell University Medical College and Pediatrician-in-Chief of the Department of Pediatrics at New York-Presbyterian Hospital. She was one of the few women in the country to serve as Chair of a major division of a medical college, and her tenure lasted for 22 years. While Chairman, New founded and directed the 8-bed Children's Clinical Research Center, a clinical research center in pediatrics with groundbreaking research in pediatric endocrinology, hematology, and immunology, during the emergence of AIDS. In 2004, New was recruited to the Mount Sinai School of Medicine as Professor of Pediatrics and H Document 4::: B)) Local hormones are a large group of signaling molecules that do not circulate within the blood. Local hormones are produced by nerve and gland cells and bind to either neighboring cells or the same type of cell that produced them. Local hormones are activated and inactivated quickly. They are released during physical work and exercise. They mainly control smooth and vascular muscle dilation. Strength of response is dependent upon the concentration of receptors of target cell and the amount of ligand ( the specific local hormone). Eicosanoids (ī′kō-să-noydz; eicosa = twenty, eidos = formed) are a primary type of local hormone. These local hormones are formed by a chain of 20 amino acid?s and fatty acids from phospholipids in the cell membrane. Eicosanoids initiate either autocrine stimulation or paracrine stimulation. There are two main types of eicosanoids: prostaglandins and leukotrienes, which initiate either autocrine stimulation or paracrine stimulation. Eicosanoids are the result of a ubiquitous pathway which first produces arachidonic acid, and then the eicosanoid product. Prostaglandins are the most diverse category of eicosanoids and are thought to be synthesized in most tissues of the body. This type of local hormone stimulates pain receptors and increases the inflammatory response. Nonsteroidal anti-inflammatory drugs stop the formation of prostaglandins, thus inhibiting these responses. Leukotrienes are a type of eicosanoids that are produced in leukocytes and function in inflammatory mediation. Paracrines (para- = beside or near) are local hormones that act on neighboring cells. This type of signaling involves the secretion of paracrine factors, which travel a short distance in the extracellular environment to affect nearby cells. These factors can be excitatory or inhibitory. There are a few families of factors that are very important in embryo development including fibroblast growth factor secreted them. Juxtacrines (juxta = near) are local The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses. Juvenile hormone modulates the activity of what? A. ecdysteroid B. growth C. estrogen D. reproduction Answer:
sciq-9034
multiple_choice
The smallest unit of structure and function in all living organisms is what?
[ "particle", "mineral", "cell", "compound" ]
C
Relavent Documents: Document 0::: Cellular components are the complex biomolecules and structures of which cells, and thus living organisms, are composed. Cells are the structural and functional units of life. The smallest organisms are single cells, while the largest organisms are assemblages of trillions of cells. DNA, double stranded macromolecule that carries the hereditary information of the cell and found in all living cells; each cell carries chromosome(s) having a distinctive DNA sequence. Examples include macromolecules such as proteins and nucleic acids, biomolecular complexes such as a ribosome, and structures such as membranes, and organelles. While the majority of cellular components are located within the cell itself, some may exist in extracellular areas of an organism. Cellular components may also be called biological matter or biological material. Most biological matter has the characteristics of soft matter, being governed by relatively small energies. All known life is made of biological matter. To be differentiated from other theoretical or fictional life forms, such life may be called carbon-based, cellular, organic, biological, or even simply living – as some definitions of life exclude hypothetical types of biochemistry. See also Cell (biology) Cell biology Biomolecule Organelle Tissue (biology) External links https://web.archive.org/web/20130918033010/http://bioserv.fiu.edu/~walterm/FallSpring/review1_fall05_chap_cell3.htm Document 1::: In biology, cell theory is a scientific theory first formulated in the mid-nineteenth century, that organisms are made up of cells, that they are the basic structural/organizational unit of all organisms, and that all cells come from pre-existing cells. Cells are the basic unit of structure in all organisms and also the basic unit of reproduction. The theory was once universally accepted, but now some biologists consider non-cellular entities such as viruses living organisms, and thus disagree with the first tenet. As of 2021: "expert opinion remains divided roughly a third each between yes, no and don’t know". As there is no universally accepted definition of life, discussion still continues. History With continual improvements made to microscopes over time, magnification technology became advanced enough to discover cells. This discovery is largely attributed to Robert Hooke, and began the scientific study of cells, known as cell biology. When observing a piece of cork under the scope, he was able to see pores. This was shocking at the time as it was believed no one else had seen these. To further support his theory, Matthias Schleiden and Theodor Schwann both also studied cells of both animal and plants. What they discovered were significant differences between the two types of cells. This put forth the idea that cells were not only fundamental to plants, but animals as well. Microscopes The discovery of the cell was made possible through the invention of the microscope. In the first century BC, Romans were able to make glass. They discovered that objects appeared to be larger under the glass. The expanded use of lenses in eyeglasses in the 13th century probably led to wider spread use of simple microscopes (magnifying glasses) with limited magnification. Compound microscopes, which combine an objective lens with an eyepiece to view a real image achieving much higher magnification, first appeared in Europe around 1620. In 1665, Robert Hooke used a microscope Document 2::: The cell is the basic structural and functional unit of all forms of life. Every cell consists of cytoplasm enclosed within a membrane, and contains many macromolecules such as proteins, DNA and RNA, as well as many small molecules of nutrients and metabolites. The term comes from the Latin word meaning 'small room'. Cells can acquire specified function and carry out various tasks within the cell such as replication, DNA repair, protein synthesis, and motility. Cells are capable of specialization and mobility within the cell. Most plant and animal cells are only visible under a light microscope, with dimensions between 1 and 100 micrometres. Electron microscopy gives a much higher resolution showing greatly detailed cell structure. Organisms can be classified as unicellular (consisting of a single cell such as bacteria) or multicellular (including plants and animals). Most unicellular organisms are classed as microorganisms. The study of cells and how they work has led to many other studies in related areas of biology, including: discovery of DNA, cancer systems biology, aging and developmental biology. Cell biology is the study of cells, which were discovered by Robert Hooke in 1665, who named them for their resemblance to cells inhabited by Christian monks in a monastery. Cell theory, first developed in 1839 by Matthias Jakob Schleiden and Theodor Schwann, states that all organisms are composed of one or more cells, that cells are the fundamental unit of structure and function in all living organisms, and that all cells come from pre-existing cells. Cells emerged on Earth about 4 billion years ago. Discovery With continual improvements made to microscopes over time, magnification technology became advanced enough to discover cells. This discovery is largely attributed to Robert Hooke, and began the scientific study of cells, known as cell biology. When observing a piece of cork under the scope, he was able to see pores. This was shocking at the time as i Document 3::: This list of life sciences comprises the branches of science that involve the scientific study of life – such as microorganisms, plants, and animals including human beings. This science is one of the two major branches of natural science, the other being physical science, which is concerned with non-living matter. Biology is the overall natural science that studies life, with the other life sciences as its sub-disciplines. Some life sciences focus on a specific type of organism. For example, zoology is the study of animals, while botany is the study of plants. Other life sciences focus on aspects common to all or many life forms, such as anatomy and genetics. Some focus on the micro-scale (e.g. molecular biology, biochemistry) other on larger scales (e.g. cytology, immunology, ethology, pharmacy, ecology). Another major branch of life sciences involves understanding the mindneuroscience. Life sciences discoveries are helpful in improving the quality and standard of life and have applications in health, agriculture, medicine, and the pharmaceutical and food science industries. For example, it has provided information on certain diseases which has overall aided in the understanding of human health. Basic life science branches Biology – scientific study of life Anatomy – study of form and function, in plants, animals, and other organisms, or specifically in humans Astrobiology – the study of the formation and presence of life in the universe Bacteriology – study of bacteria Biotechnology – study of combination of both the living organism and technology Biochemistry – study of the chemical reactions required for life to exist and function, usually a focus on the cellular level Bioinformatics – developing of methods or software tools for storing, retrieving, organizing and analyzing biological data to generate useful biological knowledge Biolinguistics – the study of the biology and evolution of language. Biological anthropology – the study of humans, non-hum Document 4::: Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas. Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below: During adiabatic expansion of an ideal gas, its temperatureincreases decreases stays the same Impossible to tell/need more information The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well. Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses. The smallest unit of structure and function in all living organisms is what? A. particle B. mineral C. cell D. compound Answer:
sciq-1501
multiple_choice
What unit are atomic radii typically measured in?
[ "powers", "picometers", "radons", "sporozoans" ]
B
Relavent Documents: Document 0::: Quantities, Units and Symbols in Physical Chemistry, also known as the Green Book, is a compilation of terms and symbols widely used in the field of physical chemistry. It also includes a table of physical constants, tables listing the properties of elementary particles, chemical elements, and nuclides, and information about conversion factors that are commonly used in physical chemistry. The Green Book is published by the International Union of Pure and Applied Chemistry (IUPAC) and is based on published, citeable sources. Information in the Green Book is synthesized from recommendations made by IUPAC, the International Union of Pure and Applied Physics (IUPAP) and the International Organization for Standardization (ISO), including recommendations listed in the IUPAP Red Book Symbols, Units, Nomenclature and Fundamental Constants in Physics and in the ISO 31 standards. History, list of editions, and translations to non-English languages The third edition of the Green Book () was first published by IUPAC in 2007. A second printing of the third edition was released in 2008; this printing made several minor revisions to the 2007 text. A third printing of the third edition was released in 2011. The text of the third printing is identical to that of the second printing. A Japanese translation of the third edition of the Green Book () was published in 2009. A French translation of the third edition of the Green Book () was published in 2012. A Portuguese translation (Brazilian Portuguese and European Portuguese) of the third edition of the Green Book () was published in 2018, with updated values of the physical constants and atomic weights; it is referred to as the "Livro Verde". A concise four-page summary of the most important material in the Green Book was published in the July–August 2011 issue of Chemistry International, the IUPAC news magazine. The second edition of the Green Book () was first published in 1993. It was reprinted in 1995, 1996, and 1998. Document 1::: Mache (symbol ME from German Mache-Einheit, plural Maches) is a unit of volumic radioactivity named for the Austrian physicist Heinrich Mache. It was defined as the quantity of radon (ignoring its daughter isotopes; in practice, mostly radon-222) per litre of air which ionises a sustained current of 0.001 esu (0.001 StatAmpere). 1 ME = 3.64 Eman = 3.64×10−10 Ci/L = 13.4545 Bq/L. Document 2::: The mass recorded by a mass spectrometer can refer to different physical quantities depending on the characteristics of the instrument and the manner in which the mass spectrum is displayed. Units The dalton (symbol: Da) is the standard unit that is used for indicating mass on an atomic or molecular scale (atomic mass). The unified atomic mass unit (symbol: u) is equivalent to the dalton. One dalton is approximately the mass of one a single proton or neutron. The unified atomic mass unit has a value of . The amu without the "unified" prefix is an obsolete unit based on oxygen, which was replaced in 1961. Molecular mass The molecular mass (abbreviated Mr) of a substance, formerly also called molecular weight and abbreviated as MW, is the mass of one molecule of that substance, relative to the unified atomic mass unit u (equal to 1/12 the mass of one atom of 12C). Due to this relativity, the molecular mass of a substance is commonly referred to as the relative molecular mass, and abbreviated to Mr. Average mass The average mass of a molecule is obtained by summing the average atomic masses of the constituent elements. For example, the average mass of natural water with formula H2O is 1.00794 + 1.00794 + 15.9994 = 18.01528 Da. Mass number The mass number, also called the nucleon number, is the number of protons and neutrons in an atomic nucleus. The mass number is unique for each isotope of an element and is written either after the element name or as a superscript to the left of an element's symbol. For example, carbon-12 (12C) has 6 protons and 6 neutrons. Nominal mass The nominal mass for an element is the mass number of its most abundant naturally occurring stable isotope, and for an ion or molecule, the nominal mass is the sum of the nominal masses of the constituent atoms. Isotope abundances are tabulated by IUPAC: for example carbon has two stable isotopes 12C at 98.9% natural abundance and 13C at 1.1% natural abundance, thus the nominal mass of carbon i Document 3::: Quantity calculus is the formal method for describing the mathematical relations between abstract physical quantities. Its roots can be traced to Fourier's concept of dimensional analysis (1822). The basic axiom of quantity calculus is Maxwell's description of a physical quantity as the product of a "numerical value" and a "reference quantity" (i.e. a "unit quantity" or a "unit of measurement"). De Boer summarized the multiplication, division, addition, association and commutation rules of quantity calculus and proposed that a full axiomatization has yet to be completed. Measurements are expressed as products of a numeric value with a unit symbol, e.g. "12.7 m". Unlike algebra, the unit symbol represents a measurable quantity such as a meter, not an algebraic variable. A careful distinction needs to be made between abstract quantities and measurable quantities. The multiplication and division rules of quantity calculus are applied to SI base units (which are measurable quantities) to define SI derived units, including dimensionless derived units, such as the radian (rad) and steradian (sr) which are useful for clarity, although they are both algebraically equal to 1. Thus there is some disagreement about whether it is meaningful to multiply or divide units. Emerson suggests that if the units of a quantity are algebraically simplified, they then are no longer units of that quantity. Johansson proposes that there are logical flaws in the application of quantity calculus, and that the so-called dimensionless quantities should be understood as "unitless quantities". How to use quantity calculus for unit conversion and keeping track of units in algebraic manipulations is explained in the handbook Quantities, Units and Symbols in Physical Chemistry. Notes Document 4::: This article gives a list of conversion factors for several physical quantities. A number of different units (some only of historical interest) are shown and expressed in terms of the corresponding SI unit. Conversions between units in the metric system are defined by their prefixes (for example, 1 kilogram = 1000 grams, 1 milligram = 0.001 grams) and are thus not listed in this article. Exceptions are made if the unit is commonly known by another name (for example, 1 micron = 10−6 metre). Within each table, the units are listed alphabetically, and the SI units (base or derived) are highlighted. The following quantities are considered: length, area, volume, plane angle, solid angle, mass, density, time, frequency, velocity, volumetric flow rate, acceleration, force, pressure (or mechanical stress), torque (or moment of force), energy, power (or heat flow rate), action, dynamic viscosity, kinematic viscosity, electric current, electric charge, electric dipole, electromotive force (or electric potential difference), electrical resistance, capacitance, magnetic flux, magnetic flux density, inductance, temperature, information entropy, luminous intensity, luminance, luminous flux, illuminance, radiation. Length Area Volume Plane angle Solid angle Mass Notes: See Weight for detail of mass/weight distinction and conversion. Avoirdupois is a system of mass based on a pound of 16 ounces, while Troy weight is the system of mass where 12 troy ounces equals one troy pound. The symbol is used to denote standard gravity in order to avoid confusion with the (upright) g symbol for gram. Density Time Frequency Speed or velocity A velocity consists of a speed combined with a direction; the speed part of the velocity takes units of speed. Flow (volume) Acceleration Force Pressure or mechanical stress Torque or moment of force Energy Power or heat flow rate Action Dynamic viscosity Kinematic viscosity Electric current Electric charge Electric dipole Elec The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses. What unit are atomic radii typically measured in? A. powers B. picometers C. radons D. sporozoans Answer:
sciq-946
multiple_choice
What word describes any chemical that affects the body’s structure or function?
[ "gift", "treatment", "toxin", "drug" ]
D
Relavent Documents: Document 0::: In chemistry, a reagent ( ) or analytical reagent is a substance or compound added to a system to cause a chemical reaction, or test if one occurs. The terms reactant and reagent are often used interchangeably, but reactant specifies a substance consumed in the course of a chemical reaction. Solvents, though involved in the reaction mechanism, are usually not called reactants. Similarly, catalysts are not consumed by the reaction, so they are not reactants. In biochemistry, especially in connection with enzyme-catalyzed reactions, the reactants are commonly called substrates. Definitions Organic chemistry In organic chemistry, the term "reagent" denotes a chemical ingredient (a compound or mixture, typically of inorganic or small organic molecules) introduced to cause the desired transformation of an organic substance. Examples include the Collins reagent, Fenton's reagent, and Grignard reagents. Analytical chemistry In analytical chemistry, a reagent is a compound or mixture used to detect the presence or absence of another substance, e.g. by a color change, or to measure the concentration of a substance, e.g. by colorimetry. Examples include Fehling's reagent, Millon's reagent, and Tollens' reagent. Commercial or laboratory preparations In commercial or laboratory preparations, reagent-grade designates chemical substances meeting standards of purity that ensure the scientific precision and reliability of chemical analysis, chemical reactions or physical testing. Purity standards for reagents are set by organizations such as ASTM International or the American Chemical Society. For instance, reagent-quality water must have very low levels of impurities such as sodium and chloride ions, silica, and bacteria, as well as a very high electrical resistivity. Laboratory products which are less pure, but still useful and economical for undemanding work, may be designated as technical, practical, or crude grade to distinguish them from reagent versions. Biology In t Document 1::: An active ingredient is any ingredient that provides biologically active or other direct effect in the diagnosis, cure, mitigation, treatment, or prevention of disease or to affect the structure or any function of the body of humans or animals. The similar terms active pharmaceutical ingredient (abbreviated as API) and bulk active are also used in medicine. The term active substance may be used for natural products. Some medication products can contain more than one active ingredient. The traditional word for the active pharmaceutical agent is pharmacon or pharmakon (from , adapted from pharmacos) which originally denoted a magical substance or drug. The terms active constituent or active principle are often chosen when referring to the active substance of interest in a plant (such as salicylic acid in willow bark or arecoline in areca nuts), since the word "ingredient" can be taken to connote a sense of human agency (that is, something that a person combines with other substances), whereas the natural products present in plants were not added by any human agency but rather occurred naturally ("a plant doesn't have ingredients"). In contrast with the active ingredients, the inactive ingredients are usually called excipients in pharmaceutical contexts. The main excipient that serves as a medium for conveying the active ingredient is usually called the vehicle. For example, petrolatum and mineral oil are common vehicles. The term 'inactive' should not, however, be misconstrued as meaning inert. Pharmaceuticals The dosage form for a pharmaceutical contains the active pharmaceutical ingredient, which is the drug substance itself, and excipients, which are the ingredients of the tablet, or the liquid in which the active agent is suspended, or other material that is pharmaceutically inert. Drugs are chosen primarily for their active ingredients. During formulation development, the excipients are chosen carefully so that the active ingredient can reach the target si Document 2::: A natural neuroactive substance (NAS) is a chemical synthesized by neurons that affects the actions of other neurons or muscle cells. Natural neuroactive substances include neurotransmitters, neurohormones, and neuromodulators. Neurotransmitters work only between adjacent neurons through synapses. Neurohormones are released into the blood and work at a distance. Some natural neuroactive substances act as both transmitters and as hormones. Document 3::: In pharmacology and biochemistry, mode of action (MoA) describes a functional or anatomical change, resulting from the exposure of a living organism to a substance. In comparison, a mechanism of action (MOA) describes such changes at the molecular level. A mode of action is important in classifying chemicals, as it represents an intermediate level of complexity in between molecular mechanisms and physiological outcomes, especially when the exact molecular target has not yet been elucidated or is subject to debate. A mechanism of action of a chemical could be "binding to DNA" while its broader mode of action would be "transcriptional regulation". However, there is no clear consensus and the term mode of action is also often used, especially in the study of pesticides, to describe molecular mechanisms such as action on specific nuclear receptors or enzymes. See also Mechanism of action in pharmaceuticals Adverse outcome pathway Document 4::: A xenobiotic is a chemical substance found within an organism that is not naturally produced or expected to be present within the organism. It can also cover substances that are present in much higher concentrations than are usual. Natural compounds can also become xenobiotics if they are taken up by another organism, such as the uptake of natural human hormones by fish found downstream of sewage treatment plant outfalls, or the chemical defenses produced by some organisms as protection against predators. The term xenobiotics, however, is very often used in the context of pollutants such as dioxins and polychlorinated biphenyls and their effect on the biota, because xenobiotics are understood as substances foreign to an entire biological system, i.e. artificial substances, which did not exist in nature before their synthesis by humans. The term xenobiotic is derived from the Greek words ξένος (xenos) = foreigner, stranger and βίος (bios) = life, plus the Greek suffix for adjectives -τικός, -ή, -όν (-tikos, -ē, -on). Xenobiotics may be grouped as carcinogens, drugs, environmental pollutants, food additives, hydrocarbons, and pesticides. Xenobiotic metabolism The body removes xenobiotics by xenobiotic metabolism. This consists of the deactivation and the excretion of xenobiotics and happens mostly in the liver. Excretion routes are urine, feces, breath, and sweat. Hepatic enzymes are responsible for the metabolism of xenobiotics by first activating them (oxidation, reduction, hydrolysis, and/or hydration of the xenobiotic), and then conjugating the active secondary metabolite with glucuronic acid, sulfuric acid, or glutathione, followed by excretion in bile or urine. An example of a group of enzymes involved in xenobiotic metabolism is hepatic microsomal cytochrome P450. These enzymes that metabolize xenobiotics are very important for the pharmaceutical industry because they are responsible for the breakdown of medications. A species with this unique cytochrome P The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses. What word describes any chemical that affects the body’s structure or function? A. gift B. treatment C. toxin D. drug Answer:
sciq-8853
multiple_choice
Are the majority of archaea chemotrophs or photosynthetic?
[ "chemotrophs", "autotrophs", "photosynthetics", "neither" ]
A
Relavent Documents: Document 0::: Archaea ( ; : archaeon ) is a domain of single-celled organisms. These microorganisms lack cell nuclei and are therefore prokaryotes. Archaea were initially classified as bacteria, receiving the name archaebacteria (in the Archaebacteria kingdom), but this term has fallen out of use. Archaeal cells have unique properties separating them from the other two domains, Bacteria and Eukaryota. Archaea are further divided into multiple recognized phyla. Classification is difficult because most have not been isolated in a laboratory and have been detected only by their gene sequences in environmental samples. It is unknown if these are able to produce endospores. Archaea and bacteria are generally similar in size and shape, although a few archaea have very different shapes, such as the flat, square cells of Haloquadratum walsbyi. Despite this morphological similarity to bacteria, archaea possess genes and several metabolic pathways that are more closely related to those of eukaryotes, notably for the enzymes involved in transcription and translation. Other aspects of archaeal biochemistry are unique, such as their reliance on ether lipids in their cell membranes, including archaeols. Archaea use more diverse energy sources than eukaryotes, ranging from organic compounds such as sugars, to ammonia, metal ions or even hydrogen gas. The salt-tolerant Haloarchaea use sunlight as an energy source, and other species of archaea fix carbon (autotrophy), but unlike plants and cyanobacteria, no known species of archaea does both. Archaea reproduce asexually by binary fission, fragmentation, or budding; unlike bacteria, no known species of Archaea form endospores. The first observed archaea were extremophiles, living in extreme environments such as hot springs and salt lakes with no other organisms. Improved molecular detection tools led to the discovery of archaea in almost every habitat, including soil, oceans, and marshlands. Archaea are particularly numerous in the oceans, and Document 1::: In biochemistry, chemosynthesis is the biological conversion of one or more carbon-containing molecules (usually carbon dioxide or methane) and nutrients into organic matter using the oxidation of inorganic compounds (e.g., hydrogen gas, hydrogen sulfide) or ferrous ions as a source of energy, rather than sunlight, as in photosynthesis. Chemoautotrophs, organisms that obtain carbon from carbon dioxide through chemosynthesis, are phylogenetically diverse. Groups that include conspicuous or biogeochemically important taxa include the sulfur-oxidizing Gammaproteobacteria, the Campylobacterota, the Aquificota, the methanogenic archaea, and the neutrophilic iron-oxidizing bacteria. Many microorganisms in dark regions of the oceans use chemosynthesis to produce biomass from single-carbon molecules. Two categories can be distinguished. In the rare sites where hydrogen molecules (H2) are available, the energy available from the reaction between CO2 and H2 (leading to production of methane, CH4) can be large enough to drive the production of biomass. Alternatively, in most oceanic environments, energy for chemosynthesis derives from reactions in which substances such as hydrogen sulfide or ammonia are oxidized. This may occur with or without the presence of oxygen. Many chemosynthetic microorganisms are consumed by other organisms in the ocean, and symbiotic associations between chemosynthesizers and respiring heterotrophs are quite common. Large populations of animals can be supported by chemosynthetic secondary production at hydrothermal vents, methane clathrates, cold seeps, whale falls, and isolated cave water. It has been hypothesized that anaerobic chemosynthesis may support life below the surface of Mars, Jupiter's moon Europa, and other planets. Chemosynthesis may have also been the first type of metabolism that evolved on Earth, leading the way for cellular respiration and photosynthesis to develop later. Hydrogen sulfide chemosynthesis process Giant tube worms Document 2::: Cyanobacteria (), also called Cyanobacteriota or Cyanophyta, are a phylum of gram-negative bacteria that obtain energy via photosynthesis. The name cyanobacteria refers to their color (), which similarly forms the basis of cyanobacteria's common name, blue-green algae, although they are not usually scientifically classified as algae. They appear to have originated in a freshwater or terrestrial environment. Sericytochromatia, the proposed name of the paraphyletic and most basal group, is the ancestor of both the non-photosynthetic group Melainabacteria and the photosynthetic cyanobacteria, also called Oxyphotobacteria. Cyanobacteria use photosynthetic pigments, such as carotenoids, phycobilins, and various forms of chlorophyll, which absorb energy from light. Unlike heterotrophic prokaryotes, cyanobacteria have internal membranes. These are flattened sacs called thylakoids where photosynthesis is performed. Phototrophic eukaryotes such as green plants perform photosynthesis in plastids that are thought to have their ancestry in cyanobacteria, acquired long ago via a process called endosymbiosis. These endosymbiotic cyanobacteria in eukaryotes then evolved and differentiated into specialized organelles such as chloroplasts, chromoplasts, etioplasts, and leucoplasts, collectively known as plastids. Cyanobacteria are the first organisms known to have produced oxygen. By producing and releasing oxygen as a byproduct of photosynthesis, cyanobacteria are thought to have converted the early oxygen-poor, reducing atmosphere into an oxidizing one, causing the Great Oxidation Event and the "rusting of the Earth", which dramatically changed the composition of life forms on Earth. The cyanobacteria Synechocystis and Cyanothece are important model organisms with potential applications in biotechnology for bioethanol production, food colorings, as a source of human and animal food, dietary supplements and raw materials. Cyanobacteria produce a range of toxins known as cyanotox Document 3::: The following outline is provided as an overview of and topical guide to life forms: A life form (also spelled life-form or lifeform) is an entity that is living, such as plants (flora), animals (fauna), and fungi (funga). It is estimated that more than 99% of all species that ever existed on Earth, amounting to over five billion species, are extinct. Earth is the only celestial body known to harbor life forms. No form of extraterrestrial life has been discovered yet. Archaea Archaea – a domain of single-celled microorganisms, morphologically similar to bacteria, but they possess genes and several metabolic pathways that are more closely related to those of eukaryotes, notably the enzymes involved in transcription and translation. Many archaea are extremophiles, which means living in harsh environments, such as hot springs and salt lakes, but they have since been found in a broad range of habitats. Thermoproteota – a phylum of the Archaea kingdom. Initially Thermoprotei Sulfolobales – grow in terrestrial volcanic hot springs with optimum growth occurring Euryarchaeota – In the taxonomy of microorganisms Haloarchaea Halobacteriales – in taxonomy, the Halobacteriales are an order of the Halobacteria, found in water saturated or nearly saturated with salt. Methanobacteria Methanobacteriales – information including symptoms, causes, diseases, symptoms, treatments, and other medical and health issues. Methanococci Methanococcales aka Methanocaldococcus jannaschii – thermophilic methanogenic archaea, meaning that it thrives at high temperatures and produces methane Methanomicrobia Methanosarcinales – In taxonomy, the Methanosarcinales are an order of the Methanomicrobia Methanopyri Methanopyrales – In taxonomy, the Methanopyrales are an order of the methanopyri. Thermococci Thermococcales Thermoplasmata Thermoplasmatales – An order of aerobic, thermophilic archaea, in the kingdom Halophiles – organisms that thrive in high salt concentrations Ko Document 4::: Colleen Marie Cavanaugh is an American academic microbiologist best known for her studies of hydrothermal vent ecosystems. As of 2002, she is the Edward C. Jeffrey Professor of Biology in the Department of Organismic and Evolutionary Biology at Harvard University and is affiliated with the Marine Biological Laboratory and the Woods Hole Oceanographic Institution. Cavanaugh was the first to propose that the deep-sea giant tube worm, Riftia pachyptila, obtains its food from bacteria living within its cells, an insight which she had as a graduate student at Harvard. Significantly, she made the connection that these chemoautotrophic bacteria were able to play this role through their use of chemosynthesis, the biological oxidation of inorganic compounds (e.g., hydrogen sulfide) to synthesize organic matter from very simple carbon-containing molecules, thus allowing organisms such as the bacteria (and dependent organisms such as tube worms) to exist in deep ocean without sunlight. Early life and education Cavanaugh was born in Detroit, Michigan, in 1953. Cavanaugh received her undergraduate degree from the University of Michigan in 1977, where she initially studied music but ultimately majored in ecology. She says her life changed direction in her sophomore year when she heard about a course in marine ecology at the oceanographic center in Woods Hole, Massachusetts. There, her work involved wading out into chilly waters to study the mating habits of horseshoe crabs, and she described herself as "[falling] in love" with the relaxed camaraderie and exchange of ideas between biologists, geologists, and scientists from other disciplines. Cavanaugh took a Marine Ecology course as an undergraduate offered by the University of Michigan, stayed in Woods Hole afterwards (as her car needed repair) looking for a job, and ultimately replaced a "no show" in a Boston University undergraduate research program, which returned her to work with local horseshoe crabs. Cavanaugh then move The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses. Are the majority of archaea chemotrophs or photosynthetic? A. chemotrophs B. autotrophs C. photosynthetics D. neither Answer:
sciq-10648
multiple_choice
Because heating results in a greater fraction of molecules having the necessary kinetic energy to escape the surface of the liquid, what occurs more quickly when liquid is heated?
[ "oxidation", "evaporation", "absorption", "expansion" ]
B
Relavent Documents: Document 0::: Boiling is the rapid phase transition from liquid to gas or vapor; the reverse of boiling is condensation. Boiling occurs when a liquid is heated to its boiling point, so that the vapour pressure of the liquid is equal to the pressure exerted on the liquid by the surrounding atmosphere. Boiling and evaporation are the two main forms of liquid vapourization. There are two main types of boiling: nucleate boiling where small bubbles of vapour form at discrete points, and critical heat flux boiling where the boiling surface is heated above a certain critical temperature and a film of vapour forms on the surface. Transition boiling is an intermediate, unstable form of boiling with elements of both types. The boiling point of water is 100 °C or 212 °F but is lower with the decreased atmospheric pressure found at higher altitudes. Boiling water is used as a method of making it potable by killing microbes and viruses that may be present. The sensitivity of different micro-organisms to heat varies, but if water is held at for one minute, most micro-organisms and viruses are inactivated. Ten minutes at a temperature of 70 °C (158 °F) is also sufficient to inactivate most bacteria. Boiling water is also used in several cooking methods including boiling, steaming, and poaching. Types Free convection The lowest heat flux seen in boiling is only sufficient to cause [natural convection], where the warmer fluid rises due to its slightly lower density. This condition occurs only when the superheat is very low, meaning that the hot surface near the fluid is nearly the same temperature as the boiling point. Nucleate Nucleate boiling is characterised by the growth of bubbles or pops on a heated surface (heterogeneous nucleation), which rises from discrete points on a surface, whose temperature is only slightly above the temperature of the liquid. In general, the number of nucleation sites is increased by an increasing surface temperature. An irregular surface of the boiling Document 1::: The Stefan flow, occasionally called Stefan's flow, is a transport phenomenon concerning the movement of a chemical species by a flowing fluid (typically in the gas phase) that is induced to flow by the production or removal of the species at an interface. Any process that adds the species of interest to or removes it from the flowing fluid may cause the Stefan flow, but the most common processes include evaporation, condensation, chemical reaction, sublimation, ablation, adsorption, absorption, and desorption. It was named after the Slovenian physicist, mathematician, and poet Josef Stefan for his early work on calculating evaporation rates. The Stefan flow is distinct from diffusion as described by Fick's law, but diffusion almost always also occurs in multi-species systems that are experiencing the Stefan flow. In systems undergoing one of the species addition or removal processes mentioned previously, the addition or removal generates a mean flow in the flowing fluid as the fluid next to the interface is displaced by the production or removal of additional fluid by the processes occurring at the interface. The transport of the species by this mean flow is the Stefan flow. When concentration gradients of the species are also present, diffusion transports the species relative to the mean flow. The total transport rate of the species is then given by a summation of the Stefan flow and diffusive contributions. An example of the Stefan flow occurs when a droplet of liquid evaporates in air. In this case, the vapor/air mixture surrounding the droplet is the flowing fluid, and liquid/vapor boundary of the droplet is the interface. As heat is absorbed by the droplet from the environment, some of the liquid evaporates into vapor at the surface of the droplet, and flows away from the droplet as it is displaced by additional vapor evaporating from the droplet. This process causes the flowing medium to move away from the droplet at some mean speed that is dependent on Document 2::: Heat transfer is a discipline of thermal engineering that concerns the generation, use, conversion, and exchange of thermal energy (heat) between physical systems. Heat transfer is classified into various mechanisms, such as thermal conduction, thermal convection, thermal radiation, and transfer of energy by phase changes. Engineers also consider the transfer of mass of differing chemical species (mass transfer in the form of advection), either cold or hot, to achieve heat transfer. While these mechanisms have distinct characteristics, they often occur simultaneously in the same system. Heat conduction, also called diffusion, is the direct microscopic exchanges of kinetic energy of particles (such as molecules) or quasiparticles (such as lattice waves) through the boundary between two systems. When an object is at a different temperature from another body or its surroundings, heat flows so that the body and the surroundings reach the same temperature, at which point they are in thermal equilibrium. Such spontaneous heat transfer always occurs from a region of high temperature to another region of lower temperature, as described in the second law of thermodynamics. Heat convection occurs when the bulk flow of a fluid (gas or liquid) carries its heat through the fluid. All convective processes also move heat partly by diffusion, as well. The flow of fluid may be forced by external processes, or sometimes (in gravitational fields) by buoyancy forces caused when thermal energy expands the fluid (for example in a fire plume), thus influencing its own transfer. The latter process is often called "natural convection". The former process is often called "forced convection." In this case, the fluid is forced to flow by use of a pump, fan, or other mechanical means. Thermal radiation occurs through a vacuum or any transparent medium (solid or fluid or gas). It is the transfer of energy by means of photons or electromagnetic waves governed by the same laws. Overview Heat Document 3::: Bumping is a phenomenon in chemistry where homogeneous liquids boiled in a test tube or other container will superheat and, upon nucleation, rapid boiling will expel the liquid from the container. In extreme cases, the container may be broken. Cause Bumping occurs when a liquid is heated or has its pressure reduced very rapidly, typically in smooth, clean glassware. The hardest part of bubble formation is the initial formation of the bubble; once a bubble has formed, it can grow quickly. Because the liquid is typically above its boiling point, when the liquid finally starts to boil, a large vapor bubble is formed that pushes the liquid out of the test tube, typically at high speed. This rapid expulsion of boiling liquid poses a serious hazard to others and oneself in the lab. Furthermore, if a liquid is boiled and cooled back down, the chance of bumping increases on each subsequent boil, because each heating cycle progressively de-gasses the liquid, reducing the number of remaining nucleation sites. Prevention The most common way of preventing bumping is by adding one or two boiling chips to the reaction vessel. However, these alone may not prevent bumping and for this reason it is advisable to boil liquids in a boiling tube, a boiling flask, or an Erlenmeyer flask. In addition, heating test tubes should never be pointed towards any person, just in case bumping does occur. Whenever a liquid is cooled below its boiling point and re-heated to a boil, a new boiling chip will be needed, as the pores in the old boiling chip tend to fill with solvent, rendering it ineffective. A sealed capillary tube can also be placed in a boiling solution to provide a nucleation site, reducing the bumping risk and allowing its easy removal from a system. Stirring a liquid also lessens the chances of bumping, as the resulting vortex breaks up any large bubbles that might form, and the stirring itself creates bubbles. Document 4::: In fluid thermodynamics, nucleate boiling is a type of boiling that takes place when the surface temperature is hotter than the saturated fluid temperature by a certain amount but where the heat flux is below the critical heat flux. For water, as shown in the graph below, nucleate boiling occurs when the surface temperature is higher than the saturation temperature () by between . The critical heat flux is the peak on the curve between nucleate boiling and transition boiling. The heat transfer from surface to liquid is greater than that in film boiling. Nucleate boiling is common in electric kettles and is responsible for the noise that occurs before boiling occurs. It also occurs in water boilers where water is rapidly heated. Mechanism Two different regimes may be distinguished in the nucleate boiling range. When the temperature difference is between approximately above TS, isolated bubbles form at nucleation sites and separate from the surface. This separation induces considerable fluid mixing near the surface, substantially increasing the convective heat transfer coefficient and the heat flux. In this regime, most of the heat transfer is through direct transfer from the surface to the liquid in motion at the surface and not through the vapor bubbles rising from the surface. Between above TS, a second flow regime may be observed. As more nucleation sites become active, increased bubble formation causes bubble interference and coalescence. In this region the vapor escapes as jets or columns which subsequently merge into slugs of vapor. Interference between the densely populated bubbles inhibits the motion of liquid near the surface. This is observed on the graph as a change in the direction of the gradient of the curve or an inflection in the boiling curve. After this point, the heat transfer coefficient starts to reduce as the surface temperature is further increased although the product of the heat transfer coefficient and the temperature difference (t The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses. Because heating results in a greater fraction of molecules having the necessary kinetic energy to escape the surface of the liquid, what occurs more quickly when liquid is heated? A. oxidation B. evaporation C. absorption D. expansion Answer:
sciq-6965
multiple_choice
What term is used to describe anything that takes up space and has mass?
[ "opacity", "weight", "matter", "size" ]
C
Relavent Documents: Document 0::: In common usage, the mass of an object is often referred to as its weight, though these are in fact different concepts and quantities. Nevertheless, one object will always weigh more than another with less mass if both are subject to the same gravity (i.e. the same gravitational field strength). In scientific contexts, mass is the amount of "matter" in an object (though "matter" may be difficult to define), but weight is the force exerted on an object's matter by gravity. At the Earth's surface, an object whose mass is exactly one kilogram weighs approximately 9.81 newtons, the product of its mass and the gravitational field strength there. The object's weight is less on Mars, where gravity is weaker; more on Saturn, where gravity is stronger; and very small in space, far from significant sources of gravity, but it always has the same mass. Material objects at the surface of the Earth have weight despite such sometimes being difficult to measure. An object floating freely on water, for example, does not appear to have weight since it is buoyed by the water. But its weight can be measured if it is added to water in a container which is entirely supported by and weighed on a scale. Thus, the "weightless object" floating in water actually transfers its weight to the bottom of the container (where the pressure increases). Similarly, a balloon has mass but may appear to have no weight or even negative weight, due to buoyancy in air. However the weight of the balloon and the gas inside it has merely been transferred to a large area of the Earth's surface, making the weight difficult to measure. The weight of a flying airplane is similarly distributed to the ground, but does not disappear. If the airplane is in level flight, the same weight-force is distributed to the surface of the Earth as when the plane was on the runway, but spread over a larger area. A better scientific definition of mass is its description as being a measure of inertia, which is the tendency of an Document 1::: A proof mass or test mass is a known quantity of mass used in a measuring instrument as a reference for the measurement of an unknown quantity. A mass used to calibrate a weighing scale is sometimes called a calibration mass or calibration weight. A proof mass that deforms a spring in an accelerometer is sometimes called the seismic mass. In a convective accelerometer, a fluid proof mass may be employed. See also Calibration, checking or adjustment by comparison with a standard Control variable, the experimental element that is constant and unchanged throughout the course of a scientific investigation Test particle, an idealized model of an object in which all physical properties are assumed to be negligible, except for the property being studied Document 2::: In physics and mechanics, mass distribution is the spatial distribution of mass within a solid body. In principle, it is relevant also for gases or liquids, but on Earth their mass distribution is almost homogeneous. Astronomy In astronomy mass distribution has decisive influence on the development e.g. of nebulae, stars and planets. The mass distribution of a solid defines its center of gravity and influences its dynamical behaviour - e.g. the oscillations and eventual rotation. Mathematical modelling A mass distribution can be modeled as a measure. This allows point masses, line masses, surface masses, as well as masses given by a volume density function. Alternatively the latter can be generalized to a distribution. For example, a point mass is represented by a delta function defined in 3-dimensional space. A surface mass on a surface given by the equation may be represented by a density distribution , where is the mass per unit area. The mathematical modelling can be done by potential theory, by numerical methods (e.g. a great number of mass points), or by theoretical equilibrium figures. Geology In geology the aspects of rock density are involved. Rotating solids Rotating solids are affected considerably by the mass distribution, either if they are homogeneous or inhomogeneous - see Torque, moment of inertia, wobble, imbalance and stability. See also Bouguer plate Gravity Mass function Mass concentration (astronomy) External links Mass distribution of the Earth Mechanics Celestial mechanics Geophysics Mass Document 3::: To help compare different orders of magnitude, the following lists describe various mass levels between 10−59 kg and 1052 kg. The least massive thing listed here is a graviton, and the most massive thing is the observable universe. Typically, an object having greater mass will also have greater weight (see mass versus weight), especially if the objects are subject to the same gravitational field strength. Units of mass The table at right is based on the kilogram (kg), the base unit of mass in the International System of Units (SI). The kilogram is the only standard unit to include an SI prefix (kilo-) as part of its name. The gram (10−3 kg) is an SI derived unit of mass. However, the names of all SI mass units are based on gram, rather than on kilogram; thus 103 kg is a megagram (106 g), not a *kilokilogram. The tonne (t) is an SI-compatible unit of mass equal to a megagram (Mg), or 103 kg. The unit is in common use for masses above about 103 kg and is often used with SI prefixes. For example, a gigagram (Gg) or 109 g is 103 tonnes, commonly called a kilotonne. Other units Other units of mass are also in use. Historical units include the stone, the pound, the carat, and the grain. For subatomic particles, physicists use the mass equivalent to the energy represented by an electronvolt (eV). At the atomic level, chemists use the mass of one-twelfth of a carbon-12 atom (the dalton). Astronomers use the mass of the sun (). The least massive things: below 10−24 kg Unlike other physical quantities, mass–energy does not have an a priori expected minimal quantity, or an observed basic quantum as in the case of electric charge. Planck's law allows for the existence of photons with arbitrarily low energies. Consequently, there can only ever be an experimental upper bound on the mass of a supposedly massless particle; in the case of the photon, this confirmed upper bound is of the order of = . 10−24 to 10−18 kg 10−18 to 10−12 kg 10−12 to 10−6 kg 10−6 to 1 kg Document 4::: In classical physics and general chemistry, matter is any substance that has mass and takes up space by having volume. All everyday objects that can be touched are ultimately composed of atoms, which are made up of interacting subatomic particles, and in everyday as well as scientific usage, matter generally includes atoms and anything made up of them, and any particles (or combination of particles) that act as if they have both rest mass and volume. However it does not include massless particles such as photons, or other energy phenomena or waves such as light or heat. Matter exists in various states (also known as phases). These include classical everyday phases such as solid, liquid, and gas – for example water exists as ice, liquid water, and gaseous steam – but other states are possible, including plasma, Bose–Einstein condensates, fermionic condensates, and quark–gluon plasma. Usually atoms can be imagined as a nucleus of protons and neutrons, and a surrounding "cloud" of orbiting electrons which "take up space". However this is only somewhat correct, because subatomic particles and their properties are governed by their quantum nature, which means they do not act as everyday objects appear to act – they can act like waves as well as particles, and they do not have well-defined sizes or positions. In the Standard Model of particle physics, matter is not a fundamental concept because the elementary constituents of atoms are quantum entities which do not have an inherent "size" or "volume" in any everyday sense of the word. Due to the exclusion principle and other fundamental interactions, some "point particles" known as fermions (quarks, leptons), and many composites and atoms, are effectively forced to keep a distance from other particles under everyday conditions; this creates the property of matter which appears to us as matter taking up space. For much of the history of the natural sciences people have contemplated the exact nature of matter. The idea tha The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses. What term is used to describe anything that takes up space and has mass? A. opacity B. weight C. matter D. size Answer:
scienceQA-8831
multiple_choice
What do these two changes have in common? burning food on a stove water boiling on a stove
[ "Both are caused by heating.", "Both are chemical changes.", "Both are only physical changes.", "Both are caused by cooling." ]
A
Step 1: Think about each change. Food burning on a stove is a chemical change. The type of matter in the food changes. The food turns black and gives off smoke. Water boiling on the stove is a change of state. So, it is a physical change. The liquid changes into a gas, but a different type of matter is not formed. Step 2: Look at each answer choice. Both are only physical changes. Water boiling is a physical change. But burning food on a stove is not. Both are chemical changes. Burning food on a stove is a chemical change. But water boiling is not. Both are caused by heating. Both changes are caused by heating. Both are caused by cooling. Neither change is caused by cooling.
Relavent Documents: Document 0::: Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas. Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below: During adiabatic expansion of an ideal gas, its temperatureincreases decreases stays the same Impossible to tell/need more information The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well. Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in Document 1::: Physical changes are changes affecting the form of a chemical substance, but not its chemical composition. Physical changes are used to separate mixtures into their component compounds, but can not usually be used to separate compounds into chemical elements or simpler compounds. Physical changes occur when objects or substances undergo a change that does not change their chemical composition. This contrasts with the concept of chemical change in which the composition of a substance changes or one or more substances combine or break up to form new substances. In general a physical change is reversible using physical means. For example, salt dissolved in water can be recovered by allowing the water to evaporate. A physical change involves a change in physical properties. Examples of physical properties include melting, transition to a gas, change of strength, change of durability, changes to crystal form, textural change, shape, size, color, volume and density. An example of a physical change is the process of tempering steel to form a knife blade. A steel blank is repeatedly heated and hammered which changes the hardness of the steel, its flexibility and its ability to maintain a sharp edge. Many physical changes also involve the rearrangement of atoms most noticeably in the formation of crystals. Many chemical changes are irreversible, and many physical changes are reversible, but reversibility is not a certain criterion for classification. Although chemical changes may be recognized by an indication such as odor, color change, or production of a gas, every one of these indicators can result from physical change. Examples Heating and cooling Many elements and some compounds change from solids to liquids and from liquids to gases when heated and the reverse when cooled. Some substances such as iodine and carbon dioxide go directly from solid to gas in a process called sublimation. Magnetism Ferro-magnetic materials can become magnetic. The process is reve Document 2::: Thermofluids is a branch of science and engineering encompassing four intersecting fields: Heat transfer Thermodynamics Fluid mechanics Combustion The term is a combination of "thermo", referring to heat, and "fluids", which refers to liquids, gases and vapors. Temperature, pressure, equations of state, and transport laws all play an important role in thermofluid problems. Phase transition and chemical reactions may also be important in a thermofluid context. The subject is sometimes also referred to as "thermal fluids". Heat transfer Heat transfer is a discipline of thermal engineering that concerns the transfer of thermal energy from one physical system to another. Heat transfer is classified into various mechanisms, such as heat conduction, convection, thermal radiation, and phase-change transfer. Engineers also consider the transfer of mass of differing chemical species, either cold or hot, to achieve heat transfer. Sections include : Energy transfer by heat, work and mass Laws of thermodynamics Entropy Refrigeration Techniques Properties and nature of pure substances Applications Engineering : Predicting and analysing the performance of machines Thermodynamics Thermodynamics is the science of energy conversion involving heat and other forms of energy, most notably mechanical work. It studies and interrelates the macroscopic variables, such as temperature, volume and pressure, which describe physical, thermodynamic systems. Fluid mechanics Fluid Mechanics the study of the physical forces at work during fluid flow. Fluid mechanics can be divided into fluid kinematics, the study of fluid motion, and fluid kinetics, the study of the effect of forces on fluid motion. Fluid mechanics can further be divided into fluid statics, the study of fluids at rest, and fluid dynamics, the study of fluids in motion. Some of its more interesting concepts include momentum and reactive forces in fluid flow and fluid machinery theory and performance. Sections include: Flu Document 3::: Heat transfer is a discipline of thermal engineering that concerns the generation, use, conversion, and exchange of thermal energy (heat) between physical systems. Heat transfer is classified into various mechanisms, such as thermal conduction, thermal convection, thermal radiation, and transfer of energy by phase changes. Engineers also consider the transfer of mass of differing chemical species (mass transfer in the form of advection), either cold or hot, to achieve heat transfer. While these mechanisms have distinct characteristics, they often occur simultaneously in the same system. Heat conduction, also called diffusion, is the direct microscopic exchanges of kinetic energy of particles (such as molecules) or quasiparticles (such as lattice waves) through the boundary between two systems. When an object is at a different temperature from another body or its surroundings, heat flows so that the body and the surroundings reach the same temperature, at which point they are in thermal equilibrium. Such spontaneous heat transfer always occurs from a region of high temperature to another region of lower temperature, as described in the second law of thermodynamics. Heat convection occurs when the bulk flow of a fluid (gas or liquid) carries its heat through the fluid. All convective processes also move heat partly by diffusion, as well. The flow of fluid may be forced by external processes, or sometimes (in gravitational fields) by buoyancy forces caused when thermal energy expands the fluid (for example in a fire plume), thus influencing its own transfer. The latter process is often called "natural convection". The former process is often called "forced convection." In this case, the fluid is forced to flow by use of a pump, fan, or other mechanical means. Thermal radiation occurs through a vacuum or any transparent medium (solid or fluid or gas). It is the transfer of energy by means of photons or electromagnetic waves governed by the same laws. Overview Heat Document 4::: Boiling is the rapid phase transition from liquid to gas or vapor; the reverse of boiling is condensation. Boiling occurs when a liquid is heated to its boiling point, so that the vapour pressure of the liquid is equal to the pressure exerted on the liquid by the surrounding atmosphere. Boiling and evaporation are the two main forms of liquid vapourization. There are two main types of boiling: nucleate boiling where small bubbles of vapour form at discrete points, and critical heat flux boiling where the boiling surface is heated above a certain critical temperature and a film of vapour forms on the surface. Transition boiling is an intermediate, unstable form of boiling with elements of both types. The boiling point of water is 100 °C or 212 °F but is lower with the decreased atmospheric pressure found at higher altitudes. Boiling water is used as a method of making it potable by killing microbes and viruses that may be present. The sensitivity of different micro-organisms to heat varies, but if water is held at for one minute, most micro-organisms and viruses are inactivated. Ten minutes at a temperature of 70 °C (158 °F) is also sufficient to inactivate most bacteria. Boiling water is also used in several cooking methods including boiling, steaming, and poaching. Types Free convection The lowest heat flux seen in boiling is only sufficient to cause [natural convection], where the warmer fluid rises due to its slightly lower density. This condition occurs only when the superheat is very low, meaning that the hot surface near the fluid is nearly the same temperature as the boiling point. Nucleate Nucleate boiling is characterised by the growth of bubbles or pops on a heated surface (heterogeneous nucleation), which rises from discrete points on a surface, whose temperature is only slightly above the temperature of the liquid. In general, the number of nucleation sites is increased by an increasing surface temperature. An irregular surface of the boiling The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses. What do these two changes have in common? burning food on a stove water boiling on a stove A. Both are caused by heating. B. Both are chemical changes. C. Both are only physical changes. D. Both are caused by cooling. Answer:
sciq-2991
multiple_choice
Static friction acts on objects when they are resting on what?
[ "their laurels", "a pool of water", "a surface", "a single point" ]
C
Relavent Documents: Document 0::: Sliding is a type of motion between two surfaces in contact. This can be contrasted to rolling motion. Both types of motion may occur in bearings. The relative motion or tendency toward such motion between two surfaces is resisted by friction. Friction may damage or "wear" the surfaces in contact. However, wear can be reduced by lubrication. The science and technology of friction, lubrication, and wear is known as tribology. Sliding may occur between two objects of arbitrary shape, whereas rolling friction is the frictional force associated with the rotational movement of a somewhat disclike or other circular object along a surface. Generally, the frictional force of rolling friction is less than that associated with sliding kinetic friction. Typical values for the coefficient of rolling friction are less than that of sliding friction. Correspondingly sliding friction typically produces greater sound and thermal bi-products. One of the most common examples of sliding friction is the movement of braking motor vehicle tires on a roadway, a process which generates considerable heat and sound, and is typically taken into account in assessing the magnitude of roadway noise pollution. Sliding friction Sliding friction (also called kinetic friction) is a contact force that resists the sliding motion of two objects or an object and a surface. Sliding friction is almost always less than that of static friction; this is why it is easier to move an object once it starts moving rather than to get the object to begin moving from a rest position. Where , is the force of kinetic friction. is the coefficient of kinetic friction, and N is the normal force. Examples of sliding friction Sledding Pushing an object across a surface Rubbing one's hands together (The friction force generates heat.) A car sliding on ice A car skidding as it turns a corner Opening a window Almost any motion where there is contact between an object and a surface Falling down a bowling Document 1::: In mechanics, friction torque is the torque caused by the frictional force that occurs when two objects in contact move. Like all torques, it is a rotational force that may be measured in newton meters or pounds-feet. Engineering Friction torque can be disruptive in engineering. There are a variety of measures engineers may choose to take to eliminate these disruptions. Ball bearings are an example of an attempt to minimize the friction torque. Friction torque can also be an asset in engineering. Bolts and nuts, or screws are often designed to be fastened with a given amount of torque, where the friction is adequate during use or operation for the bolt, nut, or screw to remain safely fastened. This is true with such applications as lug nuts retaining wheels to vehicles, or equipment subjected to vibration with sufficiently well-attached bolts, nuts, or screws to prevent the vibration from shaking them loose. Examples When a cyclist applies the brake to the forward wheel, the bicycle tips forward due to the frictional torque between the wheel and the ground. When a golf ball hits the ground it begins to spin in part because of the friction torque applied to the golf ball from the friction between the golf ball and the ground. See also Torque Force Engineering Mechanics Moment (physics) Document 2::: Belt friction is a term describing the friction forces between a belt and a surface, such as a belt wrapped around a bollard. When a force applies a tension to one end of a belt or rope wrapped around a curved surface, the frictional force between the two surfaces increases with the amount of wrap about the curved surface, and only part of that force (or resultant belt tension) is transmitted to the other end of the belt or rope. Belt friction can be modeled by the Belt friction equation. In practice, the theoretical tension acting on the belt or rope calculated by the belt friction equation can be compared to the maximum tension the belt can support. This helps a designer of such a system determine how many times the belt or rope must be wrapped around a curved surface to prevent it from slipping. Mountain climbers and sailing crews demonstrate a working knowledge of belt friction when accomplishing tasks with ropes, pulleys, bollards and capstans. Equation The equation used to model belt friction is, assuming the belt has no mass and its material is a fixed composition: where is the tension of the pulling side, is the tension of the resisting side, is the static friction coefficient, which has no units, and is the angle, in radians, formed by the first and last spots the belt touches the pulley, with the vertex at the center of the pulley. The tension on the pulling side of the belt and pulley has the ability to increase exponentially if the magnitude of the belt angle increases (e.g. it is wrapped around the pulley segment numerous times). Generalization for a rope lying on an arbitrary orthotropic surface If a rope is laying in equilibrium under tangential forces on a rough orthotropic surface then three following conditions (all of them) are satisfied: 1. No separation – normal reaction is positive for all points of the rope curve: , where is a normal curvature of the rope curve. 2. Dragging coefficient of friction and angle are satisfying Document 3::: For a rigid object in contact with a fixed environment and acted upon by gravity in the vertical direction, its support polygon is a horizontal region over which the center of mass must lie to achieve static stability. For example, for an object resting on a horizontal surface (e.g. a table), the support polygon is the convex hull of its "footprint" on the table. The support polygon succinctly represents the conditions necessary for an object to be at equilibrium under gravity. That is, if the object's center of mass lies over the support polygon, then there exist a set of forces over the region of contact that exactly counteracts the forces of gravity. Note that this is a necessary condition for stability, but not a sufficient one. Derivation Let the object be in contact at a finite number of points . At each point , let be the set of forces that can be applied on the object at that point. Here, is known as the friction cone, and for the Coulomb model of friction, is actually a cone with apex at the origin, extending to infinity in the normal direction of the contact. Let be the (unspecified) forces at the contact points. To balance the object in static equilibrium, the following Newton-Euler equations must be met on : for all where is the force of gravity on the object, and is its center of mass. The first two equations are the Newton-Euler equations, and the third requires all forces to be valid. If there is no set of forces that meet all these conditions, the object will not be in equilibrium. The second equation has no dependence on the vertical component of the center of mass, and thus if a solution exists for one , the same solution works for all . Therefore, the set of all that have solutions to the above conditions is a set that extends infinitely in the up and down directions. The support polygon is simply the projection of this set on the horizontal plane. These results can easily be extended to different friction models and Document 4::: The frictionless plane is a concept from the writings of Galileo Galilei. In his 1638 The Two New Sciences, Galileo presented a formula that predicted the motion of an object moving down an inclined plane. His formula was based upon his past experimentation with free-falling bodies. However, his model was not based upon experimentation with objects moving down an inclined plane, but from his conceptual modeling of the forces acting upon the object. Galileo understood the mechanics of the inclined plane as the combination of horizontal and vertical vectors; the result of gravity acting upon the object, diverted by the slope of the plane. However, Galileo's equations do not contemplate friction, and therefore do not perfectly predict the results of an actual experiment. This is because some energy is always lost when one mass applies a non-zero normal force to another. Therefore, the observed speed, acceleration and distance traveled should be less than Galileo predicts. This energy is lost in forms like sound and heat. However, from Galileo's predictions of an object moving down an inclined plane in a frictionless environment, he created the theoretical foundation for extremely fruitful real-world experimental prediction. Frictionless planes do not exist in the real world. However, if they did, one can be almost certain that objects on them would behave exactly as Galileo predicts. Despite their nonexistence, they have considerable value in the design of engines, motors, roadways, and even tow-truck beds, to name a few examples. The effect of friction on an object moving down an inclined plane can be calculated as where is the force of friction exerted by the object and the inclined plane on each other, parallel to the surface of the plane, is the normal force exerted by the object and the plane on each other, directed perpendicular to the plane, and is the coefficient of kinetic friction. Unless the inclined plane is in a vacuum, a (usually) small amount of The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses. Static friction acts on objects when they are resting on what? A. their laurels B. a pool of water C. a surface D. a single point Answer: