id
stringlengths 6
15
| question_type
stringclasses 1
value | question
stringlengths 15
683
| choices
listlengths 4
4
| answer
stringclasses 5
values | explanation
stringclasses 481
values | prompt
stringlengths 1.75k
10.9k
|
---|---|---|---|---|---|---|
sciq-2264
|
multiple_choice
|
What is a capillary bed that filters blood principally based on particle size?
|
[
"plexus",
"buds",
"glomerulus",
"vein"
] |
C
|
Relavent Documents:
Document 0:::
The endothelium (: endothelia) is a single layer of squamous endothelial cells that line the interior surface of blood vessels and lymphatic vessels. The endothelium forms an interface between circulating blood or lymph in the lumen and the rest of the vessel wall. Endothelial cells form the barrier between vessels and tissue and control the flow of substances and fluid into and out of a tissue.
Endothelial cells in direct contact with blood are called vascular endothelial cells whereas those in direct contact with lymph are known as lymphatic endothelial cells. Vascular endothelial cells line the entire circulatory system, from the heart to the smallest capillaries.
These cells have unique functions that include fluid filtration, such as in the glomerulus of the kidney, blood vessel tone, hemostasis, neutrophil recruitment, and hormone trafficking. Endothelium of the interior surfaces of the heart chambers is called endocardium. An impaired function can lead to serious health issues throughout the body.
Structure
The endothelium is a thin layer of single flat (squamous) cells that line the interior surface of blood vessels and lymphatic vessels.
Endothelium is of mesodermal origin. Both blood and lymphatic capillaries are composed of a single layer of endothelial cells called a monolayer. In straight sections of a blood vessel, vascular endothelial cells typically align and elongate in the direction of fluid flow.
Terminology
The foundational model of anatomy, an index of terms used to describe anatomical structures, makes a distinction between endothelial cells and epithelial cells on the basis of which tissues they develop from, and states that the presence of vimentin rather than keratin filaments separates these from epithelial cells. Many considered the endothelium a specialized epithelial tissue.
Function
The endothelium forms an interface between circulating blood or lymph in the lumen and the rest of the vessel wall. This forms a barrier between v
Document 1:::
Anatomical terminology is used to describe microanatomical (or histological) structures. This helps describe precisely the structure, layout and position of an object, and minimises ambiguity. An internationally accepted lexicon is Terminologia Histologica.
Layout
Epithelia and endothelia
Epithelial cells line body surfaces, and are described according to their shape, with three principal shapes: squamous, columnar, and cuboidal.
Squamous epithelium has cells that are wider than their height (flat and scale-like).
Cuboidal epithelium has cells whose height and width are approximately the same (cube shaped).
Columnar epithelium has cells taller than they are wide (column-shaped).
Endothelium refers to cells that line the interior surface of blood vessels and lymphatic vessels, forming an interface between circulating blood or lymph in the lumen and the rest of the vessel wall. It is a thin layer of simple, or single-layered, squamous cells called endothelial cells. Endothelial cells in direct contact with blood are called vascular endothelial cells, whereas those in direct contact with lymph are known as lymphatic endothelial cells.
Epithelium can be arranged in a single layer of cells described as "simple", or more than one layer, described as "stratified". By layer, epithelium is classed as either simple epithelium, only one cell thick (unilayered) or stratified epithelium as stratified squamous epithelium, stratified cuboidal epithelium, and stratified columnar epithelium that are two or more cells thick (multi-layered), and both types of layering can be made up of any of the cell shapes. However, when taller simple columnar epithelial cells are viewed in cross section showing several nuclei appearing at different heights, they can be confused with stratified epithelia. This kind of epithelium is therefore described as pseudostratified columnar epithelium.
Transitional epithelium has cells that can change from squamous to cuboidal, depending on the amoun
Document 2:::
The blood circulatory system is a system of organs that includes the heart, blood vessels, and blood which is circulated throughout the entire body of a human or other vertebrate. It includes the cardiovascular system, or vascular system, that consists of the heart and blood vessels (from Greek kardia meaning heart, and from Latin vascula meaning vessels). The circulatory system has two divisions, a systemic circulation or circuit, and a pulmonary circulation or circuit. Some sources use the terms cardiovascular system and vascular system interchangeably with the circulatory system.
The network of blood vessels are the great vessels of the heart including large elastic arteries, and large veins; other arteries, smaller arterioles, capillaries that join with venules (small veins), and other veins. The circulatory system is closed in vertebrates, which means that the blood never leaves the network of blood vessels. Some invertebrates such as arthropods have an open circulatory system. Diploblasts such as sponges, and comb jellies lack a circulatory system.
Blood is a fluid consisting of plasma, red blood cells, white blood cells, and platelets; it is circulated around the body carrying oxygen and nutrients to the tissues and collecting and disposing of waste materials. Circulated nutrients include proteins and minerals and other components include hemoglobin, hormones, and gases such as oxygen and carbon dioxide. These substances provide nourishment, help the immune system to fight diseases, and help maintain homeostasis by stabilizing temperature and natural pH.
In vertebrates, the lymphatic system is complementary to the circulatory system. The lymphatic system carries excess plasma (filtered from the circulatory system capillaries as interstitial fluid between cells) away from the body tissues via accessory routes that return excess fluid back to blood circulation as lymph. The lymphatic system is a subsystem that is essential for the functioning of the bloo
Document 3:::
High endothelial venules (HEV) are specialized post-capillary venules characterized by plump endothelial cells as opposed to the usual flatter endothelial cells found in regular venules. HEVs enable lymphocytes circulating in the blood to directly enter a lymph node (by crossing through the HEV).
In humans, HEVs are found in all secondary lymphoid organs (with the exception of spleen, where blood exits through open arterioles and enters the red pulp), including hundreds of lymph nodes dispersed in the body, tonsils and adenoids in the pharynx, Peyer's patches (PIs) in the small intestine, appendix, and small aggregates of lymphoid tissue in the stomach and large intestine. In contrast to the endothelial cells from other vessels, the high endothelial cells of HEVs have a distinctive appearance, consisting of a cuboidal morphology and with various receptors to interact with leukocytes (express specialized ligands for lymphocytes and are able to support high levels of lymphocyte extravasation). HEVs enable naïve lymphocytes to move in and out of the lymph nodes from the circulatory system. HEV cells express addressins, which are specific adhesion molecules that attach to the L-selectins on lymphocytes and anchor them to the HEV wall in preparation for crossing the endothelium.
The endothelial cells of HEVs have a 'plump' appearance different from the flat morphology of endothelial cells that line other vessels, and are therefore called high endothelial cells by reference to their thickness. Another characteristic of HEVs, revealed by light-microscopic examination, is the presence of a large number of lymphocytes within their walls. This illustrates the function of HEVs in lymphocyte recruitment and explains why these vessels were implicated in lymphocyte traffic from the time of their initial description.
The need for HEV
In order to have an adaptive immune response occur, T cells need to be activated. T cells become activated by recognising foreign antigens bound
Document 4:::
Vascular recruitment is the increase in the number of perfused capillaries in response to a stimulus. I.e., the more you exercise regularly, the more oxygen can reach your muscles.
Vascular recruitment may also be called capillary recruitment.
Vascular recruitment in skeletal muscle
The term «vascular recruitment» or «capillary recruitment» usually refers to the increase in the number perfused capillaries in skeletal muscle in response to a stimulus. The most important stimulus in humans is regular exercise. Vascular recruitment in skeletal muscle is thought to enhance the capillary surface area for oxygen exchange and decrease the oxygen diffusion distance.
Other stimuli are possible. Insulin can act as a stimulus for vascular recruitment in skeletal muscle. This process may also improve glucose delivery to skeletal muscle by increasing the surface area for diffusion. That insulin can act in this way has been proposed based on increases in limb blood flow and skeletal muscle blood volume which occurred after hyperinsulinemia.
The exact extent of capillary recruitment in intact skeletal muscle in response to regular exercise or insulin is unknown, because non-invasive measurement techniques are not yet extremely precise.
Being overweight or obese may negatively interfere with vascular recruitment in skeletal muscle.
Vascular recruitment in the lung
Vascular recruitment in the lung (i.e., in the pulmonary microcirculation) may be noteworthy to healthcare professionals in emergency medicine, because it may increase evidence of lung injury, and increase pulmonary capillary protein leak.
Vascular recruitment in the brain
Vascular recruitment in the brain is thought to lead to new capillaries and increase the cerebral blood flow.
Controversy
The existence of vascular recruitment in response to a stimulus has been disputed by some researchers. However, most researchers accept that vascular recruitment exists.
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is a capillary bed that filters blood principally based on particle size?
A. plexus
B. buds
C. glomerulus
D. vein
Answer:
|
|
sciq-3935
|
multiple_choice
|
What is the term for coiled structures made of dna and proteins?
|
[
"platelets",
"neutrons",
"nodes",
"chromosomes"
] |
D
|
Relavent Documents:
Document 0:::
Cellular components are the complex biomolecules and structures of which cells, and thus living organisms, are composed. Cells are the structural and functional units of life. The smallest organisms are single cells, while the largest organisms are assemblages of trillions of cells. DNA, double stranded macromolecule that carries the hereditary information of the cell and found in all living cells; each cell carries chromosome(s) having a distinctive DNA sequence.
Examples include macromolecules such as proteins and nucleic acids, biomolecular complexes such as a ribosome, and structures such as membranes, and organelles. While the majority of cellular components are located within the cell itself, some may exist in extracellular areas of an organism.
Cellular components may also be called biological matter or biological material. Most biological matter has the characteristics of soft matter, being governed by relatively small energies. All known life is made of biological matter. To be differentiated from other theoretical or fictional life forms, such life may be called carbon-based, cellular, organic, biological, or even simply living – as some definitions of life exclude hypothetical types of biochemistry.
See also
Cell (biology)
Cell biology
Biomolecule
Organelle
Tissue (biology)
External links
https://web.archive.org/web/20130918033010/http://bioserv.fiu.edu/~walterm/FallSpring/review1_fall05_chap_cell3.htm
Document 1:::
Biomolecular structure is the intricate folded, three-dimensional shape that is formed by a molecule of protein, DNA, or RNA, and that is important to its function. The structure of these molecules may be considered at any of several length scales ranging from the level of individual atoms to the relationships among entire protein subunits. This useful distinction among scales is often expressed as a decomposition of molecular structure into four levels: primary, secondary, tertiary, and quaternary. The scaffold for this multiscale organization of the molecule arises at the secondary level, where the fundamental structural elements are the molecule's various hydrogen bonds. This leads to several recognizable domains of protein structure and nucleic acid structure, including such secondary-structure features as alpha helixes and beta sheets for proteins, and hairpin loops, bulges, and internal loops for nucleic acids.
The terms primary, secondary, tertiary, and quaternary structure were introduced by Kaj Ulrik Linderstrøm-Lang in his 1951 Lane Medical Lectures at Stanford University.
Primary structure
The primary structure of a biopolymer is the exact specification of its atomic composition and the chemical bonds connecting those atoms (including stereochemistry). For a typical unbranched, un-crosslinked biopolymer (such as a molecule of a typical intracellular protein, or of DNA or RNA), the primary structure is equivalent to specifying the sequence of its monomeric subunits, such as amino acids or nucleotides.
The primary structure of a protein is reported starting from the amino N-terminus to the carboxyl C-terminus, while the primary structure of DNA or RNA molecule is known as the nucleic acid sequence reported from the 5' end to the 3' end.
The nucleic acid sequence refers to the exact sequence of nucleotides that comprise the whole molecule. Often, the primary structure encodes sequence motifs that are of functional importance. Some examples of such motif
Document 2:::
In cell biology, microtrabeculae were a hypothesised fourth element of the cytoskeleton (the other three being microfilaments, microtubules and intermediate filaments), proposed by Keith Porter based on images obtained from high-voltage electron microscopy of whole cells in the 1970s. The images showed short, filamentous structures of unknown molecular composition associated with known cytoplasmic structures. It is now generally accepted that microtrabeculae are nothing more than an artifact of certain types of fixation treatment, although the complexity of the cell's cytoskeleton is not yet fully understood.
Document 3:::
The cell is the basic structural and functional unit of all forms of life. Every cell consists of cytoplasm enclosed within a membrane, and contains many macromolecules such as proteins, DNA and RNA, as well as many small molecules of nutrients and metabolites. The term comes from the Latin word meaning 'small room'.
Cells can acquire specified function and carry out various tasks within the cell such as replication, DNA repair, protein synthesis, and motility. Cells are capable of specialization and mobility within the cell.
Most plant and animal cells are only visible under a light microscope, with dimensions between 1 and 100 micrometres. Electron microscopy gives a much higher resolution showing greatly detailed cell structure. Organisms can be classified as unicellular (consisting of a single cell such as bacteria) or multicellular (including plants and animals). Most unicellular organisms are classed as microorganisms.
The study of cells and how they work has led to many other studies in related areas of biology, including: discovery of DNA, cancer systems biology, aging and developmental biology.
Cell biology is the study of cells, which were discovered by Robert Hooke in 1665, who named them for their resemblance to cells inhabited by Christian monks in a monastery. Cell theory, first developed in 1839 by Matthias Jakob Schleiden and Theodor Schwann, states that all organisms are composed of one or more cells, that cells are the fundamental unit of structure and function in all living organisms, and that all cells come from pre-existing cells. Cells emerged on Earth about 4 billion years ago.
Discovery
With continual improvements made to microscopes over time, magnification technology became advanced enough to discover cells. This discovery is largely attributed to Robert Hooke, and began the scientific study of cells, known as cell biology. When observing a piece of cork under the scope, he was able to see pores. This was shocking at the time as i
Document 4:::
Cell physiology is the biological study of the activities that take place in a cell to keep it alive. The term physiology refers to normal functions in a living organism. Animal cells, plant cells and microorganism cells show similarities in their functions even though they vary in structure.
General characteristics
There are two types of cells: prokaryotes and eukaryotes.
Prokaryotes were the first of the two to develop and do not have a self-contained nucleus. Their mechanisms are simpler than later-evolved eukaryotes, which contain a nucleus that envelops the cell's DNA and some organelles.
Prokaryotes
Prokaryotes have DNA located in an area called the nucleoid, which is not separated from other parts of the cell by a membrane. There are two domains of prokaryotes: bacteria and archaea. Prokaryotes have fewer organelles than eukaryotes. Both have plasma membranes and ribosomes (structures that synthesize proteins and float free in cytoplasm). Two unique characteristics of prokaryotes are fimbriae (finger-like projections on the surface of a cell) and flagella (threadlike structures that aid movement).
Eukaryotes
Eukaryotes have a nucleus where DNA is contained. They are usually larger than prokaryotes and contain many more organelles. The nucleus, the feature of a eukaryote that distinguishes it from a prokaryote, contains a nuclear envelope, nucleolus and chromatin. In cytoplasm, endoplasmic reticulum (ER) synthesizes membranes and performs other metabolic activities. There are two types, rough ER (containing ribosomes) and smooth ER (lacking ribosomes). The Golgi apparatus consists of multiple membranous sacs, responsible for manufacturing and shipping out materials such as proteins. Lysosomes are structures that use enzymes to break down substances through phagocytosis, a process that comprises endocytosis and exocytosis. In the mitochondria, metabolic processes such as cellular respiration occur. The cytoskeleton is made of fibers that support the str
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is the term for coiled structures made of dna and proteins?
A. platelets
B. neutrons
C. nodes
D. chromosomes
Answer:
|
|
sciq-3056
|
multiple_choice
|
After a massive eruption what will a mountain collapsing above an empty chamber create?
|
[
"crater",
"caldera",
"valley",
"cavern"
] |
B
|
Relavent Documents:
Document 0:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 1:::
The mid-24th century BCE climate anomaly is the period, between 2354–2345 BCE, of consistently, reduced annual temperatures that are reconstructed from consecutive abnormally narrow, Irish oak tree rings. These tree rings are indicative of a period of catastrophically reduced growth in Irish trees during that period. This range of dates also matches the transition from the Neolithic to the Bronze Age in the British Isles and a period of widespread societal collapse in the Near East. It has been proposed that this anomalous downturn in the climate might have been the result of comet debris suspended in the atmosphere.
In 1997, Marie-Agnès Courty proposed that a natural disaster involving wildfires, floods, and an air blast of over 100 megatons power occurred about 2350 BCE. This proposal is based on unusual "dust" deposits which have been reported from archaeological sites in Mesopotamia that are a few hundred kilometres from each other. In later papers, Courty subsequently revised the date of this event from 2350 BCE to 2000 BCE.
Based only upon the analysis of satellite imagery, Umm al Binni lake in southern Iraq has been suggested as a possible extraterrestrial impact crater and possible cause of this natural disaster. More recent sources have argued for a formation of the lake through the subsidence of the underlying basement fault blocks. Baillie and McAneney's 2015 discussion of this climate anomaly discusses its abnormally narrow Irish tree rings and the anomalous dust deposits of Courty. However, this paper lacks any mention of Umm al Binni lake.
See also
4.2-kiloyear event, c. 2200 BCE
Great Flood (China), c. 2300 BCE
Document 2:::
In mathematical psychology and education theory, a knowledge space is a combinatorial structure used to formulate mathematical models describing the progression of a human learner. Knowledge spaces were introduced in 1985 by Jean-Paul Doignon and Jean-Claude Falmagne, and remain in extensive use in the education theory. Modern applications include two computerized tutoring systems, ALEKS and the defunct RATH.
Formally, a knowledge space assumes that a domain of knowledge is a collection of concepts or skills, each of which must be eventually mastered. Not all concepts are interchangeable; some require other concepts as prerequisites. Conversely, competency at one skill may ease the acquisition of another through similarity. A knowledge space marks out which collections of skills are feasible: they can be learned without mastering any other skills. Under reasonable assumptions, the collection of feasible competencies forms the mathematical structure known as an antimatroid.
Researchers and educators usually explore the structure of a discipline's knowledge space as a latent class model.
Motivation
Knowledge Space Theory attempts to address shortcomings of standardized testing when used in educational psychometry. Common tests, such as the SAT and ACT, compress a student's knowledge into a very small range of ordinal ranks, in the process effacing the conceptual dependencies between questions. Consequently, the tests cannot distinguish between true understanding and guesses, nor can they identify a student's particular weaknesses, only the general proportion of skills mastered. The goal of knowledge space theory is to provide a language by which exams can communicate
What the student can do and
What the student is ready to learn.
Model structure
Knowledge Space Theory-based models presume that an educational subject can be modeled as a finite set of concepts, skills, or topics. Each feasible state of knowledge about is then a subset of ; the set of
Document 3:::
Maui Nui is a modern geologists' name given to a prehistoric Hawaiian island and the corresponding modern biogeographic region. Maui Nui is composed of four modern islands: Maui, Molokaʻi, Lānaʻi, and Kahoʻolawe. Administratively, the four modern islands comprise Maui County (and a tiny part of Molokaʻi called Kalawao County). Long after the breakup of Maui Nui, the four modern islands retained plant and animal life similar to each other. Thus, Maui Nui is not only a prehistoric island but also a modern biogeographic region.
Geology
Maui Nui formed and broke up during the Pleistocene Epoch, which lasted from about 2.58 million to 11,700 years ago.
Maui Nui is built from seven shield volcanoes. The three oldest are Penguin Bank, West Molokaʻi, and East Molokaʻi, which probably range from slightly over to slightly less than 2 million years old. The four younger volcanoes are Lāna‘i, West Maui, Kaho‘olawe, and Haleakalā, which probably formed between 1.5 and 2 million years ago.
At its prime 1.2 million years ago, Maui Nui was , 50% larger than today's Hawaiʻi Island. The island of Maui Nui included four modern islands (Maui, Molokaʻi, Lānaʻi, and Kahoʻolawe) and landmass west of Molokaʻi called Penguin Bank, which is now completely submerged.
Maui Nui broke up as rising sea levels flooded the connections between the volcanoes. The breakup was complex because global sea levels rose and fell intermittently during the Quaternary glaciation. About 600,000 years ago, the connection between Molokaʻi and the island of Lāna‘i/Maui/Kahoʻolawe became intermittent. About 400,000 years ago, the connection between Lāna‘i and Maui/Kahoʻolawe also became intermittent. The connection between Maui and Kahoʻolawe was permanently broken between 200,000 and 150,000 years ago. Maui, Lāna‘i, and Molokaʻi were connected intermittently thereafter, most recently about 18,000 years ago during the Last Glacial Maximum.
Today, the sea floor between these four islands is relatively shallow
Document 4:::
Sand boils or sand volcanoes occur when water under pressure wells up through a bed of sand. The water looks like it is boiling up from the bed of sand, hence the name.
Sand volcano
A sand volcano or sand blow is a cone of sand formed by the ejection of sand onto a surface from a central point. The sand builds up as a cone with slopes at the sand's angle of repose. A crater is commonly seen at the summit. The cone looks like a small volcanic cone and can range in size from millimetres to metres in diameter.
The process is often associated with soil liquefaction and the ejection of fluidized sand that can occur in water-saturated sediments during an earthquake. The New Madrid Seismic Zone exhibited many such features during the 1811–12 New Madrid earthquakes. Adjacent sand blows aligned in a row along a linear fracture within fine-grained surface sediments are just as common, and can still be seen in the New Madrid area.
In the past few years, much effort has gone into the mapping of liquefaction features to study ancient earthquakes. The basic idea is to map zones that are susceptible to the process and then go in for a closer look. The presence or absence of soil liquefaction features is strong evidence of past earthquake activity, or lack thereof.
These are to be contrasted with mud volcanoes, which occur in areas of geyser or subsurface gas venting.
Flood protection structures
Sand boils can be a mechanism contributing to liquefaction and levee failure during floods. This effect is caused by a difference in pressure on two sides of a levee or dike, most likely during a flood. This process can result in internal erosion, whereby the removal of soil particles results in a pipe through the embankment. The creation of the pipe will quickly pick up pace and will eventually result in failure of the embankment.
A sand boil is difficult to stop. The most effective method is by creating a body of water above the boil to create enough pressure to slow the flow of
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
After a massive eruption what will a mountain collapsing above an empty chamber create?
A. crater
B. caldera
C. valley
D. cavern
Answer:
|
|
sciq-929
|
multiple_choice
|
What is a circuit that consists of one loop called?
|
[
"speaking circuit",
"dramatic circuit",
"series circuit",
"episode circuit"
] |
C
|
Relavent Documents:
Document 0:::
Mathematical methods are integral to the study of electronics.
Mathematics in electronics
Electronics engineering careers usually include courses in calculus (single and multivariable), complex analysis, differential equations (both ordinary and partial), linear algebra and probability. Fourier analysis and Z-transforms are also subjects which are usually included in electrical engineering programs. Laplace transform can simplify computing RLC circuit behaviour.
Basic applications
A number of electrical laws apply to all electrical networks. These include
Faraday's law of induction: Any change in the magnetic environment of a coil of wire will cause a voltage (emf) to be "induced" in the coil.
Gauss's Law: The total of the electric flux out of a closed surface is equal to the charge enclosed divided by the permittivity.
Kirchhoff's current law: the sum of all currents entering a node is equal to the sum of all currents leaving the node or the sum of total current at a junction is zero
Kirchhoff's voltage law: the directed sum of the electrical potential differences around a circuit must be zero.
Ohm's law: the voltage across a resistor is the product of its resistance and the current flowing through it.at constant temperature.
Norton's theorem: any two-terminal collection of voltage sources and resistors is electrically equivalent to an ideal current source in parallel with a single resistor.
Thévenin's theorem: any two-terminal combination of voltage sources and resistors is electrically equivalent to a single voltage source in series with a single resistor.
Millman's theorem: the voltage on the ends of branches in parallel is equal to the sum of the currents flowing in every branch divided by the total equivalent conductance.
See also Analysis of resistive circuits.
Circuit analysis is the study of methods to solve linear systems for an unknown variable.
Circuit analysis
Components
There are many electronic components currently used and they all have thei
Document 1:::
The circuit topology of an electronic circuit is the form taken by the network of interconnections of the circuit components. Different specific values or ratings of the components are regarded as being the same topology. Topology is not concerned with the physical layout of components in a circuit, nor with their positions on a circuit diagram; similarly to the mathematical concept of topology, it is only concerned with what connections exist between the components. There may be numerous physical layouts and circuit diagrams that all amount to the same topology.
Strictly speaking, replacing a component with one of an entirely different type is still the same topology. In some contexts, however, these can loosely be described as different topologies. For instance, interchanging inductors and capacitors in a low-pass filter results in a high-pass filter. These might be described as high-pass and low-pass topologies even though the network topology is identical. A more correct term for these classes of object (that is, a network where the type of component is specified but not the absolute value) is prototype network.
Electronic network topology is related to mathematical topology. In particular, for networks which contain only two-terminal devices, circuit topology can be viewed as an application of graph theory. In a network analysis of such a circuit from a topological point of view, the network nodes are the vertices of graph theory, and the network branches are the edges of graph theory.
Standard graph theory can be extended to deal with active components and multi-terminal devices such as integrated circuits. Graphs can also be used in the analysis of infinite networks.
Circuit diagrams
The circuit diagrams in this article follow the usual conventions in electronics; lines represent conductors, filled small circles represent junctions of conductors, and open small circles represent terminals for connection to the outside world. In most cases, imped
Document 2:::
In electrical engineering, electrical terms are associated into pairs called duals. A dual of a relationship is formed by interchanging voltage and current in an expression. The dual expression thus produced is of the same form, and the reason that the dual is always a valid statement can be traced to the duality of electricity and magnetism.
Here is a partial list of electrical dualities:
voltage – current
parallel – serial (circuits)
resistance – conductance
voltage division – current division
impedance – admittance
capacitance – inductance
reactance – susceptance
short circuit – open circuit
Kirchhoff's current law – Kirchhoff's voltage law.
Thévenin's theorem – Norton's theorem
History
The use of duality in circuit theory is due to Alexander Russell who published his ideas in 1904.
Examples
Constitutive relations
Resistor and conductor (Ohm's law)
Capacitor and inductor – differential form
Capacitor and inductor – integral form
Voltage division — current division
Impedance and admittance
Resistor and conductor
Capacitor and inductor
See also
Duality (electricity and magnetism)
Duality (mechanical engineering)
Dual impedance
Dual graph
Mechanical–electrical analogies
List of dualities
Document 3:::
A circuit diagram (or: wiring diagram, electrical diagram, elementary diagram, electronic schematic) is a graphical representation of an electrical circuit. A pictorial circuit diagram uses simple images of components, while a schematic diagram shows the components and interconnections of the circuit using standardized symbolic representations. The presentation of the interconnections between circuit components in the schematic diagram does not necessarily correspond to the physical arrangements in the finished device.
Unlike a block diagram or layout diagram, a circuit diagram shows the actual electrical connections. A drawing meant to depict the physical arrangement of the wires and the components they connect is called artwork or layout, physical design, or wiring diagram.
Circuit diagrams are used for the design (circuit design), construction (such as PCB layout), and maintenance of electrical and electronic equipment.
In computer science, circuit diagrams are useful when visualizing expressions using Boolean algebra.
Symbols
Circuit diagrams are pictures with symbols that have differed from country to country and have changed over time, but are now to a large extent internationally standardized. Simple components often had symbols intended to represent some feature of the physical construction of the device. For example, the symbol for a resistor dates back to the time when that component was made from a long piece of wire wrapped in such a manner as to not produce inductance, which would have made it a coil. These wirewound resistors are now used only in high-power applications, smaller resistors being cast from carbon composition (a mixture of carbon and filler) or fabricated as an insulating tube or chip coated with a metal film. The internationally standardized symbol for a resistor is therefore now simplified to an oblong, sometimes with the value in ohms written inside, instead of the zig-zag symbol. A less common symbol is simply a series of peaks o
Document 4:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is a circuit that consists of one loop called?
A. speaking circuit
B. dramatic circuit
C. series circuit
D. episode circuit
Answer:
|
|
ai2_arc-1002
|
multiple_choice
|
A teacher builds a model of a hydrogen atom. A red golf ball is used for a proton, and a green golf ball is used for an electron. Which is not accurate concerning the model?
|
[
"number of particles",
"relative mass of particles",
"types of particles present",
"charges of particles present"
] |
B
|
Relavent Documents:
Document 0:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 1:::
In mathematical psychology and education theory, a knowledge space is a combinatorial structure used to formulate mathematical models describing the progression of a human learner. Knowledge spaces were introduced in 1985 by Jean-Paul Doignon and Jean-Claude Falmagne, and remain in extensive use in the education theory. Modern applications include two computerized tutoring systems, ALEKS and the defunct RATH.
Formally, a knowledge space assumes that a domain of knowledge is a collection of concepts or skills, each of which must be eventually mastered. Not all concepts are interchangeable; some require other concepts as prerequisites. Conversely, competency at one skill may ease the acquisition of another through similarity. A knowledge space marks out which collections of skills are feasible: they can be learned without mastering any other skills. Under reasonable assumptions, the collection of feasible competencies forms the mathematical structure known as an antimatroid.
Researchers and educators usually explore the structure of a discipline's knowledge space as a latent class model.
Motivation
Knowledge Space Theory attempts to address shortcomings of standardized testing when used in educational psychometry. Common tests, such as the SAT and ACT, compress a student's knowledge into a very small range of ordinal ranks, in the process effacing the conceptual dependencies between questions. Consequently, the tests cannot distinguish between true understanding and guesses, nor can they identify a student's particular weaknesses, only the general proportion of skills mastered. The goal of knowledge space theory is to provide a language by which exams can communicate
What the student can do and
What the student is ready to learn.
Model structure
Knowledge Space Theory-based models presume that an educational subject can be modeled as a finite set of concepts, skills, or topics. Each feasible state of knowledge about is then a subset of ; the set of
Document 2:::
The SAT Subject Test in Biology was the name of a one-hour multiple choice test given on biology by the College Board. A student chose whether to take the test depending upon college entrance requirements for the schools in which the student is planning to apply. Until 1994, the SAT Subject Tests were known as Achievement Tests; and from 1995 until January 2005, they were known as SAT IIs. Of all SAT subject tests, the Biology E/M test was the only SAT II that allowed the test taker a choice between the ecological or molecular tests. A set of 60 questions was taken by all test takers for Biology and a choice of 20 questions was allowed between either the E or M tests. This test was graded on a scale between 200 and 800. The average for Molecular is 630 while Ecological is 591.
On January 19 2021, the College Board discontinued all SAT Subject tests, including the SAT Subject Test in Biology E/M. This was effective immediately in the United States, and the tests were to be phased out by the following summer for international students. This was done as a response to changes in college admissions due to the impact of the COVID-19 pandemic on education.
Format
This test had 80 multiple-choice questions that were to be answered in one hour. All questions had five answer choices. Students received one point for each correct answer, lost ¼ of a point for each incorrect answer, and received 0 points for questions left blank. The student's score was based entirely on his or her performance in answering the multiple-choice questions.
The questions covered a broad range of topics in general biology. There were more specific questions related respectively on ecological concepts (such as population studies and general Ecology) on the E test and molecular concepts such as DNA structure, translation, and biochemistry on the M test.
Preparation
The College Board suggested a year-long course in biology at the college preparatory level, as well as a one-year course in algebra, a
Document 3:::
Advanced Placement (AP) Physics C: Electricity and Magnetism (also known as AP Physics C: E&M or AP E&M) is an introductory physics course administered by the College Board as part of its Advanced Placement program. It is intended to proxy a second-semester calculus-based university course in electricity and magnetism. The content of Physics C: E&M overlaps with that of AP Physics 2, but Physics 2 is algebra-based and covers other topics outside of electromagnetism, while Physics C is calculus-based and only covers electromagnetism. Physics C: E&M may be combined with its mechanics counterpart to form a year-long course that prepares for both exams.
Course content
E&M is equivalent to an introductory college course in electricity and magnetism for physics or engineering majors. The course modules are:
Electrostatics
Conductors, capacitors, and dielectrics
Electric circuits
Magnetic fields
Electromagnetism.
Methods of calculus are used wherever appropriate in formulating physical principles and in applying them to physical problems. Therefore, students should have completed or be concurrently enrolled in a calculus class.
AP test
The course culminates in an optional exam for which high-performing students may receive some credit towards their college coursework, depending on the institution.
Registration
The AP examination for AP Physics C: Electricity and Magnetism is separate from the AP examination for AP Physics C: Mechanics. Before 2006, test-takers paid only once and were given the choice of taking either one or two parts of the Physics C test.
Format
The exam is typically administered on a Monday afternoon in May. The exam is configured in two categories: a 35-question multiple choice section and a 3-question free response section. Test takers are allowed to use an approved calculator during the entire exam. The test is weighted such that each section is worth half of the final score. This and AP Physics C: Mechanics are the shortest AP exams, with
Document 4:::
Advanced Placement (AP) Physics C: Mechanics (also known as AP Mechanics) is an introductory physics course administered by the College Board as part of its Advanced Placement program. It is intended to proxy a one-semester calculus-based university course in mechanics. The content of Physics C: Mechanics overlaps with that of AP Physics 1, but Physics 1 is algebra-based, while Physics C is calculus-based. Physics C: Mechanics may be combined with its electricity and magnetism counterpart to form a year-long course that prepares for both exams.
Course content
Intended to be equivalent to an introductory college course in mechanics for physics or engineering majors, the course modules are:
Kinematics
Newton's laws of motion
Work, energy and power
Systems of particles and linear momentum
Circular motion and rotation
Oscillations and gravitation.
Methods of calculus are used wherever appropriate in formulating physical principles and in applying them to physical problems. Therefore, students should have completed or be concurrently enrolled in a Calculus I class.
This course is often compared to AP Physics 1: Algebra Based for its similar course material involving kinematics, work, motion, forces, rotation, and oscillations. However, AP Physics 1: Algebra Based lacks concepts found in Calculus I, like derivatives or integrals.
This course may be combined with AP Physics C: Electricity and Magnetism to make a unified Physics C course that prepares for both exams.
AP test
The course culminates in an optional exam for which high-performing students may receive some credit towards their college coursework, depending on the institution.
Registration
The AP examination for AP Physics C: Mechanics is separate from the AP examination for AP Physics C: Electricity and Magnetism. Before 2006, test-takers paid only once and were given the choice of taking either one or two parts of the Physics C test.
Format
The exam is typically administered on a Monday aftern
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
A teacher builds a model of a hydrogen atom. A red golf ball is used for a proton, and a green golf ball is used for an electron. Which is not accurate concerning the model?
A. number of particles
B. relative mass of particles
C. types of particles present
D. charges of particles present
Answer:
|
|
sciq-7486
|
multiple_choice
|
Examples of what process are evident in many biological systems because cells are surrounded by semipermeable membranes?
|
[
"osmosis",
"apoptosis",
"photosynthesis",
"reproduction"
] |
A
|
Relavent Documents:
Document 0:::
Cell theory has its origins in seventeenth century microscopy observations, but it was nearly two hundred years before a complete cell membrane theory was developed to explain what separates cells from the outside world. By the 19th century it was accepted that some form of semi-permeable barrier must exist around a cell. Studies of the action of anesthetic molecules led to the theory that this barrier might be made of some sort of fat (lipid), but the structure was still unknown. A series of pioneering experiments in 1925 indicated that this barrier membrane consisted of two molecular layers of lipids—a lipid bilayer. New tools over the next few decades confirmed this theory, but controversy remained regarding the role of proteins in the cell membrane. Eventually the fluid mosaic model was composed in which proteins “float” in a fluid lipid bilayer "sea". Although simplistic and incomplete, this model is still widely referenced today.
[It is found in 1838.]]
Early barrier theories
Since the invention of the microscope in the seventeenth century it has been known that plant and animal tissue is composed of cells : the cell was discovered by Robert Hooke. The plant cell wall was easily visible even with these early microscopes but no similar barrier was visible on animal cells, though it stood to reason that one must exist. By the mid 19th century, this question was being actively investigated and Moritz Traube noted that this outer layer must be semipermeable to allow transport of ions. Traube had no direct evidence for the composition of this film, though, and incorrectly asserted that it was formed by an interfacial reaction of the cell protoplasm with the extracellular fluid.
The lipid nature of the cell membrane was first correctly intuited by Georg Hermann Quincke in 1888, who noted that a cell generally forms a spherical shape in water and, when broken in half, forms two smaller spheres. The only other known material to exhibit this behavior was oil. He al
Document 1:::
This is a list of articles on biophysics.
0–9
5-HT3 receptor
A
ACCN1
ANO1
AP2 adaptor complex
Aaron Klug
Acid-sensing ion channel
Activating function
Active transport
Adolf Eugen Fick
Afterdepolarization
Aggregate modulus
Aharon Katzir
Alan Lloyd Hodgkin
Alexander Rich
Alexander van Oudenaarden
Allan McLeod Cormack
Alpha-3 beta-4 nicotinic receptor
Alpha-4 beta-2 nicotinic receptor
Alpha-7 nicotinic receptor
Alpha helix
Alwyn Jones (biophysicist)
Amoeboid movement
Andreas Mershin
Andrew Huxley
Animal locomotion
Animal locomotion on the water surface
Anita Goel
Antiporter
Aquaporin 2
Aquaporin 3
Aquaporin 4
Archibald Hill
Ariel Fernandez
Arthropod exoskeleton
Arthropod leg
Avery Gilbert
B
BEST2
BK channel
Bacterial outer membrane
Balance (ability)
Bat
Bat wing development
Bert Sakmann
Bestrophin 1
Biased random walk (biochemistry)
Bioelectrochemical reactor
Bioelectrochemistry
Biofilm
Biological material
Biological membrane
Biomechanics
Biomechanics of sprint running
Biophysical Society
Biophysics
Bird flight
Bird migration
Bisindolylmaleimide
Bleb (cell biology)
Boris Pavlovich Belousov
Brian Matthews (biochemist)
Britton Chance
Brush border
Bulk movement
Document 2:::
The cell membrane (also known as the plasma membrane or cytoplasmic membrane, and historically referred to as the plasmalemma) is a biological membrane that separates and protects the interior of a cell from the outside environment (the extracellular space). The cell membrane consists of a lipid bilayer, made up of two layers of phospholipids with cholesterols (a lipid component) interspersed between them, maintaining appropriate membrane fluidity at various temperatures. The membrane also contains membrane proteins, including integral proteins that span the membrane and serve as membrane transporters, and peripheral proteins that loosely attach to the outer (peripheral) side of the cell membrane, acting as enzymes to facilitate interaction with the cell's environment. Glycolipids embedded in the outer lipid layer serve a similar purpose. The cell membrane controls the movement of substances in and out of a cell, being selectively permeable to ions and organic molecules. In addition, cell membranes are involved in a variety of cellular processes such as cell adhesion, ion conductivity, and cell signalling and serve as the attachment surface for several extracellular structures, including the cell wall and the carbohydrate layer called the glycocalyx, as well as the intracellular network of protein fibers called the cytoskeleton. In the field of synthetic biology, cell membranes can be artificially reassembled.
History
While Robert Hooke's discovery of cells in 1665 led to the proposal of the cell theory, Hooke misled the cell membrane theory that all cells contained a hard cell wall since only plant cells could be observed at the time. Microscopists focused on the cell wall for well over 150 years until advances in microscopy were made. In the early 19th century, cells were recognized as being separate entities, unconnected, and bound by individual cell walls after it was found that plant cells could be separated. This theory extended to include animal cells to su
Document 3:::
Semipermeable membrane is a type of biological or synthetic, polymeric membrane that will allow certain molecules or ions to pass through it by osmosis. The rate of passage depends on the pressure, concentration, and temperature of the molecules or solutes on either side, as well as the permeability of the membrane to each solute. Depending on the membrane and the solute, permeability may depend on solute size, solubility, properties, or chemistry. How the membrane is constructed to be selective in its permeability will determine the rate and the permeability. Many natural and synthetic materials which are rather thick are also semipermeable. One example of this is the thin film on the inside of the egg.
Biological membranes are selectively permeable, with the passage of molecules controlled by facilitated diffusion, passive transport or active transport regulated by proteins embedded in the membrane.
Biological membranes
An example of a biological semi-permeable membrane is the lipid bilayer, on which is based the plasma membrane that surrounds all biological cells. A group of phospholipids (consisting of a phosphate head and two fatty acid tails) arranged into a double layer, the phospholipid bilayer is a semipermeable membrane that is very specific in its permeability. The hydrophilic phosphate heads are in the outside layer and exposed to the water content outside and within the cell. The hydrophobic tails are the layer hidden in the inside of the membrane. Cholesterol molecules are also found throughout the plasma membrane and act as a buffer of membrane fluidity. The phospholipid bilayer is most permeable to small, uncharged solutes. Protein channels are embedded in or through phospholipids, and, collectively, this model is known as the fluid mosaic model. Aquaporins are protein channel pores permeable to water.
Cellular communication
Information can also pass through the plasma membrane when signaling molecules bind to receptors in the cell membrane. Th
Document 4:::
Cell physiology is the biological study of the activities that take place in a cell to keep it alive. The term physiology refers to normal functions in a living organism. Animal cells, plant cells and microorganism cells show similarities in their functions even though they vary in structure.
General characteristics
There are two types of cells: prokaryotes and eukaryotes.
Prokaryotes were the first of the two to develop and do not have a self-contained nucleus. Their mechanisms are simpler than later-evolved eukaryotes, which contain a nucleus that envelops the cell's DNA and some organelles.
Prokaryotes
Prokaryotes have DNA located in an area called the nucleoid, which is not separated from other parts of the cell by a membrane. There are two domains of prokaryotes: bacteria and archaea. Prokaryotes have fewer organelles than eukaryotes. Both have plasma membranes and ribosomes (structures that synthesize proteins and float free in cytoplasm). Two unique characteristics of prokaryotes are fimbriae (finger-like projections on the surface of a cell) and flagella (threadlike structures that aid movement).
Eukaryotes
Eukaryotes have a nucleus where DNA is contained. They are usually larger than prokaryotes and contain many more organelles. The nucleus, the feature of a eukaryote that distinguishes it from a prokaryote, contains a nuclear envelope, nucleolus and chromatin. In cytoplasm, endoplasmic reticulum (ER) synthesizes membranes and performs other metabolic activities. There are two types, rough ER (containing ribosomes) and smooth ER (lacking ribosomes). The Golgi apparatus consists of multiple membranous sacs, responsible for manufacturing and shipping out materials such as proteins. Lysosomes are structures that use enzymes to break down substances through phagocytosis, a process that comprises endocytosis and exocytosis. In the mitochondria, metabolic processes such as cellular respiration occur. The cytoskeleton is made of fibers that support the str
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Examples of what process are evident in many biological systems because cells are surrounded by semipermeable membranes?
A. osmosis
B. apoptosis
C. photosynthesis
D. reproduction
Answer:
|
|
sciq-10301
|
multiple_choice
|
What is the name of an uncharged particle whose weight is nearly equal to that of a proton?
|
[
"proton",
"a nuclei",
"electron",
"a neutron"
] |
D
|
Relavent Documents:
Document 0:::
The elementary charge, usually denoted by , is a fundamental physical constant, defined as the electric charge carried by a single proton or, equivalently, the magnitude of the negative electric charge carried by a single electron, which has charge −1 .
In the SI system of units, the value of the elementary charge is exactly defined as = coulombs, or 160.2176634 zeptocoulombs (zC). Since the 2019 redefinition of SI base units, the seven SI base units are defined by seven fundamental physical constants, of which the elementary charge is one.
In the centimetre–gram–second system of units (CGS), the corresponding quantity is .
Robert A. Millikan and Harvey Fletcher's oil drop experiment first directly measured the magnitude of the elementary charge in 1909, differing from the modern accepted value by just 0.6%. Under assumptions of the then-disputed atomic theory, the elementary charge had also been indirectly inferred to ~3% accuracy from blackbody spectra by Max Planck in 1901 and (through the Faraday constant) at order-of-magnitude accuracy by Johann Loschmidt's measurement of the Avogadro number in 1865.
As a unit
In some natural unit systems, such as the system of atomic units, e functions as the unit of electric charge. The use of elementary charge as a unit was promoted by George Johnstone Stoney in 1874 for the first system of natural units, called Stoney units. Later, he proposed the name electron for this unit. At the time, the particle we now call the electron was not yet discovered and the difference between the particle electron and the unit of charge electron was still blurred. Later, the name electron was assigned to the particle and the unit of charge e lost its name. However, the unit of energy electronvolt (eV) is a remnant of the fact that the elementary charge was once called electron.
In other natural unit systems, the unit of charge is defined as with the result that
where is the fine-structure constant, is the speed of light, is
Document 1:::
The subatomic scale is the domain of physical size that encompasses objects smaller than an atom. It is the scale at which the atomic constituents, such as the nucleus containing protons and neutrons, and the electrons in their orbitals, become apparent.
The subatomic scale includes the many thousands of times smaller subnuclear scale, which is the scale of physical size at which constituents of the protons and neutrons - particularly quarks - become apparent.
See also
Astronomical scale the opposite end of the spectrum
Subatomic particles
Document 2:::
This is a list of known and hypothesized particles.
Standard Model elementary particles
Elementary particles are particles with no measurable internal structure; that is, it is unknown whether they are composed of other particles. They are the fundamental objects of quantum field theory. Many families and sub-families of elementary particles exist. Elementary particles are classified according to their spin. Fermions have half-integer spin while bosons have integer spin. All the particles of the Standard Model have been experimentally observed, including the Higgs boson in 2012. Many other hypothetical elementary particles, such as the graviton, have been proposed, but not observed experimentally.
Fermions
Fermions are one of the two fundamental classes of particles, the other being bosons. Fermion particles are described by Fermi–Dirac statistics and have quantum numbers described by the Pauli exclusion principle. They include the quarks and leptons, as well as any composite particles consisting of an odd number of these, such as all baryons and many atoms and nuclei.
Fermions have half-integer spin; for all known elementary fermions this is . All known fermions except neutrinos, are also Dirac fermions; that is, each known fermion has its own distinct antiparticle. It is not known whether the neutrino is a Dirac fermion or a Majorana fermion. Fermions are the basic building blocks of all matter. They are classified according to whether they interact via the strong interaction or not. In the Standard Model, there are 12 types of elementary fermions: six quarks and six leptons.
Quarks
Quarks are the fundamental constituents of hadrons and interact via the strong force. Quarks are the only known carriers of fractional charge, but because they combine in groups of three quarks (baryons) or in pairs of one quark and one antiquark (mesons), only integer charge is observed in nature. Their respective antiparticles are the antiquarks, which are identical except th
Document 3:::
In particle physics, the electron mass (symbol: ) is the mass of a stationary electron, also known as the invariant mass of the electron. It is one of the fundamental constants of physics. It has a value of about or about , which has an energy-equivalent of about or about
Terminology
The term "rest mass" is sometimes used because in special relativity the mass of an object can be said to increase in a frame of reference that is moving relative to that object (or if the object is moving in a given frame of reference). Most practical measurements are carried out on moving electrons. If the electron is moving at a relativistic velocity, any measurement must use the correct expression for mass. Such correction becomes substantial for electrons accelerated by voltages of over .
For example, the relativistic expression for the total energy, , of an electron moving at speed is
where
is the speed of light;
is the Lorentz factor,
is the "rest mass", or more simply just the "mass" of the electron.
This quantity is frame invariant and velocity independent. However, some texts group the Lorentz factor with the mass factor to define a new quantity called the relativistic mass, .
Determination
Since the electron mass determines a number of observed effects in atomic physics, there are potentially many ways to determine its mass from an experiment, if the values of other physical constants are already considered known.
Historically, the mass of the electron was determined directly from combining two measurements. The mass-to-charge ratio of the electron was first estimated by Arthur Schuster in 1890 by measuring the deflection of "cathode rays" due to a known magnetic field in a cathode ray tube. Seven years later J. J. Thomson showed that cathode rays consist of streams of particles, to be called electrons, and made more precise measurements of their mass-to-charge ratio again using a cathode ray tube.
The second measurement was of the charge of the electron. T
Document 4:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is the name of an uncharged particle whose weight is nearly equal to that of a proton?
A. proton
B. a nuclei
C. electron
D. a neutron
Answer:
|
|
sciq-6270
|
multiple_choice
|
What are the largest phylum of the animal kingdom?
|
[
"Chelicerata",
"Crustacea",
"Pycnogonida",
"arthropods"
] |
D
|
Relavent Documents:
Document 0:::
Animals are multicellular, eukaryotic organisms in the biological kingdom Animalia. With few exceptions, animals consume organic material, breathe oxygen, have myocytes and are able to move, can reproduce sexually, and grow from a hollow sphere of cells, the blastula, during embryonic development. As of 2022, 2.16 million living animal species have been described—of which around 1.05 million are insects, over 85,000 are molluscs, and around 65,000 are vertebrates. It has been estimated there are around 7.77 million animal species. Animals range in length from to . They have complex interactions with each other and their environments, forming intricate food webs. The scientific study of animals is known as zoology.
Most living animal species are in Bilateria, a clade whose members have a bilaterally symmetric body plan. The Bilateria include the protostomes, containing animals such as nematodes, arthropods, flatworms, annelids and molluscs, and the deuterostomes, containing the echinoderms and the chordates, the latter including the vertebrates. Life forms interpreted as early animals were present in the Ediacaran biota of the late Precambrian. Many modern animal phyla became clearly established in the fossil record as marine species during the Cambrian explosion, which began around 539 million years ago. 6,331 groups of genes common to all living animals have been identified; these may have arisen from a single common ancestor that lived 650 million years ago.
Historically, Aristotle divided animals into those with blood and those without. Carl Linnaeus created the first hierarchical biological classification for animals in 1758 with his Systema Naturae, which Jean-Baptiste Lamarck expanded into 14 phyla by 1809. In 1874, Ernst Haeckel divided the animal kingdom into the multicellular Metazoa (now synonymous with Animalia) and the Protozoa, single-celled organisms no longer considered animals. In modern times, the biological classification of animals relies on ad
Document 1:::
History of Animals (, Ton peri ta zoia historion, "Inquiries on Animals"; , "History of Animals") is one of the major texts on biology by the ancient Greek philosopher Aristotle, who had studied at Plato's Academy in Athens. It was written in the fourth century BC; Aristotle died in 322 BC.
Generally seen as a pioneering work of zoology, Aristotle frames his text by explaining that he is investigating the what (the existing facts about animals) prior to establishing the why (the causes of these characteristics). The book is thus an attempt to apply philosophy to part of the natural world. Throughout the work, Aristotle seeks to identify differences, both between individuals and between groups. A group is established when it is seen that all members have the same set of distinguishing features; for example, that all birds have feathers, wings, and beaks. This relationship between the birds and their features is recognized as a universal.
The History of Animals contains many accurate eye-witness observations, in particular of the marine biology around the island of Lesbos, such as that the octopus had colour-changing abilities and a sperm-transferring tentacle, that the young of a dogfish grow inside their mother's body, or that the male of a river catfish guards the eggs after the female has left. Some of these were long considered fanciful before being rediscovered in the nineteenth century. Aristotle has been accused of making errors, but some are due to misinterpretation of his text, and others may have been based on genuine observation. He did however make somewhat uncritical use of evidence from other people, such as travellers and beekeepers.
The History of Animals had a powerful influence on zoology for some two thousand years. It continued to be a primary source of knowledge until zoologists in the sixteenth century, such as Conrad Gessner, all influenced by Aristotle, wrote their own studies of the subject.
Context
Aristotle (384–322 BC) studied at Plat
Document 2:::
Animals are multicellular eukaryotic organisms in the biological kingdom Animalia. With few exceptions, animals consume organic material, breathe oxygen, are able to move, reproduce sexually, and grow from a hollow sphere of cells, the blastula, during embryonic development. Over 1.5 million living animal species have been described—of which around 1 million are insects—but it has been estimated there are over 7 million in total. Animals range in size from 8.5 millionths of a metre to long and have complex interactions with each other and their environments, forming intricate food webs. The study of animals is called zoology.
Animals may be listed or indexed by many criteria, including taxonomy, status as endangered species, their geographical location, and their portrayal and/or naming in human culture.
By common name
List of animal names (male, female, young, and group)
By aspect
List of common household pests
List of animal sounds
List of animals by number of neurons
By domestication
List of domesticated animals
By eating behaviour
List of herbivorous animals
List of omnivores
List of carnivores
By endangered status
IUCN Red List endangered species (Animalia)
United States Fish and Wildlife Service list of endangered species
By extinction
List of extinct animals
List of extinct birds
List of extinct mammals
List of extinct cetaceans
List of extinct butterflies
By region
Lists of amphibians by region
Lists of birds by region
Lists of mammals by region
Lists of reptiles by region
By individual (real or fictional)
Real
Lists of snakes
List of individual cats
List of oldest cats
List of giant squids
List of individual elephants
List of historical horses
List of leading Thoroughbred racehorses
List of individual apes
List of individual bears
List of giant pandas
List of individual birds
List of individual bovines
List of individual cetaceans
List of individual dogs
List of oldest dogs
List of individual monkeys
List of individual pigs
List of w
Document 3:::
Caminalcules are a fictive group of animal-like life forms, which were created as a tool for better understanding phylogenetics in real organisms. They were created by Joseph H. Camin (University of Kansas) and consist of 29 living 'species' and 48 fossil forms.
The name of the taxon Caminalcules seems to come from Camin's last name and Antonie van Leeuwenhoek's animalcules.
History
Joseph H. Camin (1922–1979) drew the Caminalcules in the early 1960s or possibly even earlier to study the nature of taxonomic judgement. He assured that there was genetic continuity in the Caminalcules by the preservation of all characters except for changes that he desired in all successive animals. He did not keep track of the changes he made in the different species. The images of the Caminalcules were made using master stencils. The images of the living OTUs (29 species) were made available in the early 1960s; those of the fossil ones (48 species) later in the decade. These images were copied using xerography. Copies of all OTUs were in the possession of Dr. Paul A. Ehrlich (Stanford University), Dr. W. Wayne Moss (Philadelphia Academy of Sciences) and Robert R. Sokal (State University of New York at Stony Brook) in 1983. The original drawings by Joseph H. Camin have unfortunately been lost.
The Caminalcules first appeared in print in the journal Systematic Zoology (now Systematic Biology) in 1983, four years after Camin's death in 1979. Robert R. Sokal published four succeeding papers about them, titled "A Phylogenetic Analysis of the Caminalcules." These papers included the complete set of living and fossil species, as well as their cladogram, which Sokal had received from Camin in 1970.
At a symposium dedicated to Camin, Dr. W. Wayne Moss said that "his collaborative studies on methods and principles of systematics at Kansas in the 1960s resulted in the appearance of that delightful taxon, the Caminalcules", and that "his thoughts helped to launch the infant field of pheneti
Document 4:::
Currently, the addition of many lineages of newly discovered protists (such as Telonemia, Picozoa, Hemimastigophora, Rigifilida...) and the use of phylogenomic analyses have brought a new, more accurate supergroup model. These are the current supergroups of eukaryotes:
TSAR, constituted by Telonemia and the SAR clade (Stramenopiles, Alveolata and Rhizaria). It is estimated to occupy up to half of all eukaryotic diversity, since it includes multiple major groups such as diatoms, dinoflagellates, seaweeds, ciliates, foraminiferans, radiolarians, and the apicomplexan and oomycete parasites. It essentially contains the majority of "Chromalveolata".
Haptista (also treated as a phylum), previously in "Chromalveolata", comprising the haptophyte algae and centrohelids.
Cryptista (also treated as a phylum), previously in "Chromalveolata", comprising the cryptomonads, katablepharids and the enigmatic Palpitomonas.
Archaeplastida (also treated as a kingdom), con
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What are the largest phylum of the animal kingdom?
A. Chelicerata
B. Crustacea
C. Pycnogonida
D. arthropods
Answer:
|
|
sciq-3163
|
multiple_choice
|
What is defined as a saturated layer of rock or soil?
|
[
"river bank",
"aquifer",
"culvert",
"artesian well"
] |
B
|
Relavent Documents:
Document 0:::
USDA soil taxonomy (ST) developed by the United States Department of Agriculture and the National Cooperative Soil Survey provides an elaborate classification of soil types according to several parameters (most commonly their properties) and in several levels: Order, Suborder, Great Group, Subgroup, Family, and Series. The classification was originally developed by Guy Donald Smith, former director of the U.S. Department of Agriculture's soil survey investigations.
Discussion
A taxonomy is an arrangement in a systematic manner; the USDA soil taxonomy has six levels of classification. They are, from most general to specific: order, suborder, great group, subgroup, family and series. Soil properties that can be measured quantitatively are used in this classification system – they include: depth, moisture, temperature, texture, structure, cation exchange capacity, base saturation, clay mineralogy, organic matter content and salt content. There are 12 soil orders (the top hierarchical level) in soil taxonomy. The names of the orders end with the suffix -sol. The criteria for the different soil orders include properties that reflect major differences in the genesis of soils. The orders are:
Alfisol – soils with aluminium and iron. They have horizons of clay accumulation, and form where there is enough moisture and warmth for at least three months of plant growth. They constitute 10% of soils worldwide.
Andisol – volcanic ash soils. They are young soils. They cover 1% of the world's ice-free surface.
Aridisol – dry soils forming under desert conditions which have fewer than 90 consecutive days of moisture during the growing season and are nonleached. They include nearly 12% of soils on Earth. Soil formation is slow, and accumulated organic matter is scarce. They may have subsurface zones of caliche or duripan. Many aridisols have well-developed Bt horizons showing clay movement from past periods of greater moisture.
Entisol – recently formed soils that lack well-d
Document 1:::
Soil classification deals with the systematic categorization of soils based on distinguishing characteristics as well as criteria that dictate choices in use.
Overview
Soil classification is a dynamic subject, from the structure of the system, to the definitions of classes, to the application in the field. Soil classification can be approached from the perspective of soil as a material and soil as a resource.
Inscriptions at the temple of Horus at Edfu outline a soil classification used by Tanen to determine what kind of temple to build at which site. Ancient Greek scholars produced a number of classification based on several different qualities of the soil.
Engineering
Geotechnical engineers classify soils according to their engineering properties as they relate to use for foundation support or building material. Modern engineering classification systems are designed to allow an easy transition from field observations to basic predictions of soil engineering properties and behaviors.
The most common engineering classification system for soils in North America is the Unified Soil Classification System (USCS). The USCS has three major classification groups: (1) coarse-grained soils (e.g. sands and gravels); (2) fine-grained soils (e.g. silts and clays); and (3) highly organic soils (referred to as "peat"). The USCS further subdivides the three major soil classes for clarification. It distinguishes sands from gravels by grain size, classifying some as "well-graded" and the rest as "poorly-graded". Silts and clays are distinguished by the soils' Atterberg limits, and thus the soils are separated into "high-plasticity" and "low-plasticity" soils. Moderately organic soils are considered subdivisions of silts and clays and are distinguished from inorganic soils by changes in their plasticity properties (and Atterberg limits) on drying. The European soil classification system (ISO 14688) is very similar, differing primarily in coding and in adding an "intermediate-p
Document 2:::
The World Reference Base for Soil Resources (WRB) is an international soil classification system for naming soils and creating legends for soil maps. The currently valid version is the fourth edition 2022. It is edited by a working group of the International Union of Soil Sciences (IUSS).
Background
History
Since the 19th century, several countries developed national soil classification systems. During the 20th century, the need for an international soil classification system became more and more obvious.
From 1971 to 1981, the Food and Agriculture Organization (FAO) and UNESCO published the Soil Map of the World, 10 volumes, scale 1 : 5 M). The Legend for this map, published in 1974 under the leadership of Rudi Dudal, became the FAO soil classification. Many ideas from national soil classification systems were brought together in this worldwide-applicable system, among them the idea of diagnostic horizons as established in the '7th approximation to the USDA soil taxonomy' from 1960. The next step was the Revised Legend of the Soil Map of the World, published in 1988.
In 1982, the International Soil Science Society (ISSS; now: International Union of Soil Sciences, IUSS) established a working group named International Reference Base for Soil Classification (IRB). Chair of this working group was Ernst Schlichting. Its mandate was to develop an international soil classification system that should better consider soil-forming processes than the FAO soil classification. Drafts were presented in 1982 and 1990.
In 1992, the IRB working group decided to develop a new system named World Reference Base for Soil Resources (WRB) that should further develop the Revised Legend of the FAO soil classification and include some ideas of the more systematic IRB approach. Otto Spaargaren (International Soil Reference and Information Centre) and Freddy Nachtergaele (FAO) were nominated to prepare a draft. This draft was presented at the 15th World Congress of Soil Science in Acapu
Document 3:::
The Géotechnique lecture is an biennial lecture on the topic of soil mechanics, organised by the British Geotechnical Association named after its major scientific journal Géotechnique.
This should not be confused with the annual BGA Rankine Lecture.
List of Géotechnique Lecturers
See also
Named lectures
Rankine Lecture
Terzaghi Lecture
External links
ICE Géotechnique journal
British Geotechnical Association
Document 4:::
The pore space of soil contains the liquid and gas phases of soil, i.e., everything but the solid phase that contains mainly minerals of varying sizes as well as organic compounds.
In order to understand porosity better a series of equations have not been used to express the quantitative interactions between the three phases of soil.
Macropores or fractures play a major role in infiltration rates in many soils as well as preferential flow patterns, hydraulic conductivity and evapotranspiration. Cracks are also very influential in gas exchange, influencing respiration within soils. Modeling cracks therefore helps understand how these processes work and what the effects of changes in soil cracking such as compaction, can have on these processes.
The pore space of soil may contain the habitat of plants (rhizosphere) and microorganisms.
Background
Bulk density
The bulk density of soil depends greatly on the mineral make up of soil and the degree of compaction. The density of quartz is around 2.65 g/cm3 but the bulk density of a soil may be less than half that density.
Most soils have a bulk density between 1.0 and 1.6 g/cm3 but organic soil and some friable clay may have a bulk density well below 1 g/cm3.
Core samples are taken by driving a metal core into the earth at the desired depth and soil horizon.
The samples are then oven dried and weighed.
Bulk density = (mass of oven dry soil)/volume
The bulk density of soil is inversely related to the porosity of the same soil. The more pore space in a soil the lower the value for bulk density.
Porosity
or
Porosity is a measure of the total pore space in the soil. This is measured as a volume or percent. The amount of porosity in a soil depends on the minerals that make up the soil and the amount of sorting that occurs within the soil structure. For example, a sandy soil will have larger porosity than silty sand, because the silt will fill in the gaps between the sand particles.
Pore space relations
Hydrauli
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is defined as a saturated layer of rock or soil?
A. river bank
B. aquifer
C. culvert
D. artesian well
Answer:
|
|
sciq-1675
|
multiple_choice
|
What organism is at the top of the energy pyramid?
|
[
"herbivore",
"primate",
"carnivore",
"insects"
] |
C
|
Relavent Documents:
Document 0:::
Consumer–resource interactions are the core motif of ecological food chains or food webs, and are an umbrella term for a variety of more specialized types of biological species interactions including prey-predator (see predation), host-parasite (see parasitism), plant-herbivore and victim-exploiter systems. These kinds of interactions have been studied and modeled by population ecologists for nearly a century. Species at the bottom of the food chain, such as algae and other autotrophs, consume non-biological resources, such as minerals and nutrients of various kinds, and they derive their energy from light (photons) or chemical sources. Species higher up in the food chain survive by consuming other species and can be classified by what they eat and how they obtain or find their food.
Classification of consumer types
The standard categorization
Various terms have arisen to define consumers by what they eat, such as meat-eating carnivores, fish-eating piscivores, insect-eating insectivores, plant-eating herbivores, seed-eating granivores, and fruit-eating frugivores and omnivores are meat eaters and plant eaters. An extensive classification of consumer categories based on a list of feeding behaviors exists.
The Getz categorization
Another way of categorizing consumers, proposed by South African American ecologist Wayne Getz, is based on a biomass transformation web (BTW) formulation that organizes resources into five components: live and dead animal, live and dead plant, and particulate (i.e. broken down plant and animal) matter. It also distinguishes between consumers that gather their resources by moving across landscapes from those that mine their resources by becoming sessile once they have located a stock of resources large enough for them to feed on during completion of a full life history stage.
In Getz's scheme, words for miners are of Greek etymology and words for gatherers are of Latin etymology. Thus a bestivore, such as a cat, preys on live animal
Document 1:::
The trophic level of an organism is the position it occupies in a food web. A food chain is a succession of organisms that eat other organisms and may, in turn, be eaten themselves. The trophic level of an organism is the number of steps it is from the start of the chain. A food web starts at trophic level 1 with primary producers such as plants, can move to herbivores at level 2, carnivores at level 3 or higher, and typically finish with apex predators at level 4 or 5. The path along the chain can form either a one-way flow or a food "web". Ecological communities with higher biodiversity form more complex trophic paths.
The word trophic derives from the Greek τροφή (trophē) referring to food or nourishment.
History
The concept of trophic level was developed by Raymond Lindeman (1942), based on the terminology of August Thienemann (1926): "producers", "consumers", and "reducers" (modified to "decomposers" by Lindeman).
Overview
The three basic ways in which organisms get food are as producers, consumers, and decomposers.
Producers (autotrophs) are typically plants or algae. Plants and algae do not usually eat other organisms, but pull nutrients from the soil or the ocean and manufacture their own food using photosynthesis. For this reason, they are called primary producers. In this way, it is energy from the sun that usually powers the base of the food chain. An exception occurs in deep-sea hydrothermal ecosystems, where there is no sunlight. Here primary producers manufacture food through a process called chemosynthesis.
Consumers (heterotrophs) are species that cannot manufacture their own food and need to consume other organisms. Animals that eat primary producers (like plants) are called herbivores. Animals that eat other animals are called carnivores, and animals that eat both plants and other animals are called omnivores.
Decomposers (detritivores) break down dead plant and animal material and wastes and release it again as energy and nutrients into
Document 2:::
An ecological pyramid (also trophic pyramid, Eltonian pyramid, energy pyramid, or sometimes food pyramid') is a graphical representation designed to show the biomass or bioproductivity at each trophic level in an ecosystem.
A pyramid of energy shows how much energy is retained in the form of new biomass from each trophic level, while a pyramid of biomass shows how much biomass (the amount of living or organic matter present in an organism) is present in the organisms. There is also a pyramid of numbers representing the number of individual organisms at each trophic level. Pyramids of energy are normally upright, but other pyramids can be inverted(pyramid of biomass for marine region) or take other shapes.(spindle shaped pyramid)
Ecological pyramids begin with producers on the bottom (such as plants) and proceed through the various trophic levels (such as herbivores that eat plants, then carnivores that eat flesh, then omnivores that eat both plants and flesh, and so on). The highest level is the top of the food chain.
Biomass can be measured by a bomb calorimeter.
Pyramid of Energy
A pyramid of energy or pyramid of productivity shows the production or turnover (the rate at which energy or mass is transferred from one trophic level to the next) of biomass at each trophic level. Instead of showing a single snapshot in time, productivity pyramids show the flow of energy through the food chain. Typical units are grams per square meter per year or calories per square meter per year. As with the others, this graph shows producers at the bottom and higher trophic levels on top.
When an ecosystem is healthy, this graph produces a standard ecological pyramid. This is because, in order for the ecosystem to sustain itself, there must be more energy at lower trophic levels than there is at higher trophic levels. This allows organisms on the lower levels to not only maintain a stable population, but also to transfer energy up the pyramid. The exception to this generalizati
Document 3:::
The soil food web is the community of organisms living all or part of their lives in the soil. It describes a complex living system in the soil and how it interacts with the environment, plants, and animals.
Food webs describe the transfer of energy between species in an ecosystem. While a food chain examines one, linear, energy pathway through an ecosystem, a food web is more complex and illustrates all of the potential pathways. Much of this transferred energy comes from the sun. Plants use the sun’s energy to convert inorganic compounds into energy-rich, organic compounds, turning carbon dioxide and minerals into plant material by photosynthesis. Plant flowers exude energy-rich nectar above ground and plant roots exude acids, sugars, and ectoenzymes into the rhizosphere, adjusting the pH and feeding the food web underground.
Plants are called autotrophs because they make their own energy; they are also called producers because they produce energy available for other organisms to eat. Heterotrophs are consumers that cannot make their own food. In order to obtain energy they eat plants or other heterotrophs.
Above ground food webs
In above ground food webs, energy moves from producers (plants) to primary consumers (herbivores) and then to secondary consumers (predators). The phrase, trophic level, refers to the different levels or steps in the energy pathway. In other words, the producers, consumers, and decomposers are the main trophic levels. This chain of energy transferring from one species to another can continue several more times, but eventually ends. At the end of the food chain, decomposers such as bacteria and fungi break down dead plant and animal material into simple nutrients.
Methodology
The nature of soil makes direct observation of food webs difficult. Since soil organisms range in size from less than 0.1 mm (nematodes) to greater than 2 mm (earthworms) there are many different ways to extract them. Soil samples are often taken using a metal
Document 4:::
An energy budget is a balance sheet of energy income against expenditure. It is studied in the field of Energetics which deals with the study of energy transfer and transformation from one form to another. Calorie is the basic unit of measurement. An organism in a laboratory experiment is an open thermodynamic system, exchanging energy with its surroundings in three ways - heat, work and the potential energy of biochemical compounds.
Organisms use ingested food resources (C=consumption) as building blocks in the synthesis of tissues (P=production) and as fuel in the metabolic process that power this synthesis and other physiological processes (R=respiratory loss). Some of the resources are lost as waste products (F=faecal loss, U=urinary loss). All these aspects of metabolism can be represented in energy units. The basic model of energy budget may be shown as:
P = C - R - U - F or
P = C - (R + U + F) or
C = P + R + U + F
All the aspects of metabolism can be represented in energy units (e.g. joules (J);1 calorie = 4.2 kJ).
Energy used for metabolism will be
R = C - (F + U + P)
Energy used in the maintenance will be
R + F + U = C - P
Endothermy and ectothermy
Energy budget allocation varies for endotherms and ectotherms. Ectotherms rely on the environment as a heat source while endotherms maintain their body temperature through the regulation of metabolic processes. The heat produced in association with metabolic processes facilitates the active lifestyles of endotherms and their ability to travel far distances over a range of temperatures in the search for food. Ectotherms are limited by the ambient temperature of the environment around them but the lack of substantial metabolic heat production accounts for an energetically inexpensive metabolic rate. The energy demands for ectotherms are generally one tenth of that required for endotherms.
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What organism is at the top of the energy pyramid?
A. herbivore
B. primate
C. carnivore
D. insects
Answer:
|
|
sciq-398
|
multiple_choice
|
What is the largest cartilaginous fish?
|
[
"dolphin",
"blue whale",
"whale shark",
"sturgeon"
] |
C
|
Relavent Documents:
Document 0:::
Fish intelligence is the resultant of the process of acquiring, storing in memory, retrieving, combining, comparing, and using in new contexts information and conceptual skills" as it applies to fish.
According to Culum Brown from Macquarie University, "Fish are more intelligent than they appear. In many areas, such as memory, their cognitive powers match or exceed those of ‘higher’ vertebrates including non-human primates."
Fish hold records for the relative brain weights of vertebrates. Most vertebrate species have similar brain-to-body mass ratios. The deep sea bathypelagic bony-eared assfish has the smallest ratio of all known vertebrates. At the other extreme, the electrogenic elephantnose fish, an African freshwater fish, has one of the largest brain-to-body weight ratios of all known vertebrates (slightly higher than humans) and the highest brain-to-body oxygen consumption ratio of all known vertebrates (three times that for humans).
Brain
Fish typically have quite small brains relative to body size compared with other vertebrates, typically one-fifteenth the brain mass of a similarly sized bird or mammal. However, some fish have relatively large brains, most notably mormyrids and sharks, which have brains about as massive relative to body weight as birds and marsupials.
The cerebellum of cartilaginous and bony fishes is large and complex. In at least one important respect, it differs in internal structure from the mammalian cerebellum: The fish cerebellum does not contain discrete deep cerebellar nuclei. Instead, the primary targets of Purkinje cells are a distinct type of cell distributed across the cerebellar cortex, a type not seen in mammals. The circuits in the cerebellum are similar across all classes of vertebrates, including fish, reptiles, birds, and mammals. There is also an analogous brain structure in cephalopods with well-developed brains, such as octopuses. This has been taken as evidence that the cerebellum performs functions important to
Document 1:::
The Bachelor of Science in Aquatic Resources and Technology (B.Sc. in AQT) (or Bachelor of Aquatic Resource) is an undergraduate degree that prepares students to pursue careers in the public, private, or non-profit sector in areas such as marine science, fisheries science, aquaculture, aquatic resource technology, food science, management, biotechnology and hydrography. Post-baccalaureate training is available in aquatic resource management and related areas.
The Department of Animal Science and Export Agriculture, at the Uva Wellassa University of Badulla, Sri Lanka, has the largest enrollment of undergraduate majors in Aquatic Resources and Technology, with about 200 students as of 2014.
The Council on Education for Aquatic Resources and Technology includes undergraduate AQT degrees in the accreditation review of Aquatic Resources and Technology programs and schools.
See also
Marine Science
Ministry of Fisheries and Aquatic Resources Development
Document 2:::
The Digital Fish Library (DFL) is a University of California San Diego project funded by the Biological Infrastructure Initiative (DBI) of the National Science Foundation (NSF). The DFL creates 2D and 3D visualizations of the internal and external anatomy of fish obtained with magnetic resonance imaging (MRI) methods and makes these publicly available on the web.
The information core for the Digital Fish Library is generated using high-resolution MRI scanners housed at the Center for functional magnetic resonance imaging (CfMRI) multi-user facility at UC San Diego. These instruments use magnetic fields to take 3D images of animal tissues, allowing researchers to non-invasively see inside them and quantitatively describe their 3D anatomy. Fish specimens are obtained from the Marine Vertebrate Collection at Scripps Institute of Oceanography (SIO) and imaged by staff from UC San Diego's Center for Scientific Computation in Imaging (CSCI).
As of February 2010, the Digital Fish Library contains almost 300 species covering all five classes of fish, 56 of 60 orders, and close to 200 of the 521 fish families as described by Nelson, 2006. DFL imaging has also contributed to a number of published peer-reviewed scientific studies.
Digital Fish Library work has been featured in the media, including two National Geographic documentaries: Magnetic Navigator and Ultimate Shark.
Document 3:::
Sharks are a group of elasmobranch fish characterized by a cartilaginous skeleton, five to seven gill slits on the sides of the head, and pectoral fins that are not fused to the head. Modern sharks are classified within the clade Selachimorpha (or Selachii) and are the sister group to the Batoidea (rays and kin). Some sources extend the term "shark" as an informal category including extinct members of Chondrichthyes (cartilaginous fish) with a shark-like morphology, such as hybodonts. Shark-like chondrichthyans such as Cladoselache and Doliodus first appeared in the Devonian Period (419–359 million years), though some fossilized chondrichthyan-like scales are as old as the Late Ordovician (458–444 million years ago). The oldest modern sharks (selachimorphs) are known from the Early Jurassic, about 200 million years ago.
Sharks range in size from the small dwarf lanternshark (Etmopterus perryi), a deep sea species that is only in length, to the whale shark (Rhincodon typus), the largest fish in the world, which reaches approximately in length. They are found in all seas and are common to depths up to . They generally do not live in freshwater, although there are a few known exceptions, such as the bull shark and the river shark, which can be found in both seawater and freshwater. Sharks have a covering of dermal denticles that protects their skin from damage and parasites in addition to improving their fluid dynamics. They have numerous sets of replaceable teeth.
Several species are apex predators, which are organisms that are at the top of their food chain. Select examples include the tiger shark, blue shark, great white shark, mako shark, thresher shark, and hammerhead shark.
Sharks are caught by humans for shark meat or shark fin soup. Many shark populations are threatened by human activities. Since 1970, shark populations have been reduced by 71%, mostly from overfishing.
Etymology
Until the 16th century, sharks were known to mariners as "sea dogs". This is
Document 4:::
Marine vertebrates are vertebrates that live in marine environments. These are the marine fish and the marine tetrapods (primarily seabirds, marine reptiles, and marine mammals). Vertebrates are a subphylum of chordates that have a vertebral column (backbone). The vertebral column provides the central support structure for an internal skeleton. The internal skeleton gives shape, support, and protection to the body and can provide a means of anchoring fins or limbs to the body. The vertebral column also serves to house and protect the spinal cord that lies within the column.
Marine vertebrates can be divided into two groups, marine fish and marine tetrapods.
Marine fish
Fish fall into two main groups: fish with bony internal skeletons and fish with cartilaginous internal skeletons. Fish anatomy and physiology generally includes a two-chambered heart, eyes adapted to seeing underwater, and a skin protected by scales and mucous. They typically breathe by extracting oxygen from water through gills. Fish use fins to propel and stabilise themselves in the water. Over 33,000 species of fish have been described as of 2017, of which about 20,000 are marine fish.
Jawless fish
Hagfish form a class of about 20 species of eel-shaped, slime-producing marine fish. They are the only known living animals that have a skull but no vertebral column. Lampreys form a superclass containing 38 known extant species of jawless fish. The adult lamprey is characterized by a toothed, funnel-like sucking mouth. Although they are well known for boring into the flesh of other fish to suck their blood, only 18 species of lampreys are actually parasitic. Together hagfish and lampreys are the sister group to vertebrates. Living hagfish remain similar to hagfish from around 300 million years ago. The lampreys are a very ancient lineage of vertebrates, though their exact relationship to hagfishes and jawed vertebrates is still a matter of dispute. Molecular analysis since 1992 has suggested that h
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is the largest cartilaginous fish?
A. dolphin
B. blue whale
C. whale shark
D. sturgeon
Answer:
|
|
sciq-11548
|
multiple_choice
|
What drives the turbine in a thermonuclear reactor?
|
[
"magnets",
"heated oil",
"wind energy",
"heated water or steam"
] |
D
|
Relavent Documents:
Document 0:::
Micropower describes the use of very small electric generators and prime movers or devices to convert heat or motion to electricity, for use close to the generator. The generator is typically integrated with microelectronic devices and produces "several watts of power or less." These devices offer the promise of a power source for portable electronic devices which is lighter weight and has a longer operating time than batteries.
Microturbine technology
The components of any turbine engine — the gas compressor, the combustion chamber, and the turbine rotor — are fabricated from etched silicon, much like integrated circuits. The technology holds the promise of ten times the operating time of a battery of the same weight as the micropower unit, and similar efficiency to large utility gas turbines. Researchers at Massachusetts Institute of Technology have thus far succeeded in fabricating the parts for such a micro turbine out of six etched and stacked silicon wafers, and are working toward combining them into a functioning engine about the size of a U.S. quarter coin.
Researchers at Georgia Tech have built a micro generator 10 mm wide, which spins a magnet above an array of coils fabricated on a silicon chip. The device spins at 100,000 revolutions per minute, and produces 1.1 watts of electrical power, sufficient to operate a cell phone. Their goal is to produce 20 to 50 watts, sufficient to power a laptop computer.
Scientists at Lehigh University are developing a hydrogen generator on a silicon chip that can convert methanol, diesel, or gasoline into fuel for a microengine or a miniature fuel cell.
Professor Sanjeev Mukerjee of Northeastern University's chemistry department is developing fuel cells for the military that will burn hydrogen to power portable electronic equipment, such as night vision goggles, computers, and communication equipment. In his system, a cartridge of methanol would be used to produce hydrogen to run a small fuel cell for up to 5,000 ho
Document 1:::
The tip-speed ratio, λ, or TSR for wind turbines is the ratio between the tangential speed of the tip of a blade and the actual speed of the wind, . The tip-speed ratio is related to efficiency, with the optimum varying with blade design. Higher tip speeds result in higher noise levels and require stronger blades due to larger centrifugal forces.
The tip speed of the blade can be calculated as times R, where is the rotational speed of the rotor in radians/second, and R is the rotor radius in metres. Therefore, we can also write:
where is the wind speed in metres/second at the height of the blade hub.
Cp–λ curves
The power coefficient, is a quantity that expresses what fraction of the power in the wind is being extracted by the wind turbine. It is generally assumed to be a function of both tip-speed ratio and pitch angle. Below is a plot of the variation of the power coefficient with variations in the tip-speed ratio when the pitch is held constant:
The case for variable speed wind turbines
Originally, wind turbines were fixed speed. This has the benefit that the rotor speed in the generator is constant, thus the frequency of the AC voltage is fixed. This allows the wind turbine to be directly connected to a transmission system. However, from the figure above, we can see that the power coefficient is a function of the tip-speed ratio. By extension, the efficiency of the wind turbine is a function of the tip-speed ratio.
Ideally, one would like to have a turbine operating at the maximum value of at all wind speeds. This means that as the wind speed changes, the rotor speed must change to such that . A wind turbine with a variable rotor speed is called a variable speed wind turbine. Whilst this does mean that the wind turbine operates at or close to for a range of wind speeds, the frequency of the AC voltage generator will not be constant. This can be seen in the following equation:
where is the rotor angular speed, is the frequency of the AC volta
Document 2:::
In electricity generation, a generator is a device that converts motion-based power (potential and kinetic energy) or fuel-based power (chemical energy) into electric power for use in an external circuit. Sources of mechanical energy include steam turbines, gas turbines, water turbines, internal combustion engines, wind turbines and even hand cranks. The first electromagnetic generator, the Faraday disk, was invented in 1831 by British scientist Michael Faraday. Generators provide nearly all the power for electrical grids.
In addition to electricity- and motion-based designs, photovoltaic and fuel cell powered generators use solar power and hydrogen-based fuels, respectively, to generate electrical output.
The reverse conversion of electrical energy into mechanical energy is done by an electric motor, and motors and generators are very similar. Many motors can generate electricity from mechanical energy.
Terminology
Electromagnetic generators fall into one of two broad categories, dynamos and alternators.
Dynamos generate pulsing direct current through the use of a commutator.
Alternators generate alternating current.
Mechanically, a generator consists of a rotating part and a stationary part which together form a magnetic circuit:
Rotor: The rotating part of an electrical machine.
Stator: The stationary part of an electrical machine, which surrounds the rotor.
One of these parts generates a magnetic field, the other has a wire winding in which the changing field induces an electric current:
Field winding or field (permanent) magnets: The magnetic field-producing component of an electrical machine. The magnetic field of the dynamo or alternator can be provided by either wire windings called field coils or permanent magnets. Electrically-excited generators include an excitation system to produce the field flux. A generator using permanent magnets (PMs) is sometimes called a magneto, or a permanent magnet synchronous generator (PMSG).
Armature: The power-p
Document 3:::
A Stirling radioisotope generator (SRG) is a type of radioisotope generator based on a Stirling engine powered by a large radioisotope heater unit. The hot end of the Stirling converter reaches high temperature and heated helium drives the piston, with heat being rejected at the cold end of the engine. A generator or alternator converts the motion into electricity. Given the very constrained supply of plutonium, the Stirling converter is notable for producing about four times as much electric power from the plutonium fuel as compared to a radioisotope thermoelectric generator (RTG).
The Stirling generators were extensively tested on Earth by NASA, but their development was cancelled in 2013 before they could be deployed on actual spacecraft missions. A similar NASA project still under development, called Kilopower, also utilizes Stirling engines, but uses a small uranium fission reactor as the heat source.
History
Stirling and Brayton-cycle technology development has been conducted at NASA Glenn Research Center (formerly NASA Lewis) since the early 1970s. The Space Demonstrator Engine (SPDE) was the earliest 12.5 kWe per cylinder engine that was designed, built and tested. A later engine of this size, the Component Test Power Converter (CTPC), used a "Starfish" heat-pipe heater head, instead of the pumped-loop used by the SPDE. In the 1992-93 time period, this work was stopped due to the termination of the related SP-100 nuclear power system work and NASA's new emphasis on "better, faster, cheaper" systems and missions.
In 2020, a free-piston Stirling power converter reached 15 years of maintenance-free and degradation-free cumulative operation in the Stirling Research Laboratory at NASA Glenn. This duration equals the operational design life of the MMRTG, and is representative of typical mission concepts designed to explore the outer planets or even more distant Kuiper Belt Objects. This unit, called the Technology Demonstration Converter (TDC) #13, is the old
Document 4:::
This page lists examples of the power in watts produced by various sources of energy. They are grouped by orders of magnitude from small to large.
Below 1 W
1 to 102 W
103 to 108 W
The productive capacity of electrical generators operated by utility companies is often measured in MW. Few things can sustain the transfer or consumption of energy on this scale; some of these events or entities include: lightning strikes, naval craft (such as aircraft carriers and submarines), engineering hardware, and some scientific research equipment (such as supercolliders and large lasers).
For reference, about 10,000 100-watt lightbulbs or 5,000 computer systems would be needed to draw 1 MW. Also, 1 MW is approximately 1360 horsepower. Modern high-power diesel-electric locomotives typically have a peak power of 3–5 MW, while a typical modern nuclear power plant produces on the order of 500–2000 MW peak output.
109 to 1014 W
1015 to 1026 W
Over 1027 W
See also
Orders of magnitude (energy)
Orders of magnitude (voltage)
World energy resources and consumption
International System of Units (SI)
SI prefix
Notes
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What drives the turbine in a thermonuclear reactor?
A. magnets
B. heated oil
C. wind energy
D. heated water or steam
Answer:
|
|
sciq-5252
|
multiple_choice
|
Plant-like protists are commonly called what?
|
[
"fungus",
"sponge",
"seaweed",
"algae"
] |
D
|
Relavent Documents:
Document 0:::
Marine protists are defined by their habitat as protists that live in marine environments, that is, in the saltwater of seas or oceans or the brackish water of coastal estuaries. Life originated as marine single-celled prokaryotes (bacteria and archaea) and later evolved into more complex eukaryotes. Eukaryotes are the more developed life forms known as plants, animals, fungi and protists. Protists are the eukaryotes that cannot be classified as plants, fungi or animals. They are mostly single-celled and microscopic. The term protist came into use historically as a term of convenience for eukaryotes that cannot be strictly classified as plants, animals or fungi. They are not a part of modern cladistics because they are paraphyletic (lacking a common ancestor for all descendants).
Most protists are too small to be seen with the naked eye. They are highly diverse organisms currently organised into 18 phyla, but not easy to classify. Studies have shown high protist diversity exists in oceans, deep sea-vents and river sediments, suggesting large numbers of eukaryotic microbial communities have yet to be discovered. There has been little research on mixotrophic protists, but recent studies in marine environments found mixotrophic protists contribute a significant part of the protist biomass. Since protists are eukaryotes (and not prokaryotes) they possess within their cell at least one nucleus, as well as organelles such as mitochondria and Golgi bodies. Many protist species can switch between asexual reproduction and sexual reproduction involving meiosis and fertilization.
In contrast to the cells of prokaryotes, the cells of eukaryotes are highly organised. Plants, animals and fungi are usually multi-celled and are typically macroscopic. Most protists are single-celled and microscopic. But there are exceptions. Some single-celled marine protists are macroscopic. Some marine slime molds have unique life cycles that involve switching between unicellular, colonial, and
Document 1:::
A protist ( ) or protoctist is any eukaryotic organism that is not an animal, plant, or fungus. Protists do not form a natural group, or clade, but an artificial grouping of several independent clades that evolved from the last eukaryotic common ancestor.
Protists were historically regarded as a separate taxonomic kingdom known as Protista or Protoctista. With the advent of phylogenetic analysis and electron microscopy studies, the use of Protista as a formal taxon was gradually abandoned. In modern classifications, protists are spread across several eukaryotic clades called supergroups, such as Archaeplastida (which includes plants), SAR, Obazoa (which includes fungi and animals), Amoebozoa and Excavata.
Protists represent an extremely large genetic and ecological diversity in all environments, including extreme habitats. Their diversity, larger than for all other eukaryotes, has only been discovered in recent decades through the study of environmental DNA, and is still in the process of being fully described. They are present in all ecosystems as important components of the biogeochemical cycles and trophic webs. They exist abundantly and ubiquitously in a variety of forms that evolved multiple times independently, such as free-living algae, amoebae and slime moulds, or as important parasites. Together, they compose an amount of biomass that doubles that of animals. They exhibit varied types of nutrition (such as phototrophy, phagotrophy or osmotrophy), sometimes combining them (in mixotrophy). They present unique adaptations not present in multicellular animals, fungi or land plants. The study of protists is termed protistology.
Definition
There is not a single accepted definition of what protists are. As a paraphyletic assemblage of diverse biological groups, they have historically been regarded as a catch-all taxon that includes any eukaryotic organism (i.e., living beings whose cells possess a nucleus) that is not an animal, a land plant or a dikaryon fung
Document 2:::
A protist is any eukaryotic organism (that is, an organism whose cells contain a cell nucleus) that is not an animal, plant, or fungus. While it is likely that protists share a common ancestor, the last eukaryotic common ancestor, the exclusion of other eukaryotes means that protists do not form a natural group, or clade. Therefore, some protists may be more closely related to animals, plants, or fungi than they are to other protists. However, like algae, invertebrates and protozoans, the grouping is used for convenience.
Many protists have neither hard parts nor resistant spores, and their fossils are extremely rare or unknown. Examples of such groups include the apicomplexans, most ciliates, some green algae (the Klebsormidiales), choanoflagellates, oomycetes, brown algae, yellow-green algae, Excavata (e.g., euglenids). Some of these have been found preserved in amber (fossilized tree resin) or under unusual conditions (e.g., Paleoleishmania, a kinetoplastid).
Others are relatively common in the fossil record, as the diatoms, golden algae, haptophytes (coccoliths), silicoflagellates, tintinnids (ciliates), dinoflagellates, green algae, red algae, heliozoans, radiolarians, foraminiferans, ebriids and testate amoebae (euglyphids, arcellaceans). Some are used as paleoecological indicators to reconstruct ancient environments.
More probable eukaryote fossils begin to appear at about 1.8 billion years ago, the acritarchs, spherical fossils of likely algal protists. Another possible representative of early fossil eukaryotes are the Gabonionta.
Modern classifications
Systematists today do not treat Protista as a formal taxon, but the term "protist" is still commonly used for convenience in two ways. The most popular contemporary definition is a phylogenetic one, that identifies a paraphyletic group: a protist is any eukaryote that is not an animal, (land) plant, or (true) fungus; this definition excludes many unicellular groups, like the Microsporidia (fungi), many C
Document 3:::
Plectenchyma (from Greek πλέκω pleko 'I weave' and ἔγχυμα enchyma 'infusion', i.e., 'a woven tissue') is the general term employed to designate all types of fungal tissues.
The two most common types of tissues are prosenchyma and pseudoparenchyma.
The hyphae specifically become fused together.
Notes
Fungal morphology and anatomy
Document 4:::
Ultrastructural identity is a concept in biology. It asserts that evolutionary lineages of eukaryotes in general and protists in particular can be distinguished by complements and arrangements of cellular organelles. These ultrastructural components can be visualized by electron microscopy.
The concept emerged following the application of electron microscopy to protists.
Protists
Early ultrastructural studies revealed that many previously accepted groupings of protists based on optical microscopy included organisms with differing cellular organelles. Those groups included amoebae, flagellates, heliozoa, radiolaria, sporozoa, slime molds, and chromophytic algae. They were deemed likely to be polyphyletic, and their inclusion in efforts to assemble a phylogenetic tree would cause confusion. As an example of this work, German cell biologist Christian Bardele established unexpected diversity with the simply organized heliozoa. His work made it evident that heliozoa were not monophyletic and subsequent studies revealed that the heliozoa was composed of seven types of organisms: actinophryids, centrohelids, ciliophryids, desmothoracids, dimporphids, gymnosphaerids and taxopodids.
A critical advance was made by British phycologist David Hibberd. He demonstrated that two types of chromophytic algae, previously presumed to be closely related, had different organizations that were revealed by electron microscopy. The number and organization of locomotor organelles differed (chrysophyte - two flagella; haptophyte - two flagella and haponema), the surfaces of which differed (chrysophyte - with tripartite flagellar hairs now regarded as apomorphic for stramenopiles; haptophyte - naked), as did the transitional zone between axoneme and basal body (chrysophyte with helix); as did flagellar anchorage systems; presence or absence of embellishments on the cell surface (chrysophyte - naked; haptophyte - with scales), plastids especially eyespot, location and functions of dictyosom
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Plant-like protists are commonly called what?
A. fungus
B. sponge
C. seaweed
D. algae
Answer:
|
|
sciq-7223
|
multiple_choice
|
Besides depth soundings, what else are bathymetric maps made of?
|
[
"solar data",
"density measures",
"intensity data",
"sonar data"
] |
D
|
Relavent Documents:
Document 0:::
Bathymetry (; ) is the study of underwater depth of ocean floors (seabed topography), lake floors, or river floors. In other words, bathymetry is the underwater equivalent to hypsometry or topography. The first recorded evidence of water depth measurements are from Ancient Egypt over 3000 years ago. Bathymetric charts (not to be confused with hydrographic charts), are typically produced to support safety of surface or sub-surface navigation, and usually show seafloor relief or terrain as contour lines (called depth contours or isobaths) and selected depths (soundings), and typically also provide surface navigational information. Bathymetric maps (a more general term where navigational safety is not a concern) may also use a Digital Terrain Model and artificial illumination techniques to illustrate the depths being portrayed. The global bathymetry is sometimes combined with topography data to yield a global relief model. Paleobathymetry is the study of past underwater depths.
Synonyms include seafloor mapping, seabed mapping, seafloor imaging and seabed imaging. Bathymetric measurements are conducted with various methods, from depth sounding, sonar and Lidar techniques, to buoys and satellite altimetry. Various methods have advantages and disadvantages and the specific method used depends upon the scale of the area under study, financial means, desired measurement accuracy, and additional variables. Despite modern computer-based research, the ocean seabed in many locations is less measured than the topography of Mars.
Seabed topography
Measurement
Originally, bathymetry involved the measurement of ocean depth through depth sounding. Early techniques used pre-measured heavy rope or cable lowered over a ship's side. This technique measures the depth only a singular point at a time, and is therefore inefficient. It is also subject to movements of the ship and currents moving the line out of true and therefore is not accurate.
The data used to make bathymetric maps
Document 1:::
A bathymetric chart is a type of isarithmic map that depicts the submerged topography and physiographic features of ocean and sea bottoms. Their primary purpose is to provide detailed depth contours of ocean topography as well as provide the size, shape and distribution of underwater features. Topographic maps display elevation above ground and are complementary to bathymetric charts. Charts use a series of lines and points at equal intervals to showcase depth or elevation. A closed shape with increasingly smaller shapes inside of it can indicate an ocean trench or a seamount, or underwater mountain, depending on whether the depths increase or decrease going inward.
Bathymetric surveys and charts are associated with the science of oceanography, particularly marine geology, and underwater engineering or other specialized purposes. Bathymetric data used to produce charts can also be converted to bathymetric profiles which are vertical sections through a feature.
History
Ancient Egypt
The use of bathymetry and the development of bathymetric charts dates back around the 19th century BC to ancient Egypt. Depictions on tomb walls such as the bas-relief carvings of Deir al-Bahri commissioned by Queen Hatshepsut in the 16th century BC show ancient mariners using long slender poles as sounding poles to determine the depth of the Nile River and into the Nile River Delta.
Ancient Greece
The first written account and mapped records of sounding did not occur until 1000 years after the Egyptians had begun sounding and mapping the Nile. The Greek historian Herodotus writes of a sounding in 66 feet of water of the mouth of the Nile in the river delta. He writes of yellow mud being brought up similar to the same that was deposited with the yearly floods. These accounts show a heightened awareness of regional depths and seafloor characteristics among ancient mariners and demonstrate that discoveries in bathymetry and the use of bathymetric charts had progressed significantly.
Document 2:::
An underwater survey is a survey performed in an underwater environment or conducted remotely on an underwater object or region. Survey can have several meanings. The word originates in Medieval Latin with meanings of looking over and detailed study of a subject. One meaning is the accurate measurement of a geographical region, usually with the intention of plotting the positions of features as a scale map of the region. This meaning is often used in scientific contexts, and also in civil engineering and mineral extraction. Another meaning, often used in a civil, structural, or marine engineering context, is the inspection of a structure or vessel to compare actual condition with the specified nominal condition, usually with the purpose of reporting on the actual condition and compliance with, or deviations from, the nominal condition, for quality control, damage assessment, valuation, insurance, maintenance, and similar purposes. In other contexts it can mean inspection of a region to establish presence and distribution of specified content, such as living organisms, either to establish a baseline, or to compare with a baseline.
These types of survey may be done in or of the underwater environment, in which case they may be referred to as underwater surveys, which may include bathymetric, hydrographic, and geological surveys, archaeological surveys, ecological surveys, and structural or vessel safety surveys. In some cases they can be done by remote sensing, using a variety of tools, and sometimes by direct human intervention, usually by a professional diver. Underwater surveys are an essential part of the planning, and often of quality control and monitoring, of underwater construction, dredging, mineral extraction, ecological monitoring, and archaeological investigations. They are often required as part of an ecological impact study.
Types
The types of underwater survey include, but are not necessarily restricted to, archeological, bathymetric and hydrographic
Document 3:::
The Index to Marine & Lacustrine Geological Samples is a collaboration between multiple institutions and agencies that operate geological sample repositories. The purpose of the database is to help researchers locate sea floor and lakebed cores, grabs, dredges, and drill samples in their collections.
Sample material is available from participating institutions unless noted as unavailable.
Data include basic collection and storage information. Lithology, texture, age, principal investigator, province, weathering/metamorphism, glass remarks, and descriptive comments are included for some samples. Links are provided to related data and information at the institutions and at NCEI.
Data are coded by individual institutions, several of which receive funding from the US National Science Foundation. For more information see the NSF Division of Ocean Sciences Data and Sample Policy.
The Index is endorsed by the Intergovernmental Oceanographic Commission, Committee on International Oceanographic Data and Information Exchange (IODE-XIV.2).
The index is maintained by the National Centers for Environmental Information (NCEI), formerly the National Geophysical Data Center (NGDC), and collocated World Data Center for Geophysics, Boulder, Colorado. NCEI is part of the National Environmental Satellite, Data and Information Service of the National Oceanic & Atmospheric Administration, U. S. Department of Commerce.
Searches and data downloads are available via a JSP and an ArcIMS interface. Data selections can be downloaded in tab-delimited or shapefile form, depending on the interface used. Both WMS and WFS interfaces are also available.
The Index was created in 1977 in response to a meeting of Curators of Marine Geological Samples, sponsored by the U.S. National Science Foundation. The Curators' group continues to meet every 2–3 years.
Dataset Digital Object Identifier
DOI:10.7289/V5H41PB8
Web site
The Index to Marine and Lacustrine Geological Samples
Participating Ins
Document 4:::
A submarine pipeline (also known as marine, subsea or offshore pipeline) is a pipeline that is laid on the seabed or below it inside a trench. In some cases, the pipeline is mostly on-land but in places it crosses water expanses, such as small seas, straits and rivers. Submarine pipelines are used primarily to carry oil or gas, but transportation of water is also important. A distinction is sometimes made between a flowline and a pipeline. The former is an intrafield pipeline, in the sense that it is used to connect subsea wellheads, manifolds and the platform within a particular development field. The latter, sometimes referred to as an export pipeline, is used to bring the resource to shore. Sizeable pipeline construction projects need to take into account many factors, such as the offshore ecology, geohazards and environmental loading – they are often undertaken by multidisciplinary, international teams.
Route selection
One of the earliest and most critical tasks in a submarine pipeline planning exercise is the route selection. This selection has to consider a variety of issues, some of a political nature, but most others dealing with geohazards, physical factors along the prospective route, and other uses of the seabed in the area considered. This task begins with a fact-finding exercise, which is a standard desk study that includes a survey of geological maps, bathymetry, fishing charts, aerial and satellite photography, as well as information from navigation authorities.
Physical factors
The primary physical factor to be considered in submarine pipeline construction is the state of the seabed – whether it is smooth (i.e., relatively flat) or uneven (corrugated, with high points and low points). If it is uneven, the pipeline will include free spans when it connects two high points, leaving the section in between unsupported. If an unsupported section is too long, the bending stress exerted onto it (due to its weight) may be excessive. Vibration from current
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Besides depth soundings, what else are bathymetric maps made of?
A. solar data
B. density measures
C. intensity data
D. sonar data
Answer:
|
|
sciq-8061
|
multiple_choice
|
Monatomic cations have the same name as their parent element. what happens to monatomic anions?
|
[
"ends in - de",
"ends in - be",
"ends in - fe",
"ends in - se"
] |
A
|
Relavent Documents:
Document 0:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 1:::
Advanced Placement (AP) Physics C: Electricity and Magnetism (also known as AP Physics C: E&M or AP E&M) is an introductory physics course administered by the College Board as part of its Advanced Placement program. It is intended to proxy a second-semester calculus-based university course in electricity and magnetism. The content of Physics C: E&M overlaps with that of AP Physics 2, but Physics 2 is algebra-based and covers other topics outside of electromagnetism, while Physics C is calculus-based and only covers electromagnetism. Physics C: E&M may be combined with its mechanics counterpart to form a year-long course that prepares for both exams.
Course content
E&M is equivalent to an introductory college course in electricity and magnetism for physics or engineering majors. The course modules are:
Electrostatics
Conductors, capacitors, and dielectrics
Electric circuits
Magnetic fields
Electromagnetism.
Methods of calculus are used wherever appropriate in formulating physical principles and in applying them to physical problems. Therefore, students should have completed or be concurrently enrolled in a calculus class.
AP test
The course culminates in an optional exam for which high-performing students may receive some credit towards their college coursework, depending on the institution.
Registration
The AP examination for AP Physics C: Electricity and Magnetism is separate from the AP examination for AP Physics C: Mechanics. Before 2006, test-takers paid only once and were given the choice of taking either one or two parts of the Physics C test.
Format
The exam is typically administered on a Monday afternoon in May. The exam is configured in two categories: a 35-question multiple choice section and a 3-question free response section. Test takers are allowed to use an approved calculator during the entire exam. The test is weighted such that each section is worth half of the final score. This and AP Physics C: Mechanics are the shortest AP exams, with
Document 2:::
There are four Advanced Placement (AP) Physics courses administered by the College Board as part of its Advanced Placement program: the algebra-based Physics 1 and Physics 2 and the calculus-based Physics C: Mechanics and Physics C: Electricity and Magnetism. All are intended to be at the college level. Each AP Physics course has an exam for which high-performing students may receive credit toward their college coursework.
AP Physics 1 and 2
AP Physics 1 and AP Physics 2 were introduced in 2015, replacing AP Physics B. The courses were designed to emphasize critical thinking and reasoning as well as learning through inquiry. They are algebra-based and do not require any calculus knowledge.
AP Physics 1
AP Physics 1 covers Newtonian mechanics, including:
Unit 1: Kinematics
Unit 2: Dynamics
Unit 3: Circular Motion and Gravitation
Unit 4: Energy
Unit 5: Momentum
Unit 6: Simple Harmonic Motion
Unit 7: Torque and Rotational Motion
Until 2020, the course also covered topics in electricity (including Coulomb's Law and resistive DC circuits), mechanical waves, and sound. These units were removed because they are included in AP Physics 2.
AP Physics 2
AP Physics 2 covers the following topics:
Unit 1: Fluids
Unit 2: Thermodynamics
Unit 3: Electric Force, Field, and Potential
Unit 4: Electric Circuits
Unit 5: Magnetism and Electromagnetic Induction
Unit 6: Geometric and Physical Optics
Unit 7: Quantum, Atomic, and Nuclear Physics
AP Physics C
From 1969 to 1972, AP Physics C was a single course with a single exam that covered all standard introductory university physics topics, including mechanics, fluids, electricity and magnetism, optics, and modern physics. In 1973, the College Board split the course into AP Physics C: Mechanics and AP Physics C: Electricity and Magnetism. The exam was also split into two separate 90-minute tests, each equivalent to a semester-length calculus-based college course. Until 2006, both exams could be taken for a single
Document 3:::
{|class="wikitable"
|-
! colspan="2" | Common monatomic anions
|-
| hydride
| H−
|-
| fluoride
| F−
|-
| chloride
| Cl−
|-
| bromide
| Br−
|-
| iodide
| I−
|-
| oxide
| O2−
|-
| sulfide
| S2−
|-
| nitride
| N3−
|-
| phosphide
| P3−
|-
|}
Document 4:::
Advanced Level (A-Level) Mathematics is a qualification of further education taken in the United Kingdom (and occasionally other countries as well). In the UK, A-Level exams are traditionally taken by 17-18 year-olds after a two-year course at a sixth form or college. Advanced Level Further Mathematics is often taken by students who wish to study a mathematics-based degree at university, or related degree courses such as physics or computer science.
Like other A-level subjects, mathematics has been assessed in a modular system since the introduction of Curriculum 2000, whereby each candidate must take six modules, with the best achieved score in each of these modules (after any retake) contributing to the final grade. Most students will complete three modules in one year, which will create an AS-level qualification in their own right and will complete the A-level course the following year—with three more modules.
The system in which mathematics is assessed is changing for students starting courses in 2017 (as part of the A-level reforms first introduced in 2015), where the reformed specifications have reverted to a linear structure with exams taken only at the end of the course in a single sitting.
In addition, while schools could choose freely between taking Statistics, Mechanics or Discrete Mathematics (also known as Decision Mathematics) modules with the ability to specialise in one branch of applied Mathematics in the older modular specification, in the new specifications, both Mechanics and Statistics were made compulsory, with Discrete Mathematics being made exclusive as an option to students pursuing a Further Mathematics course. The first assessment opportunity for the new specification is 2018 and 2019 for A-levels in Mathematics and Further Mathematics, respectively.
2000s specification
Prior to the 2017 reform, the basic A-Level course consisted of six modules, four pure modules (C1, C2, C3, and C4) and two applied modules in Statistics, Mechanics
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Monatomic cations have the same name as their parent element. what happens to monatomic anions?
A. ends in - de
B. ends in - be
C. ends in - fe
D. ends in - se
Answer:
|
|
ai2_arc-944
|
multiple_choice
|
Robots can perform tasks that are dangerous for humans. What is the MAJOR limitation to the use of robots?
|
[
"The assembly pieces must be very small.",
"The assembly process must remain exactly the same.",
"Robots require regular maintenance.",
"Robots must be provided with electricity."
] |
B
|
Relavent Documents:
Document 0:::
Human–robot interaction (HRI) is the study of interactions between humans and robots. Human–robot interaction is a multidisciplinary field with contributions from human–computer interaction, artificial intelligence, robotics, natural language processing, design, and psychology. A subfield known as physical human–robot interaction (pHRI) has tended to focus on device design to enable people to safely interact with robotic systems.
Origins
Human–robot interaction has been a topic of both science fiction and academic speculation even before any robots existed. Because much of active HRI development depends on natural language processing, many aspects of HRI are continuations of human communications, a field of research which is much older than robotics.
The origin of HRI as a discrete problem was stated by 20th-century author Isaac Asimov in 1941, in his novel I, Robot. Asimov coined Three Laws of Robotics, namely:
A robot may not injure a human being or, through inaction, allow a human being to come to harm.
A robot must obey the orders by human beings except where such orders would conflict with the First Law.
A robot must protect its own existence as long as such protection does not conflict with the First or Second Laws.
These three laws provide an overview of the goals engineers and researchers hold for safety in the HRI field, although the fields of robot ethics and machine ethics are more complex than these three principles. However, generally human–robot interaction prioritizes the safety of humans that interact with potentially dangerous robotics equipment. Solutions to this problem range from the philosophical approach of treating robots as ethical agents (individuals with moral agency), to the practical approach of creating safety zones. These safety zones use technologies such as lidar to detect human presence or physical barriers to protect humans by preventing any contact between machine and operator.
Although initially robots in the human–robot
Document 1:::
Henrik Iskov Christensen (born July 16, 1962 in Frederikshavn, Denmark) is a Danish roboticist and Professor of Computer Science at Dept. of Computer Science and Engineering, at the UC San Diego Jacobs School of Engineering. He is also the Director of the Contextual Robotics Institute at UC San Diego.
Prior to UC San Diego, he was a Distinguished Professor of Computer Science in the School of Interactive Computing at the Georgia Institute of Technology. At Georgia Tech, Christensen served as the founding director of the Institute for Robotics and Intelligent Machines (IRIM@GT) and the KUKA Chair of Robotics.
Previously, Christensen was the Founding Chairman of European Robotics Research Network (EURON) and an IEEE Robotics and Automation Society Distinguished Lecturer in Robotics.
Biography
Christensen received his Certificate of Apprenticeship in Mechanical Engineering from the Frederikshavn Technical School, Denmark in 1981. He received his M.Sc. and Ph.D. in Electrical Engineering from Aalborg University in 1987 and 1990, respectively. His doctoral thesis Aspects of Real Time Image Sequence Analysis was advised by Erik Granum.
After receiving his Ph.D., Christensen held teaching and research positions at Aalborg University, Oak Ridge National Laboratory, and the Royal Institute of Technology. In 2006, Christensen accepted a part-time position at the Georgia Institute of Technology as a Distinguished Professor of Computer Science and the KUKA Chair of Robotics, and transitioned to full-time in early 2007. At Georgia Tech, Christensen served as the founding director of the Center for Robotics and Intelligent Machines (RIM@GT), an interdepartmental research units consists of the College of Computing, College of Engineering, and the Georgia Tech Research Institute (GTRI). During his tenure, RIM@GT experienced an unprecedented growth, including (as of 2008) 36 faculty members as well as a dedicated interdisciplinary Ph.D. program in Robotics.
He joined UC San Die
Document 2:::
Laws of robotics are any set of laws, rules, or principles, which are intended as a fundamental framework to underpin the behavior of robots designed to have a degree of autonomy. Robots of this degree of complexity do not yet exist, but they have been widely anticipated in science fiction, films and are a topic of active research and development in the fields of robotics and artificial intelligence.
The best known set of laws are those written by Isaac Asimov in the 1940s, or based upon them, but other sets of laws have been proposed by researchers in the decades since then.
Isaac Asimov's "Three Laws of Robotics"
The best known set of laws are Isaac Asimov's "Three Laws of Robotics". These were introduced in his 1942 short story "Runaround", although they were foreshadowed in a few earlier stories. The Three Laws are:
A robot may not injure a human being or, through inaction, allow a human being to come to harm.
A robot must obey the orders by human beings except where such orders would conflict with the First Law.
A robot must protect its own existence as long as such protection does not conflict with the First or Second Laws.
In The Evitable Conflict the machines generalize the First Law to mean:
No machine may harm humanity; or, through inaction, allow humanity to come to harm.
This was refined in the end of Foundation and Earth, a zeroth law was introduced, with the original three suitably rewritten as subordinate to it:
Adaptations and extensions exist based upon this framework. As of 2021 they remain a "fictional device".
EPSRC / AHRC principles of robotics
In 2011, the Engineering and Physical Sciences Research Council (EPSRC) and the Arts and Humanities Research Council (AHRC) of United Kingdom jointly published a set of five ethical "principles for designers, builders and users of robots" in the real world, along with seven "high-level messages" intended to be conveyed, based on a September 2010 research workshop:
Robots should not be de
Document 3:::
The adaptive machine is a category of flexible industrial machinery characterized by the ability to intelligently adapt itself to the product to be produced, e.g. to move individual products through the manufacturing, assembly, inspection, packaging and other process stations required for to produce them.
Examples of intelligent adaptive capabilities
Detecting the size and shape of products in a current production batch and automatically adapting the distance of the actuators used to clamp them and to transport them through the processing line
Adapting the route a product takes through a processing line based on the specifications of the product; each product can take its own individual route through a production line and only stop at processing stations where it actually needs processing. Unnecessary stations are by-passed, via an alternate route or just passed through. This is in stark contrast to fixed-indexing systems such as belt and chain conveyors or round dials that still represent the majority of installed manufacturing equipment.
Exemplary non-characteristics
Using a multitude of work-piece holders specific to each single product variant
Operator intervention for mechanical change-over (e.g. exchanging mechanical cams)
Design approach
At first glance, the adaptive capabilities are rooted in software. But a second look reveals that machines handle physical products. For doing so machines need a proper mechanical design as well as a proper electrical design to power the mechanical movements. An adaptive machine is best designed by applying an interdisciplinary mechatronic design approach where mechanics, electrics and software as well as their interfaces and interactions are considered holistically.
Purpose
The primary function of an adaptive machine is to make production more flexible by enable greater product variety (e.g. with respect to product size and shape) and smaller batches. The ultimate goal of an adaptive machine is mass customizati
Document 4:::
Cloud robotics is a field of robotics that attempts to invoke cloud technologies such as cloud computing, cloud storage, and other Internet technologies centered on the benefits of converged infrastructure and shared services for robotics. When connected to the cloud, robots can benefit from the powerful computation, storage, and communication resources of modern data center in the cloud, which can process and share information from various robots or agent (other machines, smart objects, humans, etc.). Humans can also delegate tasks to robots remotely through networks. Cloud computing technologies enable robot systems to be endowed with powerful capability whilst reducing costs through cloud technologies. Thus, it is possible to build lightweight, low-cost, smarter robots with an intelligent "brain" in the cloud. The "brain" consists of data center, knowledge base, task planners, deep learning, information processing, environment models, communication support, etc.
Components
A cloud for robots potentially has at least six significant components:
Building a "cloud brain" for robots. It is the main object of cloud robotics.
Offering a global library of images, maps, and object data, often with geometry and mechanical properties, expert system, knowledge base (i.e. semantic web, data centres);
Massively-parallel computation on demand for sample-based statistical modelling and motion planning, task planning, multi-robot collaboration, scheduling and coordination of system;
Robot sharing of outcomes, trajectories, and dynamic control policies and robot learning support;
Human sharing of "open-source" code, data, and designs for programming, experimentation, and hardware construction;
On-demand human guidance and assistance for evaluation, learning, and error recovery;
Augmented human–robot interaction through various way (Semantics knowledge base, Apple SIRI like service etc.).
Applications
Autonomous mobile robots Google's self-driving cars are cloud robots.
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Robots can perform tasks that are dangerous for humans. What is the MAJOR limitation to the use of robots?
A. The assembly pieces must be very small.
B. The assembly process must remain exactly the same.
C. Robots require regular maintenance.
D. Robots must be provided with electricity.
Answer:
|
|
sciq-4460
|
multiple_choice
|
What is the term for chordates with a vertebral column and endoskeleton of cartilage and bone?
|
[
"insect",
"am",
"mammel",
"vertebrate"
] |
D
|
Relavent Documents:
Document 0:::
This glossary describes the terms used in formal descriptions of spiders; where applicable these terms are used in describing other arachnids.
Links within the glossary are shown .
Terms
A
Abdomen or opisthosoma: One of the two main body parts (tagmata), located towards the posterior end; see also Abdomen § Other animals
Accessory claw: Modified at the tip of the in web-building spiders; used with to grip strands of the web
Anal tubercle: A small protuberance (tubercule) above the through which the anus opens
Apodeme: see
Apophysis (plural apophyses): An outgrowth or process changing the general shape of a body part, particularly the appendages; often used in describing the male : see
Atrium (plural atria): An internal chamber at the entrance to the in female haplogyne spiders
B
Bidentate: Having two
Book lungs: Respiratory organs on the ventral side (underside) of the , in front of the , opening through narrow slits; see also Book lungs
Branchial operculum: see
Bulbus: see
C
Calamistrum (plural calamistra): Modified setae (bristles) on the of the fourth leg of spiders with a , arranged in one or more rows or in an oval shape, used to comb silk produced by the cribellum; see also Calamistrum
Caput (plural capita): see
Carapace: A hardened plate (sclerite) covering the upper (dorsal) portion of the ; see also Carapace
Carpoblem: The principal on the male ; also just called the tibial apophysis
Cephalic region or caput: The front part of the , separated from the thoracic region by the
Cephalothorax or prosoma: One of the two main body parts (tagmata), located towards the anterior end, composed of the head ( or caput) and the thorax (thoracic region), the two regions being separated by the ; covered by the and bearing the , legs, and mouthparts
Cervical groove: A shallow U-shaped groove, separating the and thoracic regions of the
Chelate: A description of a where the closes against a tooth-like process
Chelic
Document 1:::
Myomeres are blocks of skeletal muscle tissue arranged in sequence, commonly found in aquatic chordates. Myomeres are separated from adjacent myomeres by connective fascia (myosepta) and most easily seen in larval fishes or in the olm. Myomere counts are sometimes used for identifying specimens, since their number corresponds to the number of vertebrae in the adults. Location varies, with some species containing these only near the tails, while some have them located near the scapular or pelvic girdles. Depending on the species, myomeres could be arranged in an epaxial or hypaxial manner. Hypaxial refers to ventral muscles and related structures while epaxial refers to more dorsal muscles. The horizontal septum divides these two regions in vertebrates from cyclostomes to gnathostomes. In terrestrial chordates, the myomeres become fused as well as indistinct, due to the disappearance of myosepta.
Shape
The shape of myomeres varies by species. Myomeres are commonly zig-zag, "V" (lancelets), "W" (fishes), or straight (tetrapods)– shaped muscle fibers. Generally, cyclostome myomeres are arranged in vertical strips while those of jawed fishes are folded in a complex matter due to swimming capability evolution. Specifically, myomeres of elasmobranchs and eels are “W”-shaped. Contrastingly, myomeres of tetrapods run vertically and do not display complex folding. Another species with simply-lain myomeres are mudpuppies. Myomeres overlap each other in succession, meaning myomere activation also allows neighboring myomeres to activate.
Myomeres are made up of myoglobin-rich dark muscle as well as white muscle. Dark muscle, generally, functions as slow-twitch muscle fibers while white muscle is composed of fast-twitch fibers.
Function
Specifically, three types of myomeres in fish-like chordates include amphioxine (lancelet), cyclostomine (jawless fish), and gnathostomine (jawed fish). A common function shared by all of these is that they function to flex the body lateral
Document 2:::
In anatomy, a suture is a fairly rigid joint between two or more hard elements of an organism, with or without significant overlap of the elements.
Sutures are found in the skeletons or exoskeletons of a wide range of animals, in both invertebrates and vertebrates. Sutures are found in animals with hard parts from the Cambrian period to the present day. Sutures were and are formed by several different methods, and they exist between hard parts that are made from several different materials.
Vertebrate skeletons
The skeletons of vertebrate animals (fish, amphibians, reptiles, birds, and mammals) are made of bone, in which the main rigid ingredient is calcium phosphate.
Cranial sutures
The skulls of most vertebrates consist of sets of bony plates held together by cranial sutures. These sutures are held together mainly by Sharpey's fibers which grow from each bone into the adjoining one.
Sutures in the ankles of land vertebrates
In the type of crurotarsal ankle which is found in crocodilians and some other archosaurs, the astragalus is fixed to the tibia by a suture, and the joint bends around a peg on the astragalus, which fits into a socket in the calcaneum.
Invertebrate exoskeletons
In molluscs
The shells of most molluscs are made of calcium carbonate (the main constituent of limestone and chalk), and of conchiolin, a protein. For more information, see Mollusc shell.
Sutures in the shells of cephalopods
In cephalopod mollusks which have external shells (e.g. Nautilus, ammonites), the shell is divided into compartments by septa (partitions).
The septa are joined to the external shell by sutures formed by repeated invagination (they interlock like pieces of a jigsaw puzzle). The sutures are visible from the outside and often form complex and elaborate patterns.
The suture in the shells of gastropods
Nearly all snail shells (except for the shells of limpets, abalone, sea hares, etc.) can be visualized as a tube of increasing diameter, closed at the small e
Document 3:::
Cephalization is an evolutionary trend in which, over many generations, the mouth, sense organs, and nerve ganglia become concentrated at the front end of an animal, producing a head region. This is associated with movement and bilateral symmetry, such that the animal has a definite head end. This led to the formation of a highly sophisticated brain in three groups of animals, namely the arthropods, cephalopod molluscs, and vertebrates.
Animals without bilateral symmetry
Cnidaria, such as the radially symmetrical Hydrozoa, show some degree of cephalization. The Anthomedusae have a head end with their mouth, photoreceptive cells, and a concentration of neural cells.
Bilateria
Cephalization is a characteristic feature of the Bilateria, a large group containing the majority of animal phyla. These have the ability to move, using muscles, and a body plan with a front end that encounters stimuli first as the animal moves forwards, and accordingly has evolved to contain many of the body's sense organs, able to detect light, chemicals, and gravity. There is often also a collection of nerve cells able to process the information from these sense organs, forming a brain in several phyla and one or more ganglia in others.
Acoela
The Acoela are basal bilaterians, part of the Xenacoelomorpha. They are small and simple animals, and have very slightly more nerve cells at the head end than elsewhere, not forming a distinct and compact brain. This represents an early stage in cephalization.
Flatworms
The Platyhelminthes (flatworms) have a more complex nervous system than the Acoela, and are lightly cephalized, for instance having an eyespot above the brain, near the front end.
Complex active bodies
The philosopher Michael Trestman noted that three bilaterian phyla, namely the arthropods, the molluscs in the shape of the cephalopods, and the chordates, were distinctive in having "complex active bodies", something that the acoels and flatworms did not have. Any such animal, whe
Document 4:::
A plate in animal anatomy may refer to several things:
Flat bones (examples: bony plates, dermal plates) of vertebrates
an appendage of the Stegosauria group of dinosaurs
articulated armoured plates covering the head of thorax of Placodermi (literally "plate-skinned"), an extinct class of prehistoric fish (including skull, thoracic and tooth plates)
bony shields of the Ostracoderms (armored jawless fishes) such as the dermal head armour of members of the class Pteraspidomorphi that include dorsal, ventral, rostral and pineal plates
plates of a carapace, such as the dermal plates of the shell of a turtle
dermal plates partly or completely covering the body of the fish in the order Gasterosteiformes that includes the sticklebacks and relatives
plates of dermal bones of the armadillo
Zygomatic plate, a bony plate derived from the flattened front part of the zygomatic arch (cheekbone) in rodent anatomy
Other flat structures
hairy plate-like keratin scales of the pangolin
Basal plate (disambiguation), several anatomy-related meanings
Other meanings in human anatomy
Alar plate, a neural structure in the embryonic nervous system
Cribriform plate, of the ethmoid bone (horizontal lamina) received into the ethmoidal notch of the frontal bone and roofs in the nasal cavities
Epiphyseal plate, a hyaline cartilage plate in the metaphysis at each end of a long bone
Lateral pterygoid plate of the sphenoid, a broad, thin and everted bone that forms the lateral part of a horseshoe like process that extends from the inferior aspect of the sphenoid bone
Nail plate, the hard and translucent portion of the nail
Perpendicular plate of the ethmoid bone (vertical plate), a thin, flattened lamina, polygonal in form, which descends from the under surface of the cribriform plate, and assists in forming the septum of the nose
Related structures
Scute, a bony external plate or scale overlaid with horn, as on the shell of a turtle, the skin of crocodilians and the feet of
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is the term for chordates with a vertebral column and endoskeleton of cartilage and bone?
A. insect
B. am
C. mammel
D. vertebrate
Answer:
|
|
sciq-4732
|
multiple_choice
|
What kind of a resource is soil, for human purposes?
|
[
"emanates resource",
"renewable resource",
"nonrenewable resource",
"guarana resource"
] |
C
|
Relavent Documents:
Document 0:::
Soil fertility refers to the ability of soil to sustain agricultural plant growth, i.e. to provide plant habitat and result in sustained and consistent yields of high quality. It also refers to the soil's ability to supply plant/crop nutrients in the right quantities and qualities over a sustained period of time. A fertile soil has the following properties:
The ability to supply essential plant nutrients and water in adequate amounts and proportions for plant growth and reproduction; and
The absence of toxic substances which may inhibit plant growth e.g Fe2+ which leads to nutrient toxicity.
The following properties contribute to soil fertility in most situations:
Sufficient soil depth for adequate root growth and water retention;
Good internal drainage, allowing sufficient aeration for optimal root growth (although some plants, such as rice, tolerate waterlogging);
Topsoil or horizon O is with sufficient soil organic matter for healthy soil structure and soil moisture retention;
Soil pH in the range 5.5 to 7.0 (suitable for most plants but some prefer or tolerate more acid or alkaline conditions);
Adequate concentrations of essential plant nutrients in plant-available forms;
Presence of a range of microorganisms that support plant growth.
In lands used for agriculture and other human activities, maintenance of soil fertility typically requires the use of soil conservation practices. This is because soil erosion and other forms of soil degradation generally result in a decline in quality with respect to one or more of the aspects indicated above.
Soil fertilization
Bioavailable phosphorus (available to soil life) is the element in soil that is most often lacking. Nitrogen and potassium are also needed in substantial amounts. For this reason these three elements are always identified on a commercial fertilizer analysis. For example, a 10-10-15 fertilizer has 10 percent nitrogen, 10 percent available phosphorus (P2O5) and 15 percent water-soluble potassiu
Document 1:::
Resource refers to all the materials available in our environment which are technologically accessible, economically feasible and culturally sustainable and help us to satisfy our needs and wants. Resources can broadly be classified upon their availability — they are classified into renewable and non-renewable resources. They can also be classified as actual and potential on the basis of the level of development and use, on the basis of origin they can be classified as biotic and abiotic, and on the basis of their distribution, as ubiquitous and localised (private, community-owned, national and international resources). An item becomes a resource with time and developing technology. The benefits of resource utilization may include increased wealth, proper functioning of a system, or enhanced well-being. From a human perspective, a natural resource is anything obtained from the environment to satisfy human needs and wants. From a broader biological or ecological perspective, a resource satisfies the needs of a living organism (see biological resource).
The concept of resources has been developed across many established areas of work, in economics, biology and ecology, computer science, management, and human resources for example - linked to the concepts of competition, sustainability, conservation, and stewardship. In application within human society, commercial or non-commercial factors require resource allocation through resource management.
The concept of a resource can also be tied to the direction of leadership over resources, this can include the things leaders have responsibility for over the human resources, with management, help, support or direction such as in charge of a professional group, technical experts, innovative leaders, archiving expertise, academic management, association management, business management, healthcare management, military management, public administration, spiritual leadership and social networking administrator.
individuals exp
Document 2:::
USDA soil taxonomy (ST) developed by the United States Department of Agriculture and the National Cooperative Soil Survey provides an elaborate classification of soil types according to several parameters (most commonly their properties) and in several levels: Order, Suborder, Great Group, Subgroup, Family, and Series. The classification was originally developed by Guy Donald Smith, former director of the U.S. Department of Agriculture's soil survey investigations.
Discussion
A taxonomy is an arrangement in a systematic manner; the USDA soil taxonomy has six levels of classification. They are, from most general to specific: order, suborder, great group, subgroup, family and series. Soil properties that can be measured quantitatively are used in this classification system – they include: depth, moisture, temperature, texture, structure, cation exchange capacity, base saturation, clay mineralogy, organic matter content and salt content. There are 12 soil orders (the top hierarchical level) in soil taxonomy. The names of the orders end with the suffix -sol. The criteria for the different soil orders include properties that reflect major differences in the genesis of soils. The orders are:
Alfisol – soils with aluminium and iron. They have horizons of clay accumulation, and form where there is enough moisture and warmth for at least three months of plant growth. They constitute 10% of soils worldwide.
Andisol – volcanic ash soils. They are young soils. They cover 1% of the world's ice-free surface.
Aridisol – dry soils forming under desert conditions which have fewer than 90 consecutive days of moisture during the growing season and are nonleached. They include nearly 12% of soils on Earth. Soil formation is slow, and accumulated organic matter is scarce. They may have subsurface zones of caliche or duripan. Many aridisols have well-developed Bt horizons showing clay movement from past periods of greater moisture.
Entisol – recently formed soils that lack well-d
Document 3:::
Soil classification deals with the systematic categorization of soils based on distinguishing characteristics as well as criteria that dictate choices in use.
Overview
Soil classification is a dynamic subject, from the structure of the system, to the definitions of classes, to the application in the field. Soil classification can be approached from the perspective of soil as a material and soil as a resource.
Inscriptions at the temple of Horus at Edfu outline a soil classification used by Tanen to determine what kind of temple to build at which site. Ancient Greek scholars produced a number of classification based on several different qualities of the soil.
Engineering
Geotechnical engineers classify soils according to their engineering properties as they relate to use for foundation support or building material. Modern engineering classification systems are designed to allow an easy transition from field observations to basic predictions of soil engineering properties and behaviors.
The most common engineering classification system for soils in North America is the Unified Soil Classification System (USCS). The USCS has three major classification groups: (1) coarse-grained soils (e.g. sands and gravels); (2) fine-grained soils (e.g. silts and clays); and (3) highly organic soils (referred to as "peat"). The USCS further subdivides the three major soil classes for clarification. It distinguishes sands from gravels by grain size, classifying some as "well-graded" and the rest as "poorly-graded". Silts and clays are distinguished by the soils' Atterberg limits, and thus the soils are separated into "high-plasticity" and "low-plasticity" soils. Moderately organic soils are considered subdivisions of silts and clays and are distinguished from inorganic soils by changes in their plasticity properties (and Atterberg limits) on drying. The European soil classification system (ISO 14688) is very similar, differing primarily in coding and in adding an "intermediate-p
Document 4:::
The World Reference Base for Soil Resources (WRB) is an international soil classification system for naming soils and creating legends for soil maps. The currently valid version is the fourth edition 2022. It is edited by a working group of the International Union of Soil Sciences (IUSS).
Background
History
Since the 19th century, several countries developed national soil classification systems. During the 20th century, the need for an international soil classification system became more and more obvious.
From 1971 to 1981, the Food and Agriculture Organization (FAO) and UNESCO published the Soil Map of the World, 10 volumes, scale 1 : 5 M). The Legend for this map, published in 1974 under the leadership of Rudi Dudal, became the FAO soil classification. Many ideas from national soil classification systems were brought together in this worldwide-applicable system, among them the idea of diagnostic horizons as established in the '7th approximation to the USDA soil taxonomy' from 1960. The next step was the Revised Legend of the Soil Map of the World, published in 1988.
In 1982, the International Soil Science Society (ISSS; now: International Union of Soil Sciences, IUSS) established a working group named International Reference Base for Soil Classification (IRB). Chair of this working group was Ernst Schlichting. Its mandate was to develop an international soil classification system that should better consider soil-forming processes than the FAO soil classification. Drafts were presented in 1982 and 1990.
In 1992, the IRB working group decided to develop a new system named World Reference Base for Soil Resources (WRB) that should further develop the Revised Legend of the FAO soil classification and include some ideas of the more systematic IRB approach. Otto Spaargaren (International Soil Reference and Information Centre) and Freddy Nachtergaele (FAO) were nominated to prepare a draft. This draft was presented at the 15th World Congress of Soil Science in Acapu
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What kind of a resource is soil, for human purposes?
A. emanates resource
B. renewable resource
C. nonrenewable resource
D. guarana resource
Answer:
|
|
sciq-9557
|
multiple_choice
|
What is the smallest main part of the brain?
|
[
"the cerebellum",
"the stem",
"the cortex",
"the claw"
] |
B
|
Relavent Documents:
Document 0:::
The temporal lobe is one of the four major lobes of the cerebral cortex in the brain of mammals. The temporal lobe is located beneath the lateral fissure on both cerebral hemispheres of the mammalian brain.
The temporal lobe is involved in processing sensory input into derived meanings for the appropriate retention of visual memory, language comprehension, and emotion association.
Temporal refers to the head's temples.
Structure
The temporal lobe consists of structures that are vital for declarative or long-term memory. Declarative (denotative) or explicit memory is conscious memory divided into semantic memory (facts) and episodic memory (events). Medial temporal lobe structures that are critical for long-term memory include the hippocampus, along with the surrounding hippocampal region consisting of the perirhinal, parahippocampal, and entorhinal neocortical regions. The hippocampus is critical for memory formation, and the surrounding medial temporal cortex is currently theorized to be critical for memory storage. The prefrontal and visual cortices are also involved in explicit memory.
Research has shown that lesions in the hippocampus of monkeys results in limited impairment of function, whereas extensive lesions that include the hippocampus and the medial temporal cortex result in severe impairment.
Function
Visual memories
The temporal lobe communicates with the hippocampus and plays a key role in the formation of explicit long-term memory modulated by the amygdala.
Processing sensory input
Auditory Adjacent areas in the superior, posterior, and lateral parts of the temporal lobes are involved in high-level auditory processing. The temporal lobe is involved in primary auditory perception, such as hearing, and holds the primary auditory cortex. The primary auditory cortex receives sensory information from the ears and secondary areas process the information into meaningful units such as speech and words. The superior temporal gyrus includes an area (wit
Document 1:::
The human brain is the central organ of the human nervous system, and with the spinal cord makes up the central nervous system. The brain consists of the cerebrum, the brainstem and the cerebellum. It controls most of the activities of the body, processing, integrating, and coordinating the information it receives from the sense organs, and making decisions as to the instructions sent to the rest of the body. The brain is contained in, and protected by, the skull bones of the head.
The cerebrum, the largest part of the human brain, consists of two cerebral hemispheres. Each hemisphere has an inner core composed of white matter, and an outer surface – the cerebral cortex – composed of grey matter. The cortex has an outer layer, the neocortex, and an inner allocortex. The neocortex is made up of six neuronal layers, while the allocortex has three or four. Each hemisphere is conventionally divided into four lobes – the frontal, temporal, parietal, and occipital lobes. The frontal lobe is associated with executive functions including self-control, planning, reasoning, and abstract thought, while the occipital lobe is dedicated to vision. Within each lobe, cortical areas are associated with specific functions, such as the sensory, motor and association regions. Although the left and right hemispheres are broadly similar in shape and function, some functions are associated with one side, such as language in the left and visual-spatial ability in the right. The hemispheres are connected by commissural nerve tracts, the largest being the corpus callosum.
The cerebrum is connected by the brainstem to the spinal cord. The brainstem consists of the midbrain, the pons, and the medulla oblongata. The cerebellum is connected to the brainstem by three pairs of nerve tracts called cerebellar peduncles. Within the cerebrum is the ventricular system, consisting of four interconnected ventricles in which cerebrospinal fluid is produced and circulated. Underneath the cerebral cortex
Document 2:::
The following diagram is provided as an overview of and topical guide to the human nervous system:
Human nervous system – the part of the human body that coordinates a person's voluntary and involuntary actions and transmits signals between different parts of the body. The human nervous system consists of two main parts: the central nervous system (CNS) and the peripheral nervous system (PNS). The CNS contains the brain and spinal cord. The PNS consists mainly of nerves, which are long fibers that connect the CNS to every other part of the body. The PNS includes motor neurons, mediating voluntary movement; the autonomic nervous system, comprising the sympathetic nervous system and the parasympathetic nervous system and regulating involuntary functions; and the enteric nervous system, a semi-independent part of the nervous system whose function is to control the gastrointestinal system.
Evolution of the human nervous system
Evolution of nervous systems
Evolution of human intelligence
Evolution of the human brain
Paleoneurology
Some branches of science that study the human nervous system
Neuroscience
Neurology
Paleoneurology
Central nervous system
The central nervous system (CNS) is the largest part of the nervous system and includes the brain and spinal cord.
Spinal cord
Brain
Brain – center of the nervous system.
Outline of the human brain
List of regions of the human brain
Principal regions of the vertebrate brain:
Peripheral nervous system
Peripheral nervous system (PNS) – nervous system structures that do not lie within the CNS.
Sensory system
A sensory system is a part of the nervous system responsible for processing sensory information. A sensory system consists of sensory receptors, neural pathways, and parts of the brain involved in sensory perception.
List of sensory systems
Sensory neuron
Perception
Visual system
Auditory system
Somatosensory system
Vestibular system
Olfactory system
Taste
Pain
Components of the nervous system
Neuron
I
Document 3:::
The following are two lists of animals ordered by the size of their nervous system. The first list shows number of neurons in their entire nervous system, indicating their overall neural complexity. The second list shows the number of neurons in the structure that has been found to be representative of animal intelligence. The human brain contains 86 billion neurons, with 16 billion neurons in the cerebral cortex.
Scientists are engaged in counting, quantification, in order to find answers to the question in the strategy of neuroscience and intelligence of "self-knowledge": how the evolution of a set of components and parameters (~1011 neurons, ~1014 synapses) of a complex system could lead to the phenomenon of the appearance of intelligence in the biological species "sapiens".
Overview
Neurons are the cells that transmit information in an animal's nervous system so that it can sense stimuli from its environment and behave accordingly. Not all animals have neurons; Trichoplax and sponges lack nerve cells altogether.
Neurons may be packed to form structures such as the brain of vertebrates or the neural ganglions of insects.
The number of neurons and their relative abundance in different parts of the brain is a determinant of neural function and, consequently, of behavior.
Whole nervous system
All numbers for neurons (except Caenorhabditis and Ciona), and all numbers for synapses (except Ciona) are estimations.
List of animal species by forebrain (cerebrum or pallium) neuron number
The question of what physical characteristic of an animal makes an animal intelligent has varied over the centuries. One early speculation was brain size (or weight, which provides the same ordering.) A second proposal was brain-to-body-mass ratio, and a third was encephalization quotient, sometimes referred to as EQ. The current best predictor is number of neurons in the forebrain, based on Herculano-Houzel's improved neuron counts. It accounts most accurately for variations
Document 4:::
There is much to be discovered about the evolution of the brain and the principles that govern it. While much has been discovered, not everything currently known is well understood. The evolution of the brain has appeared to exhibit diverging adaptations within taxonomic classes such as Mammalia and more vastly diverse adaptations across other taxonomic classes.
Brain to body size scales allometrically. This means as body size changes, so do other physiological, anatomical, and biochemical constructs connecting the brain to the body. Small bodied mammals have relatively large brains compared to their bodies whereas large mammals (such as whales) have a smaller brain to body ratios. If brain weight is plotted against body weight for primates, the regression line of the sample points can indicate the brain power of a primate species. Lemurs for example fall below this line which means that for a primate of equivalent size, we would expect a larger brain size. Humans lie well above the line indicating that humans are more encephalized than lemurs. In fact, humans are more encephalized compared to all other primates. This means that human brains have exhibited a larger evolutionary increase in its complexity relative to its size. Some of these evolutionary changes have been found to be linked to multiple genetic factors, such as proteins and other organelles.
Early history of brain development
One approach to understanding overall brain evolution is to use a paleoarchaeological timeline to trace the necessity for ever increasing complexity in structures that allow for chemical and electrical signaling. Because brains and other soft tissues do not fossilize as readily as mineralized tissues, scientists often look to other structures as evidence in the fossil record to get an understanding of brain evolution. This, however, leads to a dilemma as the emergence of organisms with more complex nervous systems with protective bone or other protective tissues that can then
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is the smallest main part of the brain?
A. the cerebellum
B. the stem
C. the cortex
D. the claw
Answer:
|
|
sciq-6711
|
multiple_choice
|
Solutions of ionic compounds and melted ionic compounds conduct what?
|
[
"electricity",
"sound waves",
"light",
"radiation"
] |
A
|
Relavent Documents:
Document 0:::
An ionic liquid (IL) is a salt in the liquid state. In some contexts, the term has been restricted to salts whose melting point is below a specific temperature, such as . While ordinary liquids such as water and gasoline are predominantly made of electrically neutral molecules, ionic liquids are largely made of ions. These substances are variously called liquid electrolytes, ionic melts, ionic fluids, fused salts, liquid salts, or ionic glasses.
Ionic liquids have many potential applications. They are powerful solvents and can be used as electrolytes. Salts that are liquid at near-ambient temperature are important for electric battery applications, and have been considered as sealants due to their very low vapor pressure.
Any salt that melts without decomposing or vaporizing usually yields an ionic liquid. Sodium chloride (NaCl), for example, melts at into a liquid that consists largely of sodium cations () and chloride anions (). Conversely, when an ionic liquid is cooled, it often forms an ionic solid—which may be either crystalline or glassy.
The ionic bond is usually stronger than the Van der Waals forces between the molecules of ordinary liquids. Because of these strong interactions, salts tend to have high lattice energies, manifested in high melting points. Some salts, especially those with organic cations, have low lattice energies and thus are liquid at or below room temperature. Examples include compounds based on the 1-ethyl-3-methylimidazolium (EMIM) cation and include: EMIM:Cl, EMIMAc (acetate anion), EMIM dicyanamide, ()()·, that melts at ; and 1-butyl-3,5-dimethylpyridinium bromide which becomes a glass below .
Low-temperature ionic liquids can be compared to ionic solutions, liquids that contain both ions and neutral molecules, and in particular to the so-called deep eutectic solvents, mixtures of ionic and non-ionic solid substances which have much lower melting points than the pure compounds. Certain mixtures of nitrate salts can have melt
Document 1:::
Ionic transfer is the transfer of ions from one liquid phase to another. This is related to the phase transfer catalysts which are a special type of liquid-liquid extraction which is used in synthetic chemistry.
For instance nitrate anions can be transferred between water and nitrobenzene. One way to observe this is to use a cyclic voltammetry experiment where the liquid-liquid interface is the working electrode. This can be done by placing secondary electrodes in each phase and close to interface each phase has a reference electrode. One phase is attached to a potentiostat which is set to zero volts, while the other potentiostat is driven with a triangular wave. This experiment is known as a polarised Interface between Two Immiscible Electrolyte Solutions (ITIES) experiment.
See also
Diffusion potential
Document 2:::
The use of ionic liquids in carbon capture is a potential application of ionic liquids as absorbents for use in carbon capture and sequestration. Ionic liquids, which are salts that exist as liquids near room temperature, are polar, nonvolatile materials that have been considered for many applications. The urgency of climate change has spurred research into their use in energy-related applications such as carbon capture and storage.
Carbon capture using absorption
Ionic liquids as solvents
Amines are the most prevalent absorbent in postcombustion carbon capture technology today. In particular, monoethanolamine (MEA) has been used in industrial scales in postcombustion carbon capture, as well as in other CO2 separations, such as "sweetening" of natural gas. However, amines are corrosive, degrade over time, and require large industrial facilities. Ionic liquids on the other hand, have low vapor pressures . This property results from their strong Coulombic attractive force. Vapor pressure remains low through the substance's thermal decomposition point (typically >300 °C). In principle, this low vapor pressure simplifies their use and makes them "green" alternatives. Additionally, it reduces risk of contamination of the CO2 gas stream and of leakage into the environment.
The solubility of CO2 in ionic liquids is governed primarily by the anion, less so by the cation. The hexafluorophosphate (PF6–) and tetrafluoroborate (BF4–) anions have been shown to be especially amenable to CO2 capture.
Ionic liquids have been considered as solvents in a variety of liquid-liquid extraction processes, but never commercialized. Beside that, ionic liquids have replaced the conventional volatile solvents in industry such as absorption of gases or extractive distillation. Additionally, ionic liquids are used as co-solutes for the generation of aqueous biphasic systems, or purification of biomolecules.
Process
A typical CO2 absorption process consists of a feed gas, an absorptio
Document 3:::
An electrolyte is a medium containing ions that is electrically conducting through the movement of those ions, but not conducting electrons. This includes most soluble salts, acids, and bases dissolved in a polar solvent, such as water. Upon dissolving, the substance separates into cations and anions, which disperse uniformly throughout the solvent. Solid-state electrolytes also exist. In medicine and sometimes in chemistry, the term electrolyte refers to the substance that is dissolved.
Electrically, such a solution is neutral. If an electric potential is applied to such a solution, the cations of the solution are drawn to the electrode that has an abundance of electrons, while the anions are drawn to the electrode that has a deficit of electrons. The movement of anions and cations in opposite directions within the solution amounts to a current. Some gases, such as hydrogen chloride (HCl), under conditions of high temperature or low pressure can also function as electrolytes. Electrolyte solutions can also result from the dissolution of some biological (e.g., DNA, polypeptides) or synthetic polymers (e.g., polystyrene sulfonate), termed "polyelectrolytes", which contain charged functional groups. A substance that dissociates into ions in solution or in the melt acquires the capacity to conduct electricity. Sodium, potassium, chloride, calcium, magnesium, and phosphate in a liquid phase are examples of electrolytes.
In medicine, electrolyte replacement is needed when a person has prolonged vomiting or diarrhea, and as a response to sweating due to strenuous athletic activity. Commercial electrolyte solutions are available, particularly for sick children (such as oral rehydration solution, Suero Oral, or Pedialyte) and athletes (sports drinks). Electrolyte monitoring is important in the treatment of anorexia and bulimia.
In science, electrolytes are one of the main components of electrochemical cells.
In clinical medicine, mentions of electrolytes usually refer m
Document 4:::
Conductivity (or specific conductance) of an electrolyte solution is a measure of its ability to conduct electricity. The SI unit of conductivity is siemens per meter (S/m).
Conductivity measurements are used routinely in many industrial and environmental applications as a fast, inexpensive and reliable way of measuring the ionic content in a solution. For example, the measurement of product conductivity is a typical way to monitor and continuously trend the performance of water purification systems.
In many cases, conductivity is linked directly to the total dissolved solids (TDS). High quality deionized water has a conductivity of about 0.05 μS/cm at 25 °C, typical drinking water is in the range of 200–800 μS/cm, while sea water is about 50 mS/cm (or 0.05 S/cm).
Conductivity is traditionally determined by connecting the electrolyte in a Wheatstone bridge. Dilute solutions follow Kohlrausch's Laws of concentration dependence and additivity of ionic contributions. Lars Onsager gave a theoretical explanation of Kohlrausch's law by extending Debye–Hückel theory.
Units
The SI unit of conductivity is S/m and, unless otherwise qualified, it refers to 25 °C. More generally encountered is the traditional unit of μS/cm.
The commonly used standard cell has a width of 1 cm, and thus for very pure water in equilibrium with air would have a resistance of about 106 ohms, known as a megohm. Ultra-pure water could achieve 18 megohms or more. Thus in the past, megohm-cm was used, sometimes abbreviated to "megohm". Sometimes, conductivity is given in "microsiemens" (omitting the distance term in the unit). While this is an error, it can often be assumed to be equal to the traditional μS/cm. Often, by typographic limitations μS/cm is expressed as uS/cm.
The conversion of conductivity to the total dissolved solids depends on the chemical composition of the sample and can vary between 0.54 and 0.96. Typically, the conversion is done assuming that the solid is sodium chloride;
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Solutions of ionic compounds and melted ionic compounds conduct what?
A. electricity
B. sound waves
C. light
D. radiation
Answer:
|
|
scienceQA-6012
|
multiple_choice
|
What do these two changes have in common?
cutting your fingernails
stretching a rubber band
|
[
"Both are caused by cooling.",
"Both are caused by heating.",
"Both are chemical changes.",
"Both are only physical changes."
] |
D
|
Step 1: Think about each change.
Cutting your fingernails is a physical change. Your fingernails are shorter after you cut them. But the pieces are still made of the same type of matter as the uncut fingernails.
Stretching a rubber band is a physical change. The rubber band gets longer. But it is still made of the same type of matter as before.
Step 2: Look at each answer choice.
Both are only physical changes.
Both changes are physical changes. No new matter is created.
Both are chemical changes.
Both changes are physical changes. They are not chemical changes.
Both are caused by heating.
Neither change is caused by heating.
Both are caused by cooling.
Neither change is caused by cooling.
|
Relavent Documents:
Document 0:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 1:::
Physical changes are changes affecting the form of a chemical substance, but not its chemical composition. Physical changes are used to separate mixtures into their component compounds, but can not usually be used to separate compounds into chemical elements or simpler compounds.
Physical changes occur when objects or substances undergo a change that does not change their chemical composition. This contrasts with the concept of chemical change in which the composition of a substance changes or one or more substances combine or break up to form new substances. In general a physical change is reversible using physical means. For example, salt dissolved in water can be recovered by allowing the water to evaporate.
A physical change involves a change in physical properties. Examples of physical properties include melting, transition to a gas, change of strength, change of durability, changes to crystal form, textural change, shape, size, color, volume and density.
An example of a physical change is the process of tempering steel to form a knife blade. A steel blank is repeatedly heated and hammered which changes the hardness of the steel, its flexibility and its ability to maintain a sharp edge.
Many physical changes also involve the rearrangement of atoms most noticeably in the formation of crystals. Many chemical changes are irreversible, and many physical changes are reversible, but reversibility is not a certain criterion for classification. Although chemical changes may be recognized by an indication such as odor, color change, or production of a gas, every one of these indicators can result from physical change.
Examples
Heating and cooling
Many elements and some compounds change from solids to liquids and from liquids to gases when heated and the reverse when cooled. Some substances such as iodine and carbon dioxide go directly from solid to gas in a process called sublimation.
Magnetism
Ferro-magnetic materials can become magnetic. The process is reve
Document 2:::
A pre-STEM program is a course of study at any two-year college that prepares a student to transfer to a four-year school to earn a bachelor's degree in a STEM field.
Overview
The concept of a pre-STEM program is being developed to address America's need for more college-trained professionals in science, technology, engineering, and mathematics (STEM). It is an innovation meant to fill a gap at community colleges that do not have 'major' degree paths that students identify with on their way to earning an Associates degree. Students must complete a considerable amount of STEM coursework before transferring from a two-year school to a four-year school and earn a baccalaureate degree in a STEM field. Schools with a pre-STEM program are able to identify those students and support them with STEM-specific academic and career advising, increasing the student's chances of going on to earn a STEM baccalaureate degree in a timely fashion.
With over 50% of America's college-bound students starting their college career at public or private two-year school, and with a very small proportion of students who start college at a two-year school matriculating to and earning STEM degrees from four-year schools, pre-STEM programs have great potential for broadening participation in baccalaureate STEM studies.
Example programs
The effectiveness of pre-STEM programs is being investigated by a consortium of schools in Missouri: Moberly Area Community College, St. Charles Community College, Metropolitan Community College, and Truman State University.
A larger group of schools met at the Belknap Springs Meetings in October 2009 to discuss the challenges and opportunities presented by STEM-focused partnerships between 2-year and 4-year schools. Each program represented a two-year school and a four-year school that were trying to increase the number of people who earn a baccalaureate degree in a STEM area through various means, some of which were pre-STEM programs. Other methods includes
Document 3:::
Adaptive comparative judgement is a technique borrowed from psychophysics which is able to generate reliable results for educational assessment – as such it is an alternative to traditional exam script marking. In the approach, judges are presented with pairs of student work and are then asked to choose which is better, one or the other. By means of an iterative and adaptive algorithm, a scaled distribution of student work can then be obtained without reference to criteria.
Introduction
Traditional exam script marking began in Cambridge 1792 when, with undergraduate numbers rising, the importance of proper ranking of students was growing. So in 1792 the new Proctor of Examinations, William Farish, introduced marking, a process in which every examiner gives a numerical score to each response by every student, and the overall total mark puts the students in the final rank order. Francis Galton (1869) noted that, in an unidentified year about 1863, the Senior Wrangler scored 7,634 out of a maximum of 17,000, while the Second Wrangler scored 4,123. (The 'Wooden Spoon' scored only 237.)
Prior to 1792, a team of Cambridge examiners convened at 5pm on the last day of examining, reviewed the 19 papers each student had sat – and published their rank order at midnight. Marking solved the problems of numbers and prevented unfair personal bias, and its introduction was a step towards modern objective testing, the format it is best suited to. But the technology of testing that followed, with its major emphasis on reliability and the automatisation of marking, has been an uncomfortable partner for some areas of educational achievement: assessing writing or speaking, and other kinds of performance need something more qualitative and judgemental.
The technique of Adaptive Comparative Judgement is an alternative to marking. It returns to the pre-1792 idea of sorting papers according to their quality, but retains the guarantee of reliability and fairness. It is by far the most rel
Document 4:::
Engineering mathematics is a branch of applied mathematics concerning mathematical methods and techniques that are typically used in engineering and industry. Along with fields like engineering physics and engineering geology, both of which may belong in the wider category engineering science, engineering mathematics is an interdisciplinary subject motivated by engineers' needs both for practical, theoretical and other considerations outside their specialization, and to deal with constraints to be effective in their work.
Description
Historically, engineering mathematics consisted mostly of applied analysis, most notably: differential equations; real and complex analysis (including vector and tensor analysis); approximation theory (broadly construed, to include asymptotic, variational, and perturbative methods, representations, numerical analysis); Fourier analysis; potential theory; as well as linear algebra and applied probability, outside of analysis. These areas of mathematics were intimately tied to the development of Newtonian physics, and the mathematical physics of that period. This history also left a legacy: until the early 20th century subjects such as classical mechanics were often taught in applied mathematics departments at American universities, and fluid mechanics may still be taught in (applied) mathematics as well as engineering departments.
The success of modern numerical computer methods and software has led to the emergence of computational mathematics, computational science, and computational engineering (the last two are sometimes lumped together and abbreviated as CS&E), which occasionally use high-performance computing for the simulation of phenomena and the solution of problems in the sciences and engineering. These are often considered interdisciplinary fields, but are also of interest to engineering mathematics.
Specialized branches include engineering optimization and engineering statistics.
Engineering mathematics in tertiary educ
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What do these two changes have in common?
cutting your fingernails
stretching a rubber band
A. Both are caused by cooling.
B. Both are caused by heating.
C. Both are chemical changes.
D. Both are only physical changes.
Answer:
|
sciq-9198
|
multiple_choice
|
Today, new elements are usually named after famous scientists. the names of the elements can be cumbersome to write in full, especially when combined to form the names of compounds. therefore, each element name is abbreviated as a one- or two-letter chemical this?
|
[
"shorthand",
"title",
"symbol",
"notation"
] |
C
|
Relavent Documents:
Document 0:::
In chemical nomenclature, the IUPAC nomenclature of organic chemistry is a method of naming organic chemical compounds as recommended by the International Union of Pure and Applied Chemistry (IUPAC). It is published in the Nomenclature of Organic Chemistry (informally called the Blue Book). Ideally, every possible organic compound should have a name from which an unambiguous structural formula can be created. There is also an IUPAC nomenclature of inorganic chemistry.
To avoid long and tedious names in normal communication, the official IUPAC naming recommendations are not always followed in practice, except when it is necessary to give an unambiguous and absolute definition to a compound. IUPAC names can sometimes be simpler than older names, as with ethanol, instead of ethyl alcohol. For relatively simple molecules they can be more easily understood than non-systematic names, which must be learnt or looked over. However, the common or trivial name is often substantially shorter and clearer, and so preferred. These non-systematic names are often derived from an original source of the compound. Also, very long names may be less clear than structural formulas.
Basic principles
In chemistry, a number of prefixes, suffixes and infixes are used to describe the type and position of the functional groups in the compound.
The steps for naming an organic compound are:
Identification of the parent hydride parent hydrocarbon chain. This chain must obey the following rules, in order of precedence:
It should have the maximum number of substituents of the suffix functional group. By suffix, it is meant that the parent functional group should have a suffix, unlike halogen substituents. If more than one functional group is present, the one with highest group precedence should be used.
It should have the maximum number of multiple bonds.
It should have the maximum length.
It should have the maximum number of substituents or branches cited as prefixes
It should have the ma
Document 1:::
The names for the chemical elements 104 to 106 were the subject of a major controversy starting in the 1960s, described by some nuclear chemists as the Transfermium Wars because it concerned the elements following fermium (element 100) on the periodic table.
This controversy arose from disputes between American scientists and Soviet scientists as to which had first isolated these elements. The final resolution of this controversy in 1997 also decided the names of elements 107 to 109.
Controversy
By convention, naming rights for newly discovered chemical elements go to their discoverers. For elements 104, 105, and 106, there was a controversy between Soviet researchers at the Joint Institute for Nuclear Research and American researchers at Lawrence Berkeley National Laboratory regarding which group had discovered them first. Both parties suggested their own names for elements 104 and 105, not recognizing the other's name.
The American name of seaborgium for element 106 was also objectionable to some, because it referred to American chemist Glenn T. Seaborg who was still alive at the time this name was proposed. (Einsteinium and fermium had also been proposed as names of new elements while Einstein and Fermi were still living, but only made public after their deaths, due to Cold War secrecy.)
Opponents
The two principal groups which were involved in the conflict over element naming were:
An American group at Lawrence Berkeley Laboratory.
A Russian group at Joint Institute for Nuclear Research in Dubna.
and, as a kind of arbiter,
The IUPAC Commission on Nomenclature of Inorganic Chemistry, which introduced its own proposal to the IUPAC General Assembly.
The German group at the Gesellschaft für Schwerionenforschung (GSI) in Darmstadt, who had (undisputedly) discovered elements 107 to 109, were dragged into the controversy when the Commission suggested that the name "hahnium", proposed for element 105 by the Americans, be used for GSI's element 108 instead.
P
Document 2:::
This is an index of lists of molecules (i.e. by year, number of atoms, etc.). Millions of molecules have existed in the universe since before the formation of Earth. Three of them, carbon dioxide, water and oxygen were necessary for the growth of life. Although humanity had always been surrounded by these substances, it has not always known what they were composed of.
By century
The following is an index of list of molecules organized by time of discovery of their molecular formula or their specific molecule in case of isomers:
List of compounds
By number of carbon atoms in the molecule
List of compounds with carbon number 1
List of compounds with carbon number 2
List of compounds with carbon number 3
List of compounds with carbon number 4
List of compounds with carbon number 5
List of compounds with carbon number 6
List of compounds with carbon number 7
List of compounds with carbon number 8
List of compounds with carbon number 9
List of compounds with carbon number 10
List of compounds with carbon number 11
List of compounds with carbon number 12
List of compounds with carbon number 13
List of compounds with carbon number 14
List of compounds with carbon number 15
List of compounds with carbon number 16
List of compounds with carbon number 17
List of compounds with carbon number 18
List of compounds with carbon number 19
List of compounds with carbon number 20
List of compounds with carbon number 21
List of compounds with carbon number 22
List of compounds with carbon number 23
List of compounds with carbon number 24
List of compounds with carbon numbers 25-29
List of compounds with carbon numbers 30-39
List of compounds with carbon numbers 40-49
List of compounds with carbon numbers 50+
Other lists
List of interstellar and circumstellar molecules
List of gases
List of molecules with unusual names
See also
Molecule
Empirical formula
Chemical formula
Chemical structure
Chemical compound
Chemical bond
Coordination complex
L
Document 3:::
Steudel R 2020, Chemistry of the Non-metals: Syntheses - Structures - Bonding - Applications, in collaboration with D Scheschkewitz, Berlin, Walter de Gruyter, . ▲
An updated translation of the 5th German edition of 2013, incorporating the literature up to Spring 2019. Twenty-three nonmetals, including B, Si, Ge, As, Se, Te, and At but not Sb (nor Po). The nonmetals are identified on the basis of their electrical conductivity at absolute zero putatively being close to zero, rather than finite as in the case of metals. That does not work for As however, which has the electronic structure of a semimetal (like Sb).
Halka M & Nordstrom B 2010, "Nonmetals", Facts on File, New York,
A reading level 9+ book covering H, C, N, O, P, S, Se. Complementary books by the same authors examine (a) the post-transition metals (Al, Ga, In, Tl, Sn, Pb and Bi) and metalloids (B, Si, Ge, As, Sb, Te and Po); and (b) the halogens and noble gases.
Woolins JD 1988, Non-Metal Rings, Cages and Clusters, John Wiley & Sons, Chichester, .
A more advanced text that covers H; B; C, Si, Ge; N, P, As, Sb; O, S, Se and Te.
Steudel R 1977, Chemistry of the Non-metals: With an Introduction to Atomic Structure and Chemical Bonding, English edition by FC Nachod & JJ Zuckerman, Berlin, Walter de Gruyter, . ▲
Twenty-four nonmetals, including B, Si, Ge, As, Se, Te, Po and At.
Powell P & Timms PL 1974, The Chemistry of the Non-metals, Chapman & Hall, London, . ▲
Twenty-two nonmetals including B, Si, Ge, As and Te. Tin and antimony are shown as being intermediate between metals and nonmetals; they are later shown as either metals or nonmetals. Astatine is counted as a metal.
Document 4:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Today, new elements are usually named after famous scientists. the names of the elements can be cumbersome to write in full, especially when combined to form the names of compounds. therefore, each element name is abbreviated as a one- or two-letter chemical this?
A. shorthand
B. title
C. symbol
D. notation
Answer:
|
|
sciq-10077
|
multiple_choice
|
The tiny hair present on mucus membranes is known as?
|
[
"cilia",
"dendrites",
"collagen",
"tissue"
] |
A
|
Relavent Documents:
Document 0:::
Mucus ( ) is a slippery aqueous secretion produced by, and covering, mucous membranes. It is typically produced from cells found in mucous glands, although it may also originate from mixed glands, which contain both serous and mucous cells. It is a viscous colloid containing inorganic salts, antimicrobial enzymes (such as lysozymes), immunoglobulins (especially IgA), and glycoproteins such as lactoferrin and mucins, which are produced by goblet cells in the mucous membranes and submucosal glands. Mucus serves to protect epithelial cells in the linings of the respiratory, digestive, and urogenital systems, and structures in the visual and auditory systems from pathogenic fungi, bacteria and viruses. Most of the mucus in the body is produced in the gastrointestinal tract.
Amphibians, fish, snails, slugs, and some other invertebrates also produce external mucus from their epidermis as protection against pathogens, and to help in movement and is also produced in fish to line their gills. Plants produce a similar substance called mucilage that is also produced by some microorganisms.
Respiratory system
In the human respiratory system, mucus is part of the airway surface liquid (ASL), also known as epithelial lining fluid (ELF), that lines most of the respiratory tract. The airway surface liquid consists of a sol layer termed the periciliary liquid layer and an overlying gel layer termed the mucus layer. The periciliary liquid layer is so named as it surrounds the cilia and lies on top of the surface epithelium. The periciliary liquid layer surrounding the cilia consists of a gel meshwork of cell-tethered mucins and polysaccharides. The mucus blanket aids in the protection of the lungs by trapping foreign particles before they enter them, in particular through the nose during normal breathing.
Mucus is made up of a fluid component of around 95% water, the mucin secretions from the goblet cells, and the submucosal glands (2–3% glycoproteins), proteoglycans (0.1–0.5%),
Document 1:::
H2.00.04.4.01001: Lymphoid tissue
H2.00.05.0.00001: Muscle tissue
H2.00.05.1.00001: Smooth muscle tissue
H2.00.05.2.00001: Striated muscle tissue
H2.00.06.0.00001: Nerve tissue
H2.00.06.1.00001: Neuron
H2.00.06.2.00001: Synapse
H2.00.06.2.00001: Neuroglia
h3.01: Bones
h3.02: Joints
h3.03: Muscles
h3.04: Alimentary system
h3.05: Respiratory system
h3.06: Urinary system
h3.07: Genital system
h3.08:
Document 2:::
The posterior surfaces of the ciliary processes are covered by a bilaminar layer of black pigment cells, which is continued forward from the retina, and is named the pars ciliaris retinae.
Document 3:::
A mucous membrane or mucosa is a membrane that lines various cavities in the body of an organism and covers the surface of internal organs. It consists of one or more layers of epithelial cells overlying a layer of loose connective tissue. It is mostly of endodermal origin and is continuous with the skin at body openings such as the eyes, eyelids, ears, inside the nose, inside the mouth, lips, the genital areas, the urethral opening and the anus. Some mucous membranes secrete mucus, a thick protective fluid. The function of the membrane is to stop pathogens and dirt from entering the body and to prevent bodily tissues from becoming dehydrated.
Structure
The mucosa is composed of one or more layers of epithelial cells that secrete mucus, and an underlying lamina propria of loose connective tissue. The type of cells and type of mucus secreted vary from organ to organ and each can differ along a given tract.
Mucous membranes line the digestive, respiratory and reproductive tracts and are the primary barrier between the external world and the interior of the body; in an adult human the total surface area of the mucosa is about 400 square meters while the surface area of the skin is about 2 square meters. Along with providing a physical barrier, they also contain key parts of the immune system and serve as the interface between the body proper and the microbiome.
Examples
Some examples include:
Endometrium: the mucosa of the uterus
Gastric mucosa
Intestinal mucosa
Nasal mucosa
Olfactory mucosa
Oral mucosa
Penile mucosa
Respiratory mucosa
Vaginal mucosa
Frenulum of tongue
Anal canal
Conjunctiva
Development
Developmentally, the majority of mucous membranes are of endodermal origin. Exceptions include the palate, cheeks, floor of the mouth, gums, lips and the portion of the anal canal below the pectinate line, which are all ectodermal in origin.
Function
One of its functions is to keep the tissue moist (for example in the respiratory tract, including the mouth and nose
Document 4:::
In mammals, trichocytes are the specialized epithelial cells from which the highly mechanically resilient tissues hair and nails are formed. They can be identified by the fact that they express "hard", "trichocyte" or "hair" keratin proteins. These are modified keratins containing large amounts of the amino acid cysteine, which facilitates chemical cross-linking of these proteins to form the tough material from which hair and nail is composed. These cells give rise to non-hair non-keratinized IRSC (inner root sheath cell) as well.
See also
List of human cell types derived from the germ layers
List of distinct cell types in the adult human body
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
The tiny hair present on mucus membranes is known as?
A. cilia
B. dendrites
C. collagen
D. tissue
Answer:
|
|
sciq-1888
|
multiple_choice
|
What type of symmetry do echinoderms show?
|
[
"lateral",
"mirrored",
"parallel",
"radial"
] |
D
|
Relavent Documents:
Document 0:::
Symmetry in biology refers to the symmetry observed in organisms, including plants, animals, fungi, and bacteria. External symmetry can be easily seen by just looking at an organism. For example, the face of a human being has a plane of symmetry down its centre, or a pine cone displays a clear symmetrical spiral pattern. Internal features can also show symmetry, for example the tubes in the human body (responsible for transporting gases, nutrients, and waste products) which are cylindrical and have several planes of symmetry.
Biological symmetry can be thought of as a balanced distribution of duplicate body parts or shapes within the body of an organism. Importantly, unlike in mathematics, symmetry in biology is always approximate. For example, plant leaves – while considered symmetrical – rarely match up exactly when folded in half. Symmetry is one class of patterns in nature whereby there is near-repetition of the pattern element, either by reflection or rotation.
While sponges and placozoans represent two groups of animals which do not show any symmetry (i.e. are asymmetrical), the body plans of most multicellular organisms exhibit, and are defined by, some form of symmetry. There are only a few types of symmetry which are possible in body plans. These are radial (cylindrical), bilateral, biradial and spherical symmetry. While the classification of viruses as an "organism" remains controversial, viruses also contain icosahedral symmetry.
The importance of symmetry is illustrated by the fact that groups of animals have traditionally been defined by this feature in taxonomic groupings. The Radiata, animals with radial symmetry, formed one of the four branches of Georges Cuvier's classification of the animal kingdom. Meanwhile, Bilateria is a taxonomic grouping still used today to represent organisms with embryonic bilateral symmetry.
Radial symmetry
Organisms with radial symmetry show a repeating pattern around a central axis such that they can be separated in
Document 1:::
Asymmetry is the absence of, or a violation of, symmetry (the property of an object being invariant to a transformation, such as reflection). Symmetry is an important property of both physical and abstract systems and it may be displayed in precise terms or in more aesthetic terms. The absence of or violation of symmetry that are either expected or desired can have important consequences for a system.
In organisms
Due to how cells divide in organisms, asymmetry in organisms is fairly usual in at least one dimension, with biological symmetry also being common in at least one dimension.
Louis Pasteur proposed that biological molecules are asymmetric because the cosmic [i.e. physical] forces that preside over their formation are themselves asymmetric. While at his time, and even now, the symmetry of physical processes are highlighted, it is known that there are fundamental physical asymmetries, starting with time.
Asymmetry in biology
Asymmetry is an important and widespread trait, having evolved numerous times in many organisms and at many levels of organisation (ranging from individual cells, through organs, to entire body-shapes). Benefits of asymmetry sometimes have to do with improved spatial arrangements, such as the left human lung being smaller, and having one fewer lobes than the right lung to make room for the asymmetrical heart. In other examples, division of function between the right and left half may have been beneficial and has driven the asymmetry to become stronger. Such an explanation is usually given for mammal hand or paw preference (handedness), an asymmetry in skill development in mammals. Training the neural pathways in a skill with one hand (or paw) may take less effort than doing the same with both hands.
Nature also provides several examples of handedness in traits that are usually symmetric. The following are examples of animals with obvious left-right asymmetries:
Most snails, because of torsion during development, show remarkable as
Document 2:::
Dipleurula is a hypothetical larva of the ancestral echinoderm. It represents the type of basis of all larval forms of, at least, the eleutherozoans (all echinoderms except crinoids), where the starfish, sea urchins, sea cucumbers and brittle stars belong. The dipleurula is a bilaterally symmetrical, ciliated echinoderm larva (cilia devoted to movement, feeding and perception).
Etymology
Derives from Ancient Greek di, meaning 'two' and the small form of pleura (pleurula), meaning 'little side', i.e. 'little, two-sided [larva]'.
History
Although the term dipleurula stems from Semon (1888),
it was first systematically introduced, described and drawn by Bather (1900) in his monography on the echinoderms. The name dipleurula, two-sided, was given to stress the fact that the larva of the typically five-rayed, (approximately) radially symmetric adults show a bilateral structure. It was this bilateral structure of the larvae that identified echinoderms as bilaterian animals.
The original doliolaria schema shows a benthic, crawling, larva. However Bather could not have known yet that larval echinoderms are typically pelagic (free-floating plankton).
Structure
The hypothetical dipleurula larva bears resemblance to stages of all extant echinoderms, such as the bipinnaria and the brachiolaria of the starfish, the auricularia of the sea cucumbers, the echinopluteus larva of the sea urchins, and the ophiopluteus of the brittle stars. Also the doliolaria of the crinoids (sea-lilies and feather stars) can be attributed to the same basic pattern. Note, that the extant echinoderms represent just a small window on the extraordinary diversity of early echinoderms as known from their rich fossil record. Although there is current research on fossilized larval skeletons, the doliolaria remains a hypothesis.
The doliolaria shows a three-partite body, each of which develop paired coelomic spaces which originate from the enterocoel. This, and their close resemblance to the torn
Document 3:::
Cephalization is an evolutionary trend in which, over many generations, the mouth, sense organs, and nerve ganglia become concentrated at the front end of an animal, producing a head region. This is associated with movement and bilateral symmetry, such that the animal has a definite head end. This led to the formation of a highly sophisticated brain in three groups of animals, namely the arthropods, cephalopod molluscs, and vertebrates.
Animals without bilateral symmetry
Cnidaria, such as the radially symmetrical Hydrozoa, show some degree of cephalization. The Anthomedusae have a head end with their mouth, photoreceptive cells, and a concentration of neural cells.
Bilateria
Cephalization is a characteristic feature of the Bilateria, a large group containing the majority of animal phyla. These have the ability to move, using muscles, and a body plan with a front end that encounters stimuli first as the animal moves forwards, and accordingly has evolved to contain many of the body's sense organs, able to detect light, chemicals, and gravity. There is often also a collection of nerve cells able to process the information from these sense organs, forming a brain in several phyla and one or more ganglia in others.
Acoela
The Acoela are basal bilaterians, part of the Xenacoelomorpha. They are small and simple animals, and have very slightly more nerve cells at the head end than elsewhere, not forming a distinct and compact brain. This represents an early stage in cephalization.
Flatworms
The Platyhelminthes (flatworms) have a more complex nervous system than the Acoela, and are lightly cephalized, for instance having an eyespot above the brain, near the front end.
Complex active bodies
The philosopher Michael Trestman noted that three bilaterian phyla, namely the arthropods, the molluscs in the shape of the cephalopods, and the chordates, were distinctive in having "complex active bodies", something that the acoels and flatworms did not have. Any such animal, whe
Document 4:::
Symmetry breaking in biology is the process by which uniformity is broken, or the number of points to view invariance are reduced, to generate a more structured and improbable state. Symmetry breaking is the event where symmetry along a particular axis is lost to establish a polarity. Polarity is a measure for a biological system to distinguish poles along an axis. This measure is important because it is the first step to building complexity. For example, during organismal development, one of the first steps for the embryo is to distinguish its dorsal-ventral axis. The symmetry-breaking event that occurs here will determine which end of this axis will be the ventral side, and which end will be the dorsal side. Once this distinction is made, then all the structures that are located along this axis can develop at the proper location. As an example, during human development, the embryo needs to establish where is ‘back’ and where is ‘front’ before complex structures, such as the spine and lungs, can develop in the right location (where the lungs are placed ‘in front’ of the spine). This relationship between symmetry breaking and complexity was articulated by P.W. Anderson. He speculated that increasing levels of broken symmetry in many-body systems correlates with increasing complexity and functional specialization. In a biological perspective, the more complex an organism is, the higher number of symmetry-breaking events can be found.
The importance of symmetry breaking in biology is also reflected in the fact that it's found at all scales. Symmetry breaking can be found at the macromolecular level, at the subcellular level and even at the tissues and organ level. It's also interesting to note that most asymmetry on a higher scale is a reflection of symmetry breaking on a lower scale. Cells first need to establish a polarity through a symmetry-breaking event before tissues and organs themselves can be polar. For example, one model proposes that left-right bo
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What type of symmetry do echinoderms show?
A. lateral
B. mirrored
C. parallel
D. radial
Answer:
|
|
sciq-2790
|
multiple_choice
|
Molecules are represented by symbols that all who agree on?
|
[
"geologists",
"astrologists",
"chemists",
"physics"
] |
C
|
Relavent Documents:
Document 0:::
The Science, Technology, Engineering and Mathematics Network or STEMNET is an educational charity in the United Kingdom that seeks to encourage participation at school and college in science and engineering-related subjects (science, technology, engineering, and mathematics) and (eventually) work.
History
It is based at Woolgate Exchange near Moorgate tube station in London and was established in 1996. The chief executive is Kirsten Bodley. The STEMNET offices are housed within the Engineering Council.
Function
Its chief aim is to interest children in science, technology, engineering and mathematics. Primary school children can start to have an interest in these subjects, leading secondary school pupils to choose science A levels, which will lead to a science career. It supports the After School Science and Engineering Clubs at schools. There are also nine regional Science Learning Centres.
STEM ambassadors
To promote STEM subjects and encourage young people to take up jobs in these areas, STEMNET have around 30,000 ambassadors across the UK. these come from a wide selection of the STEM industries and include TV personalities like Rob Bell.
Funding
STEMNET used to receive funding from the Department for Education and Skills. Since June 2007, it receives funding from the Department for Children, Schools and Families and Department for Innovation, Universities and Skills, since STEMNET sits on the chronological dividing point (age 16) of both of the new departments.
See also
The WISE Campaign
Engineering and Physical Sciences Research Council
National Centre for Excellence in Teaching Mathematics
Association for Science Education
Glossary of areas of mathematics
Glossary of astronomy
Glossary of biology
Glossary of chemistry
Glossary of engineering
Glossary of physics
Document 1:::
The STEM (Science, Technology, Engineering, and Mathematics) pipeline is a critical infrastructure for fostering the development of future scientists, engineers, and problem solvers. It's the educational and career pathway that guides individuals from early childhood through to advanced research and innovation in STEM-related fields.
Description
The "pipeline" metaphor is based on the idea that having sufficient graduates requires both having sufficient input of students at the beginning of their studies, and retaining these students through completion of their academic program. The STEM pipeline is a key component of workplace diversity and of workforce development that ensures sufficient qualified candidates are available to fill scientific and technical positions.
The STEM pipeline was promoted in the United States from the 1970s onwards, as “the push for STEM (science, technology, engineering, and mathematics) education appears to have grown from a concern for the low number of future professionals to fill STEM jobs and careers and economic and educational competitiveness.”
Today, this metaphor is commonly used to describe retention problems in STEM fields, called “leaks” in the pipeline. For example, the White House reported in 2012 that 80% of minority groups and women who enroll in a STEM field switch to a non-STEM field or drop out during their undergraduate education. These leaks often vary by field, gender, ethnic and racial identity, socioeconomic background, and other factors, drawing attention to structural inequities involved in STEM education and careers.
Current efforts
The STEM pipeline concept is a useful tool for programs aiming at increasing the total number of graduates, and is especially important in efforts to increase the number of underrepresented minorities and women in STEM fields. Using STEM methodology, educational policymakers can examine the quantity and retention of students at all stages of the K–12 educational process and beyo
Document 2:::
Female education in STEM refers to child and adult female representation in the educational fields of science, technology, engineering, and mathematics (STEM). In 2017, 33% of students in STEM fields were women.
The organization UNESCO has stated that this gender disparity is due to discrimination, biases, social norms and expectations that influence the quality of education women receive and the subjects they study. UNESCO also believes that having more women in STEM fields is desirable because it would help bring about sustainable development.
Current status of girls and women in STEM education
Overall trends in STEM education
Gender differences in STEM education participation are already visible in early childhood care and education in science- and math-related play, and become more pronounced at higher levels of education. Girls appear to lose interest in STEM subjects with age, particularly between early and late adolescence. This decreased interest affects participation in advanced studies at the secondary level and in higher education. Female students represent 35% of all students enrolled in STEM-related fields of study at this level globally. Differences are also observed by disciplines, with female enrollment lowest in engineering, manufacturing and construction, natural science, mathematics and statistics and ICT fields. Significant regional and country differences in female representation in STEM studies can be observed, though, suggesting the presence of contextual factors affecting girls’ and women's engagement in these fields. Women leave STEM disciplines in disproportionate numbers during their higher education studies, in their transition to the world of work and even in their career cycle.
Learning achievement in STEM education
Data on gender differences in learning achievement present a complex picture, depending on what is measured (subject, knowledge acquisition against knowledge application), the level of education/age of students, and
Document 3:::
In mathematical psychology and education theory, a knowledge space is a combinatorial structure used to formulate mathematical models describing the progression of a human learner. Knowledge spaces were introduced in 1985 by Jean-Paul Doignon and Jean-Claude Falmagne, and remain in extensive use in the education theory. Modern applications include two computerized tutoring systems, ALEKS and the defunct RATH.
Formally, a knowledge space assumes that a domain of knowledge is a collection of concepts or skills, each of which must be eventually mastered. Not all concepts are interchangeable; some require other concepts as prerequisites. Conversely, competency at one skill may ease the acquisition of another through similarity. A knowledge space marks out which collections of skills are feasible: they can be learned without mastering any other skills. Under reasonable assumptions, the collection of feasible competencies forms the mathematical structure known as an antimatroid.
Researchers and educators usually explore the structure of a discipline's knowledge space as a latent class model.
Motivation
Knowledge Space Theory attempts to address shortcomings of standardized testing when used in educational psychometry. Common tests, such as the SAT and ACT, compress a student's knowledge into a very small range of ordinal ranks, in the process effacing the conceptual dependencies between questions. Consequently, the tests cannot distinguish between true understanding and guesses, nor can they identify a student's particular weaknesses, only the general proportion of skills mastered. The goal of knowledge space theory is to provide a language by which exams can communicate
What the student can do and
What the student is ready to learn.
Model structure
Knowledge Space Theory-based models presume that an educational subject can be modeled as a finite set of concepts, skills, or topics. Each feasible state of knowledge about is then a subset of ; the set of
Document 4:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Molecules are represented by symbols that all who agree on?
A. geologists
B. astrologists
C. chemists
D. physics
Answer:
|
|
sciq-2573
|
multiple_choice
|
Glasses are mixtures of oxides, the main component of which is silica (sio2). silica is called the glass former, while additives are referred to as this?
|
[
"addition modifiers",
"glass modifiers",
"natural modifiers",
"glass actors"
] |
B
|
Relavent Documents:
Document 0:::
This is a list of some physical properties of common glasses. Unless otherwise stated, the technical glass compositions and many experimentally determined properties are taken from one large study. Unless stated otherwise, the properties of fused silica (quartz glass) and germania glass are derived from the SciGlass glass database by forming the arithmetic mean of all the experimental values from different authors (in general more than 10 independent sources for quartz glass and Tg of germanium oxide glass).
The list is not exhaustive.
Document 1:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 2:::
In mathematical psychology and education theory, a knowledge space is a combinatorial structure used to formulate mathematical models describing the progression of a human learner. Knowledge spaces were introduced in 1985 by Jean-Paul Doignon and Jean-Claude Falmagne, and remain in extensive use in the education theory. Modern applications include two computerized tutoring systems, ALEKS and the defunct RATH.
Formally, a knowledge space assumes that a domain of knowledge is a collection of concepts or skills, each of which must be eventually mastered. Not all concepts are interchangeable; some require other concepts as prerequisites. Conversely, competency at one skill may ease the acquisition of another through similarity. A knowledge space marks out which collections of skills are feasible: they can be learned without mastering any other skills. Under reasonable assumptions, the collection of feasible competencies forms the mathematical structure known as an antimatroid.
Researchers and educators usually explore the structure of a discipline's knowledge space as a latent class model.
Motivation
Knowledge Space Theory attempts to address shortcomings of standardized testing when used in educational psychometry. Common tests, such as the SAT and ACT, compress a student's knowledge into a very small range of ordinal ranks, in the process effacing the conceptual dependencies between questions. Consequently, the tests cannot distinguish between true understanding and guesses, nor can they identify a student's particular weaknesses, only the general proportion of skills mastered. The goal of knowledge space theory is to provide a language by which exams can communicate
What the student can do and
What the student is ready to learn.
Model structure
Knowledge Space Theory-based models presume that an educational subject can be modeled as a finite set of concepts, skills, or topics. Each feasible state of knowledge about is then a subset of ; the set of
Document 3:::
Glass-like carbon, often called glassy carbon or vitreous carbon, is a non-graphitizing, or nongraphitizable, carbon which combines glassy and ceramic properties with those of graphite. The most important properties are high thermal stability, high thermal conductivity, hardness (7 Mohs), low density, low electrical resistance, low friction, extreme resistance to chemical attack, and impermeability to gases and liquids. Glassy carbon is widely used as an electrode material in electrochemistry, for high-temperature crucibles, and as a component of some prosthetic devices. It can be fabricated in different shapes, sizes and sections.
The names glassy carbon and vitreous carbon have been registered as trademarks, and IUPAC does not recommend their use as technical terms.
A historical review of glassy carbon was published in 2021.
History
Glassy carbon was first observed in the laboratories of The Carborundum Company, Manchester, UK, in the mid-1950s by Bernard Redfern, a materials scientist and diamond technologist. He noticed that Sellotape he used to hold ceramic (rocket nozzle) samples in a furnace maintained a sort of structural identity after firing in an inert atmosphere. He searched for a polymer matrix to mirror a diamond structure and discovered a resole resin that would, with special preparation, set without a catalyst. Crucibles were produced with this phenolic resin, and distributed to organisations such as UKAEA Harwell.
Redfern left The Carborundum Co., which officially wrote off all interests in the glassy carbon invention. While working at the Plessey Company laboratory (in a disused church) in Towcester, UK, Redfern received a glassy carbon crucible for duplication from UKAEA. He identified it as one he had made from markings he had engraved into the uncured precursor prior to carbonisation—it is almost impossible to engrave the finished product. The Plessey Company set up a laboratory, first in a factory previously used to make briar pipes in Li
Document 4:::
Precision glass moulding is a replicative process that allows the production of high precision optical components from glass without grinding and polishing. The process is also known as ultra-precision glass pressing. It is used to manufacture precision glass lenses for consumer products such as digital cameras, and high-end products like medical systems. The main advantage over mechanical lens production is that complex lens geometries such as aspheres can be produced cost-efficiently.
Process
The precision glass moulding process consists of six steps:
The glass blank is loaded into the lower side of the moulding tool.
Oxygen is removed from the working area by filling with nitrogen and/or evacuation of the process chamber.
The tool system is nearly closed (no contact of the upper mould) and the entire system of mould, die and glass is heated up. Infrared lamps are used for heating in most systems.
After reaching the working temperature, which is between the transition temperature and the softening point of the glass, the moulds close further and start pressing the glass in a travel-controlled process.
When the final thickness of the part has been achieved, the pressing switches over to a force-controlled process.
After moulding has been completed, the glass is cooled down and the working environment is filled with nitrogen. When the lens has cooled to the point where it can be handled, it is removed from the tool.
The process is executed on a specialized moulding machine, which precisely controls the temperature, travel, and force during the process. The tools used must withstand high temperatures and pressures, and need to be resistant to chemical interaction with the glass. The mold materials also have to be suitable for machining into the precise surface profiles.
Process chain
In order to ensure the desired quality the parts are measured between each process step. Additionally, the parts are handled and transported carefully between the processing and met
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Glasses are mixtures of oxides, the main component of which is silica (sio2). silica is called the glass former, while additives are referred to as this?
A. addition modifiers
B. glass modifiers
C. natural modifiers
D. glass actors
Answer:
|
|
sciq-5695
|
multiple_choice
|
What are non-renewable sources of energy that produce environmental damage?
|
[
"fossil fuels",
"crop fuels",
"wind",
"sunlight"
] |
A
|
Relavent Documents:
Document 0:::
A non-renewable resource (also called a finite resource) is a natural resource that cannot be readily replaced by natural means at a pace quick enough to keep up with consumption. An example is carbon-based fossil fuels. The original organic matter, with the aid of heat and pressure, becomes a fuel such as oil or gas. Earth minerals and metal ores, fossil fuels (coal, petroleum, natural gas) and groundwater in certain aquifers are all considered non-renewable resources, though individual elements are always conserved (except in nuclear reactions, nuclear decay or atmospheric escape).
Conversely, resources such as timber (when harvested sustainably) and wind (used to power energy conversion systems) are considered renewable resources, largely because their localized replenishment can occur within time frames meaningful to humans as well.
Earth minerals and metal ores
Earth minerals and metal ores are examples of non-renewable resources. The metals themselves are present in vast amounts in Earth's crust, and their extraction by humans only occurs where they are concentrated by natural geological processes (such as heat, pressure, organic activity, weathering and other processes) enough to become economically viable to extract. These processes generally take from tens of thousands to millions of years, through plate tectonics, tectonic subsidence and crustal recycling.
The localized deposits of metal ores near the surface which can be extracted economically by humans are non-renewable in human time-frames. There are certain rare earth minerals and elements that are more scarce and exhaustible than others. These are in high demand in manufacturing, particularly for the electronics industry.
Fossil fuels
Natural resources such as coal, petroleum(crude oil) and natural gas take thousands of years to form naturally and cannot be replaced as fast as they are being consumed. Eventually it is considered that fossil-based resources will become too costly to harvest and
Document 1:::
Energy quality is a measure of the ease with which a form of energy can be converted to useful work or to another form of energy: i.e. its content of thermodynamic free energy. A high quality form of energy has a high content of thermodynamic free energy, and therefore a high proportion of it can be converted to work; whereas with low quality forms of energy, only a small proportion can be converted to work, and the remainder is dissipated as heat. The concept of energy quality is also used in ecology, where it is used to track the flow of energy between different trophic levels in a food chain and in thermoeconomics, where it is used as a measure of economic output per unit of energy. Methods of evaluating energy quality often involve developing a ranking of energy qualities in hierarchical order.
Examples: Industrialization, Biology
The consideration of energy quality was a fundamental driver of industrialization from the 18th through 20th centuries. Consider for example the industrialization of New England in the 18th century. This refers to the construction of textile mills containing power looms for weaving cloth. The simplest, most economical and straightforward source of energy was provided by water wheels, extracting energy from a millpond behind a dam on a local creek. If another nearby landowner also decided to build a mill on the same creek, the construction of their dam would lower the overall hydraulic head to power the existing waterwheel, thus hurting power generation and efficiency. This eventually became an issue endemic to the entire region, reducing the overall profitability of older mills as newer ones were built. The search for higher quality energy was a major impetus throughout the 19th and 20th centuries. For example, burning coal to make steam to generate mechanical energy would not have been imaginable in the 18th century; by the end of the 19th century, the use of water wheels was long outmoded. Similarly, the quality of energy from elec
Document 2:::
Primary energy (PE) is the energy found in nature that has not been subjected to any human engineered conversion process. It encompasses energy contained in raw fuels and other forms of energy, including waste, received as input to a system. Primary energy can be non-renewable or renewable.
Total primary energy supply (TPES) is the sum of production and imports, plus or minus stock changes, minus exports and international bunker storage.
The International Recommendations for Energy Statistics (IRES) prefers total energy supply (TES) to refer to this indicator. These expressions are often used to describe the total energy supply of a national territory.
Secondary energy is a carrier of energy, such as electricity. These are produced by conversion from a primary energy source.
Primary energy is used as a measure in energy statistics in the compilation of energy balances, as well as in the field of energetics. In energetics, a primary energy source (PES) refers to the energy forms required by the energy sector to generate the supply of energy carriers used by human society. Primary energy only counts raw energy and not usable energy and fails to account well for energy losses, particularly the large losses in thermal sources. It therefore generally grossly undercounts non thermal renewable energy sources .
Examples of sources
Primary energy sources should not be confused with the energy system components (or conversion processes) through which they are converted into energy carriers.
Usable energy
Primary energy sources are transformed in energy conversion processes to more convenient forms of energy that can directly be used by society, such as electrical energy, refined fuels, or synthetic fuels such as hydrogen fuel. In the field of energetics, these forms are called energy carriers and correspond to the concept of "secondary energy" in energy statistics.
Conversion to energy carriers (or secondary energy)
Energy carriers are energy forms which have been tra
Document 3:::
The indirect land use change impacts of biofuels, also known as ILUC or iLUC (pronounced as i-luck), relates to the unintended consequence of releasing more carbon emissions due to land-use changes around the world induced by the expansion of croplands for ethanol or biodiesel production in response to the increased global demand for biofuels.
As farmers worldwide respond to higher crop prices in order to maintain the global food supply-and-demand balance, pristine lands are cleared to replace the food crops that were diverted elsewhere to biofuels' production. Because natural lands, such as rainforests and grasslands, store carbon in their soil and biomass as plants grow each year, clearance of wilderness for new farms translates to a net increase in greenhouse gas emissions. Due to this off-site change in the carbon stock of the soil and the biomass, indirect land use change has consequences in the greenhouse gas (GHG) balance of a biofuel.
Other authors have also argued that indirect land use changes produce other significant social and environmental impacts, affecting biodiversity, water quality, food prices and supply, land tenure, worker migration, and community and cultural stability.
History
The estimates of carbon intensity for a given biofuel depend on the assumptions regarding several variables. As of 2008, multiple full life cycle studies had found that corn ethanol, cellulosic ethanol and Brazilian sugarcane ethanol produce lower greenhouse gas emissions than gasoline. None of these studies, however, considered the effects of indirect land-use changes, and though land use impacts were acknowledged, estimation was considered too complex and difficult to model. A controversial paper published in February 2008 in Sciencexpress by a team led by Searchinger from Princeton University concluded that such effects offset the (positive) direct effects of both corn and cellulosic ethanol and that Brazilian sugarcane performed better, but still resulted in a sma
Document 4:::
Energy statistics refers to collecting, compiling, analyzing and disseminating data on commodities such as coal, crude oil, natural gas, electricity, or renewable energy sources (biomass, geothermal, wind or solar energy), when they are used for the energy they contain. Energy is the capability of some substances, resulting from their physico-chemical properties, to do work or produce heat. Some energy commodities, called fuels, release their energy content as heat when they burn. This heat could be used to run an internal or external combustion engine.
The need to have statistics on energy commodities became obvious during the 1973 oil crisis that brought tenfold increase in petroleum prices. Before the crisis, to have accurate data on global energy supply and demand was not deemed critical. Another concern of energy statistics today is a huge gap in energy use between developed and developing countries. As the gap narrows (see picture), the pressure on energy supply increases tremendously.
The data on energy and electricity come from three principal sources:
Energy industry
Other industries ("self-producers")
Consumers
The flows of and trade in energy commodities are measured both in physical units (e.g., metric tons), and, when energy balances are calculated, in energy units (e.g., terajoules or tons of oil equivalent). What makes energy statistics specific and different from other fields of economic statistics is the fact that energy commodities undergo greater number of transformations (flows) than other commodities. In these transformations energy is conserved, as defined by and within the limitations of the first and second laws of thermodynamics.
See also
Energy system
World energy resources and consumption
External links
Statistical Energy Database Review: Enerdata Yearbook 2012
International Energy Agency: Statistics
United Nations: Energy Statistics
The Oslo Group on Energy Statistics
DOE Energy Information Administration
Year of Ener
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What are non-renewable sources of energy that produce environmental damage?
A. fossil fuels
B. crop fuels
C. wind
D. sunlight
Answer:
|
|
sciq-11661
|
multiple_choice
|
What occurs when nature reclaims areas formerly occupied by life?
|
[
"tertiary succession",
"secondary succession",
"typical succession",
"primary succession"
] |
B
|
Relavent Documents:
Document 0:::
Primary succession is the beginning step of ecological succession after an extreme disturbance, which usually occurs in an environment devoid of vegetation and other organisms. These environments are typically lacking in soil, as disturbances like lava flow or retreating glaciers scour the environment clear of nutrients.
In contrast, secondary succession occurs on substrates that previously supported vegetation before an ecological disturbance. This occurs when smaller disturbances like floods, hurricanes, tornadoes, and fires destroy only the local plant life and leave soil nutrients for immediate establishment by intermediate community species.
Occurrence
In primary succession pioneer species like lichen, algae and fungi as well as abiotic factors like wind and water start to "normalise" the habitat or in other words start to develop soil and other important mechanisms for greater diversity to flourish. Primary succession begins on rock formations, such as volcanoes or mountains, or in a place with no organisms or soil. Primary succession leads to conditions nearer optimum for vascular plant growth; pedogenesis or the formation of soil, and the increased amount of shade are the most important processes.
These pioneer lichen, algae, and fungi are then dominated and often replaced by plants that are better adapted to less harsh conditions, these plants include vascular plants like grasses and some shrubs that are able to live in thin soils that are often mineral-based. Water and nutrient levels increase with the amount of succession exhibited.
The early stages of primary succession are dominated by species with small propagules (seed and spores) which can be dispersed long distances. The early colonizers—often algae, fungi, and lichens—stabilize the substrate. Nitrogen supplies are limited in new soils, and nitrogen-fixing species tend to play an important role early in primary succession. Unlike in primary succession, the species that dominate secondary success
Document 1:::
Plant ecology is a subdiscipline of ecology that studies the distribution and abundance of plants, the effects of environmental factors upon the abundance of plants, and the interactions among plants and between plants and other organisms. Examples of these are the distribution of temperate deciduous forests in North America, the effects of drought or flooding upon plant survival, and competition among desert plants for water, or effects of herds of grazing animals upon the composition of grasslands.
A global overview of the Earth's major vegetation types is provided by O.W. Archibold. He recognizes 11 major vegetation types: tropical forests, tropical savannas, arid regions (deserts), Mediterranean ecosystems, temperate forest ecosystems, temperate grasslands, coniferous forests, tundra (both polar and high mountain), terrestrial wetlands, freshwater ecosystems and coastal/marine systems. This breadth of topics shows the complexity of plant ecology, since it includes plants from floating single-celled algae up to large canopy forming trees.
One feature that defines plants is photosynthesis. Photosynthesis is the process of a chemical reactions to create glucose and oxygen, which is vital for plant life. One of the most important aspects of plant ecology is the role plants have played in creating the oxygenated atmosphere of earth, an event that occurred some 2 billion years ago. It can be dated by the deposition of banded iron formations, distinctive sedimentary rocks with large amounts of iron oxide. At the same time, plants began removing carbon dioxide from the atmosphere, thereby initiating the process of controlling Earth's climate. A long term trend of the Earth has been toward increasing oxygen and decreasing carbon dioxide, and many other events in the Earth's history, like the first movement of life onto land, are likely tied to this sequence of events.
One of the early classic books on plant ecology was written by J.E. Weaver and F.E. Clements. It
Document 2:::
Secondary succession is the secondary ecological succession of a plant's life. As opposed to the first, primary succession, secondary succession is a process started by an event (e.g. forest fire, harvesting, hurricane, etc.) that reduces an already established ecosystem (e.g. a forest or a wheat field) to a smaller population of species, and as such secondary succession occurs on preexisting soil whereas primary succession usually occurs in a place lacking soil. Many factors can affect secondary succession, such as trophic interaction, initial composition, and competition-colonization trade-offs. The factors that control the increase in abundance of a species during succession may be determined mainly by seed production and dispersal, micro climate; landscape structure (habitat patch size and distance to outside seed sources); bulk density, pH, and soil texture (sand and clay).
Secondary succession is the ecological succession that occurs after the initial succession has been disrupted and some plants and animals still exist. It is usually faster than primary succession as soil is already present, and seeds, roots, and the underground vegetative organs of plants may still survive in the soil.
Examples
Imperata
Imperata grasslands are caused by human activities such as logging, forest clearing for shifting cultivation, agriculture and grazing, and also by frequent fires. The latter is a frequent result of human interference. However, when not maintained by frequent fires and human disturbances, they regenerate naturally and speedily to secondary young forest. The time of succession in Imperata grassland (for example in Samboja Lestari area), Imperata cylindrica has the highest coverage but it becomes less dominant from the fourth year onwards. While Imperata decreases, the percentage of shrubs and young trees clearly increases with time. In the burned plots, Melastoma malabathricum, Eupatorium inulaefolium, Ficus sp., and Vitex pinnata. strongly increase with
Document 3:::
"Auto-" meaning self or same, and "-genic" meaning producing or causing. Autogenic succession refers to ecological succession driven by biotic factors within an ecosystem and although the mechanisms of autogenic succession have long been debated, the role of living things in shaping the progression of succession was realized early on. Presently, there is more of a consensus that the mechanisms of facilitation, tolerance, and inhibition all contribute to autogenic succession. The concept of succession is most often associated with communities of vegetation and forests, though it is applicable to a broader range of ecosystems. In contrast, allogenic succession is driven by the abiotic components of the ecosystem.
How it occurs
The plants themselves (biotic components) cause succession to occur.
Light captured by leaves
Production of detritus
Water and nutrient uptake
Nitrogen fixation
anthropogenic climate change
These aspects lead to a gradual ecological change in a particular spot of land, known as a progression of inhabiting species. Autogenic succession can be viewed as a secondary succession because of pre-existing plant life. A 2000 case study in the journal Oecologia tested the hypothesis that areas with high plant diversity could suppress weed growth more effectively than those with lower plant diversity.
Facilitation
Improvement of site factors like increased organic matter
Inhibition
Hinders species or growth
Document 4:::
Ecological classification or ecological typology is the classification of land or water into geographical units that represent variation in one or more ecological features. Traditional approaches focus on geology, topography, biogeography, soils, vegetation, climate conditions, living species, habitats, water resources, and sometimes also anthropic factors. Most approaches pursue the cartographical delineation or regionalisation of distinct areas for mapping and planning.
Approaches to classifications
Different approaches to ecological classifications have been developed in terrestrial, freshwater and marine disciplines. Traditionally these approaches have focused on biotic components (vegetation classification), abiotic components (environmental approaches) or implied ecological and evolutionary processes (biogeographical approaches). Ecosystem classifications are specific kinds of ecological classifications that consider all four elements of the definition of ecosystems: a biotic component, an abiotic complex, the interactions between and within them, and the physical space they occupy (ecotope).
Vegetation classification
Vegetation is often used to classify terrestrial ecological units. Vegetation classification can be based on vegetation structure and floristic composition. Classifications based entirely on vegetation structure overlap with land cover mapping categories.
Many schemes of vegetation classification are in use by the land, resource and environmental management agencies of different national and state jurisdictions. The International Vegetation Classification (IVC or EcoVeg) has been recently proposed but has not been yet widely adopted.
Vegetation classifications have limited use in aquatic systems, since only a handful of freshwater or marine habitats are dominated by plants (e.g. kelp forests or seagrass meadows). Also, some extreme terrestrial environments, like subterranean or cryogenic ecosystems, are not properly described in vegetation c
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What occurs when nature reclaims areas formerly occupied by life?
A. tertiary succession
B. secondary succession
C. typical succession
D. primary succession
Answer:
|
|
sciq-10919
|
multiple_choice
|
What natural destructive element cuts away at rock forming river valleys?
|
[
"water",
"lava",
"sediment",
"air"
] |
A
|
Relavent Documents:
Document 0:::
The Arkansas River Valley (usually shortened to River Valley) is a region in Arkansas defined by the Arkansas River in the western part of the state. Generally defined as the area between the Ozark and Ouachita Mountains, the River Valley is characterized by flat lowlands covered in fertile farmland and lakes periodically interrupted by high peaks. Mount Magazine, Mount Nebo, and Petit Jean Mountain compose the Tri-Peaks Region, a further subdivision of the River Valley popular with hikers and outdoors enthusiasts. In addition to the outdoor recreational activities available to residents and visitors of the region, the River Valley contains Arkansas's wine country as well as hundreds of historical sites throughout the area. It is one of six natural divisions of Arkansas.
Definition
The Arkansas River Valley is informally defined along county boundaries, including all of Logan and Sebastian counties and portions of Conway, Franklin, Johnson, Perry, Pope, and Yell counties.
Subdivisions
Arkansas Valley Hills - North and east of the Arkansas River, sometimes associated with the Ozarks
Bottomlands - Low swamps and prairies along the Arkansas River itself, wide in some places
Fort Smith metropolitan area - Sebastian, Crawford, and Franklin counties in Arkansas (also includes Le Fore and Sequoyah counties in Oklahoma)
Ozark National Forest - a small, discontinuous portion of the federally protected area is within the region
Tri-peaks Region - Region punctuated by three steep mountains: Mount Magazine, Mount Nebo and Petit Jean Mountain
Valley - south of the Arkansas River, level plains and gently rolling hills
Wine Country - American Viticultural Area near Altus
History
In the Pre-Colonial era, the River Valley was inhabited by Native American tribes, including Caddo, Cherokee, Choctaw, Osage, Tunica, and Quapaw tribes. Most first encounters describe scattered villages and individual farmsteads in the River Valley, unlike the organized "towns" and groves and o
Document 1:::
Large woody debris (LWD) are the logs, sticks, branches, and other wood that falls into streams and rivers. This debris can influence the flow and the shape of the stream channel. Large woody debris, grains, and the shape of the bed of the stream are the three main providers of flow resistance, and are thus, a major influence on the shape of the stream channel. Some stream channels have less LWD than they would naturally because of removal by watershed managers for flood control and aesthetic reasons.
The study of woody debris is important for its forestry management implications. Plantation thinning can reduce the potential for recruitment of LWD into proximal streams. The presence of large woody debris is important in the formation of pools which serve as salmon habitat in the Pacific Northwest. Entrainment of the large woody debris in a stream can also cause erosion and scouring around and under the LWD. The amount of scouring and erosion is determined by the ratio of the diameter of the piece, to the depth of the stream, and the embedding and orientation of the piece.
Influence on stream flow around bends
Large woody debris slow the flow through a bend in the stream, while accelerating flow in the constricted area downstream of the obstruction.
See also
Beaver dam
Coarse woody debris
Driftwood
Log jam
Stream restoration
Document 2:::
A grassed waterway is a to 48-metre-wide (157 ft) native grassland strip of green belt. It is generally installed in the thalweg, the deepest continuous line along a valley or watercourse, of a cultivated dry valley in order to control erosion. A study carried out on a grassed waterway during 8 years in Bavaria showed that it can lead to several other types of positive impacts, e.g. on biodiversity.
Distinctions
Confusion between "grassed waterway" and "vegetative filter strips" should be avoided. The latter are generally narrower (only a few metres wide) and rather installed along rivers as well as along or within cultivated fields. However, buffer strip can be a synonym, with shrubs and trees added to the plant component, as does a riparian zone.
Runoff and erosion mitigation
Runoff generated on cropland during storms or long winter rains concentrates in the thalweg where it can lead to rill or gully erosion.
Rills and gullies further concentrate runoff and speed up its transfer, which can worsen damage occurring downstream. This can result in a muddy flood.
In this context, a grassed waterway allows increasing soil cohesion and roughness. It also prevents the formation of rills and gullies. Furthermore, it can slow down runoff and allow its re-infiltration during long winter rains. In contrast, its infiltration capacity is generally not sufficient to reinfiltrate runoff produced by heavy spring and summer storms. It can therefore be useful to combine it with extra measures, like the installation of earthen dams across the grassed waterway, in order to buffer runoff temporarily.
Document 3:::
An endangered river is one which has the potential to partly or fully dry up, or one that is thought to have ecological issues now or in the near future. Some such issues are natural while others are the direct result of human development. Organisations including the World Wildlife Fund (WWF) have published lists of rivers at risk. The WWF's 2007 list featured the Danube, the Nile, and the Rio Grande amongst others, stating that these "once great rivers" are in danger and "can no longer be assured of reaching the sea unhindered". The US National Park Service reported in 2015 that part of the Rio Grande "often lies dry". Dangers may be the result of the natural change of conditions in the local environment but many are due to human development projects such as dams and irrigation. Humans have also caused significant water pollution which endangers life which relies upon the water source.
Importance
According to the World Resources Institute (WRI) river basins are "indispensable resources for billions of people, companies, farms, and ecosystems" but many suffer from "water stress". It states that numerous major rivers around the world have become depleted to the point that they can fail to reach their destinations. Water scarcity is a danger to all the life that depends on the source.
Dangers to rivers
Rivers face many dangers, both natural and manmade. Climate change can have a significant impact on rivers, both positive and negative. This could be seen itself as an indirect result of human developments. An area which has a reduction in rainfall or an increase in temperature can make a river fully or partially dry out. This often means that the water can no longer reach its mouth.
The flora and fauna of an area can face change and can have a substantial impact on landscape and local ecology. Invasive species can move into an area or be introduced by humans. One particular example of the latter is the case of beavers being introduced into Tierra del Fuego in 1946.
Document 4:::
Hydraulic roughness is the measure of the amount of frictional resistance water experiences when passing over land and channel features.
One roughness coefficient is Manning's n-value. Manning's n is used extensively around the world to predict the degree of roughness in channels. Flow velocity is strongly dependent on the resistance to flow. An increase in this n value will cause a decrease in the velocity of water flowing across a surface.
Manning's n
The value of Manning's n is affected by many variables. Factors like suspended load, sediment grain size, presence of bedrock or boulders in the stream channel, variations in channel width and depth, and overall sinuosity of the stream channel can all affect Manning's n value. Biological factors have the greatest overall effect on Manning's n; bank stabilization by vegetation, height of grass and brush across a floodplain, and stumps and logs creating natural dams are the main observable influences.
Biological Importance
Recent studies have found a relationship between hydraulic roughness and salmon spawning habitat; “bed-surface grain size is responsive to hydraulic roughness caused by bank irregularities, bars, and wood debris… We find that wood debris plays an important role at our study sites, not only providing hydraulic roughness but also influencing pool spacing, frequency of textural patches, and the amplitude and wavelength of bank and bar topography and their consequent roughness. Channels with progressively greater hydraulic roughness have systematically finer bed surfaces, presumably due to reduced bed shear stress, resulting in lower channel competence and diminished bed load transport capacity, both of which promote textural fining”. Textural fining of stream beds can effect more than just salmon spawning habitats, “bar and wood roughness create a greater variety of textural patches, offering a range of aquatic habitats that may promote biologic diversity or be of use to specific animals at differe
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What natural destructive element cuts away at rock forming river valleys?
A. water
B. lava
C. sediment
D. air
Answer:
|
|
sciq-2233
|
multiple_choice
|
What type of behavior is most flexible?
|
[
"innate behavior",
"adaptation",
"learned behavior",
"reflex behavior"
] |
C
|
Relavent Documents:
Document 0:::
"Fixed action pattern" is an ethological term describing an instinctive behavioral sequence that is highly stereotyped and species-characteristic. Fixed action patterns are said to be produced by the innate releasing mechanism, a "hard-wired" neural network, in response to a sign/key stimulus or releaser. Once released, a fixed action pattern runs to completion.
This term is often associated with Konrad Lorenz, who is the founder of the concept. Lorenz identified six characteristics of fixed action patterns. These characteristics state that fixed action patterns are stereotyped, complex, species-characteristic, released, triggered, and independent of experience.
Fixed action patterns have been observed in many species, but most notably in fish and birds. Classic studies by Konrad Lorenz and Niko Tinbergen involve male stickleback mating behavior and greylag goose egg-retrieval behavior.
Fixed action patterns have been shown to be evolutionarily advantageous, as they increase both fitness and speed. However, as a result of their predictability, they may also be used as a means of exploitation. An example of this exploitation would be brood parasitism.
There are four exceptions to fixed action pattern rules: reduced response threshold, vacuum activity, displacement behavior, and graded response.
Characteristics
There are 6 characteristics of fixed action patterns. Fixed action patterns are said to be stereotyped, complex, species-characteristic, released, triggered, and independent of experience.
Stereotyped: Fixed action patterns occur in rigid, predictable, and highly-structured sequences.
Complex: Fixed action patterns are not a simple reflex. They are a complex pattern of behavior.
Species-characteristic: Fixed action patterns occur in all members of a species of a certain sex and/or a given age when they have attained a specific level of arousal.
Released: Fixed action patterns occur in response to a certain sign stimulus or releaser.
Triggered: Once relea
Document 1:::
Behavior (American English) or behaviour (British English) is the range of actions and mannerisms made by individuals, organisms, systems or artificial entities in some environment. These systems can include other systems or organisms as well as the inanimate physical environment. It is the computed response of the system or organism to various stimuli or inputs, whether internal or external, conscious or subconscious, overt or covert, and voluntary or involuntary.
Taking a behavior informatics perspective, a behavior consists of actor, operation, interactions, and their properties. This can be represented as a behavior vector.
Models
Biology
Although disagreement exists as to how to precisely define behavior in a biological context, one common interpretation based on a meta-analysis of scientific literature states that "behavior is the internally coordinated responses (actions or inactions) of whole living organisms (individuals or groups) to internal or external stimuli".
A broader definition of behavior, applicable to plants and other organisms, is similar to the concept of phenotypic plasticity. It describes behavior as a response to an event or environment change during the course of the lifetime of an individual, differing from other physiological or biochemical changes that occur more rapidly, and excluding changes that are a result of development (ontogeny).
Behaviors can be either innate or learned from the environment.
Behavior can be regarded as any action of an organism that changes its relationship to its environment. Behavior provides outputs from the organism to the environment.
Human behavior
The endocrine system and the nervous system likely influence human behavior. Complexity in the behavior of an organism may be correlated to the complexity of its nervous system. Generally, organisms with more complex nervous systems have a greater capacity to learn new responses and thus adjust their behavior.
Animal behavior
Ethology is the scientifi
Document 2:::
Behavioral plasticity refers to a change in an organism's behavior that results from exposure to stimuli, such as changing environmental conditions. Behavior can change more rapidly in response to changes in internal or external stimuli than is the case for most morphological traits and many physiological traits. As a result, when organisms are confronted by new conditions, behavioral changes often occur in advance of physiological or morphological changes. For instance, larval amphibians changed their antipredator behavior within an hour after a change in cues from predators, but morphological changes in body and tail shape in response to the same cues required a week to complete.
Background
For many years, ethologists have studied the ways that behavior can change in response to changes in external stimuli or changes in the internal state of an organism. In a parallel literature, psychologists studying learning and cognition have spent years documenting the many ways that experiences in the past can affect the behavior an individual expresses at the current time. Interest in behavioral plasticity gained prominence more recently as an example of a type of phenotypic plasticity with major consequences for evolutionary biology.
Types
Behavioral plasticity can be broadly organized into two types: exogenous and endogenous. Exogenous plasticity refers to the changes in behavioral phenotype (i.e., observable behaviors) caused by an external stimulus, experience, or environment. Endogenous plasticity encompasses plastic responses that result from changes in internal cues, such as genotype, circadian rhythms, and menstruation.
These two broad categories can be further broken down into two other important classifications. When an external stimulus elicits or "activates" an immediate response (an immediate effect on behavior), then the organism is demonstrating contextual plasticity. This form of plasticity highlights the concept that external stimuli in a given context
Document 3:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 4:::
Instinct is the inherent inclination of a living organism towards a particular complex behaviour, containing innate (inborn) elements. The simplest example of an instinctive behaviour is a fixed action pattern (FAP), in which a very short to medium length sequence of actions, without variation, are carried out in response to a corresponding clearly defined stimulus.
Any behaviour is instinctive if it is performed without being based upon prior experience (that is, in the absence of learning), and is therefore an expression of innate biological factors. Sea turtles, newly hatched on a beach, will instinctively move toward the ocean. A marsupial climbs into its mother's pouch upon being born. Other examples include animal fighting, animal courtship behaviour, internal escape functions, and the building of nests. Though an instinct is defined by its invariant innate characteristics, details of its performance can be changed by experience; for example, a dog can improve its listening skills by practice.
Instincts are inborn complex patterns of behaviour that exist in most members of the species, and should be distinguished from reflexes, which are simple responses of an organism to a specific stimulus, such as the contraction of the pupil in response to bright light or the spasmodic movement of the lower leg when the knee is tapped. The absence of volitional capacity must not be confused with an inability to modify fixed action patterns. For example, people may be able to modify a stimulated fixed action pattern by consciously recognizing the point of its activation and simply stop doing it, whereas animals without a sufficiently strong volitional capacity may not be able to disengage from their fixed action patterns, once activated.
Instinctual behaviour in humans has been studied.
Early theorists
Jean Henri Fabre
Jean Henri Fabre (1823–1915) is said to be the first person to study small animals (that weren't birds) and insects, and he specifically specialized i
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What type of behavior is most flexible?
A. innate behavior
B. adaptation
C. learned behavior
D. reflex behavior
Answer:
|
|
scienceQA-1790
|
multiple_choice
|
Select the invertebrate.
|
[
"human",
"puffin",
"golden orb-weaver",
"cardinalfish"
] |
C
|
Like other spiders, a golden orb-weaver is an invertebrate. It does not have a backbone. It has an exoskeleton.
A human is a mammal. Like other mammals, a human is a vertebrate. It has a backbone.
A puffin is a bird. Like other birds, a puffin is a vertebrate. It has a backbone.
A cardinalfish is a fish. Like other fish, a cardinalfish is a vertebrate. It has a backbone.
|
Relavent Documents:
Document 0:::
Animals are multicellular, eukaryotic organisms in the biological kingdom Animalia. With few exceptions, animals consume organic material, breathe oxygen, have myocytes and are able to move, can reproduce sexually, and grow from a hollow sphere of cells, the blastula, during embryonic development. As of 2022, 2.16 million living animal species have been described—of which around 1.05 million are insects, over 85,000 are molluscs, and around 65,000 are vertebrates. It has been estimated there are around 7.77 million animal species. Animals range in length from to . They have complex interactions with each other and their environments, forming intricate food webs. The scientific study of animals is known as zoology.
Most living animal species are in Bilateria, a clade whose members have a bilaterally symmetric body plan. The Bilateria include the protostomes, containing animals such as nematodes, arthropods, flatworms, annelids and molluscs, and the deuterostomes, containing the echinoderms and the chordates, the latter including the vertebrates. Life forms interpreted as early animals were present in the Ediacaran biota of the late Precambrian. Many modern animal phyla became clearly established in the fossil record as marine species during the Cambrian explosion, which began around 539 million years ago. 6,331 groups of genes common to all living animals have been identified; these may have arisen from a single common ancestor that lived 650 million years ago.
Historically, Aristotle divided animals into those with blood and those without. Carl Linnaeus created the first hierarchical biological classification for animals in 1758 with his Systema Naturae, which Jean-Baptiste Lamarck expanded into 14 phyla by 1809. In 1874, Ernst Haeckel divided the animal kingdom into the multicellular Metazoa (now synonymous with Animalia) and the Protozoa, single-celled organisms no longer considered animals. In modern times, the biological classification of animals relies on ad
Document 1:::
Invertebrate zoology is the subdiscipline of zoology that consists of the study of invertebrates, animals without a backbone (a structure which is found only in fish, amphibians, reptiles, birds and mammals).
Invertebrates are a vast and very diverse group of animals that includes sponges, echinoderms, tunicates, numerous different phyla of worms, molluscs, arthropods and many additional phyla. Single-celled organisms or protists are usually not included within the same group as invertebrates.
Subdivisions
Invertebrates represent 97% of all named animal species, and because of that fact, this subdivision of zoology has many further
subdivisions, including but not limited to:
Arthropodology - the study of arthropods, which includes
Arachnology - the study of spiders and other arachnids
Entomology - the study of insects
Carcinology - the study of crustaceans
Myriapodology - the study of centipedes, millipedes, and other myriapods
Cnidariology - the study of Cnidaria
Helminthology - the study of parasitic worms.
Malacology - the study of mollusks, which includes
Conchology - the study of Mollusk shells.
Limacology - the study of slugs.
Teuthology - the study of cephalopods.
Invertebrate paleontology - the study of fossil invertebrates
These divisions are sometimes further divided into more specific specialties. For example, within arachnology, acarology is the study of mites and ticks; within entomology, lepidoptery is the study of butterflies and moths, myrmecology is the study of ants and so on. Marine invertebrates are all those invertebrates that exist in marine habitats.
History
Early Modern Era
In the early modern period starting in the late 16th century, invertebrate zoology saw growth in the number of publications made and improvement in the experimental practices associated with the field. (Insects are one of the most diverse groups of organisms on Earth. They play important roles in ecosystems, including pollination, natural enemies, saprophytes, and
Document 2:::
International Society for Invertebrate Morphology (ISIM) was founded during the 1st International Congress on Invertebrate Morphology, in Copenhagen, August 2008. The objectives of the society are to promote international collaboration and provide educational opportunities and training on invertebrate morphology, and to organize and promote the international congresses of invertebrate morphology, international meetings and other forms of scientific exchange.
The ISIM has its own Constitution
ISIM board 2014-2017
Gerhard Scholtz (President) Institute of Biology, Humboldt-Universität zu Berlin, Germany. https://www.biologie.hu-berlin.de/de/gruppenseiten/compzool/people/gerhard_scholtz_page
Natalia Biserova (President-Elect) Moscow State University, Moscow, Russia.
Gonzalo Giribet (Past-President) Museum of Comparative Zoology, Harvard University, Cambridge, MA, USA.
Julia Sigwart (Secretary)
Katrina Worsaae (Treasurer)
Greg Edgecombe (2nd term)
Andreas Hejnol (2nd term)
Sally Leys (2nd term)
Fernando Pardos (2nd term)
Katharina Jörger (1st term)
Marymegan Daly (1st term)
Georg Mayer (1st term)
ISIM board 2017-2020
Natalia Biserova (President), Lomonosov Moscow State University, Moscow, Russian Federation http://invert.bio.msu.ru/en/staff-en/33-biserova-en .
Andreas Wanninger (President-elect), Department of Integrative Zoology, University of Vienna, Vienna, Austria.
Gerhard Scholtz (Past-president), Department of Biology, Humboldt-Universität zu Berlin, Germany.
Julia Sigwart (Secretary), School of Biological Sciences, Queen's University Belfast, UK.
Katrine Worsaae (Treasurer), Department of Biology, University of Copenhagen, Copenhagen, Denmark.
Advisory Council:
Ariel Chipman (Israel)
D. Bruce Conn (USA)
Conrad Helm (Germany)
Xiaoya Ma (UK)
Pedro Martinez (Spain)
Ana Riesgo (Spain)
Nadezhda Rimskaya-Korsakova (Russia)
Elected 23-08-2017, Moscow
Former meetings
ICIM 1 (2008) University of Copenhagen, Denmark
ICIM 2 (2011) H
Document 3:::
A cnidariologist is a zoologist specializing in Cnidaria, a group of freshwater and marine aquatic animals that include the sea anemones, corals, and jellyfish.
Examples
Edward Thomas Browne (1866-1937)
Henry Bryant Bigelow (1879-1967)
Randolph Kirkpatrick (1863–1950)
Kamakichi Kishinouye (1867-1929)
Paul Lassenius Kramp (1887-1975)
Alfred G. Mayer (1868-1922)
See also
Document 4:::
This is a list of scientific journals which cover the field of zoology.
A
Acta Entomologica Musei Nationalis Pragae
Acta Zoologica Academiae Scientiarum Hungaricae
Acta Zoologica Bulgarica
Acta Zoológica Mexicana
Acta Zoologica: Morphology and Evolution
African Entomology
African Invertebrates
African Journal of Herpetology
African Zoology
Alces
American Journal of Primatology
Animal Biology, formerly Netherlands Journal of Zoology
Animal Cognition
Arctic
Australian Journal of Zoology
Australian Mammalogy
B
Bulgarian Journal of Agricultural Science
Bulletin of the American Museum of Natural History
C
Canadian Journal of Zoology
Caribbean Herpetology
Central European Journal of Biology
Contributions to Zoology
Copeia
Crustaceana
E
Environmental Biology of Fishes
F
Frontiers in Zoology
H
Herpetological Monographs
I
Integrative and Comparative Biology, formerly American Zoologist
International Journal of Acarology
International Journal of Primatology
J
M
Malacologia
N
North-Western Journal of Zoology
P
Physiological and Biochemical Zoology
R
Raffles Bulletin of Zoology
Rangifer
Russian Journal of Nematology
V
The Veliger
W
Worm Runner's Digest
Z
See also
List of biology journals
List of ornithology journals
List of entomology journals
Lists of academic journals
Zoology-related lists
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Select the invertebrate.
A. human
B. puffin
C. golden orb-weaver
D. cardinalfish
Answer:
|
sciq-5637
|
multiple_choice
|
What does the distribution of darts on a dartboard shows the difference between?
|
[
"Aim and precision",
"equilibrium and precision",
"accuracy and precision",
"production and precision"
] |
C
|
Relavent Documents:
Document 0:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 1:::
Adaptive comparative judgement is a technique borrowed from psychophysics which is able to generate reliable results for educational assessment – as such it is an alternative to traditional exam script marking. In the approach, judges are presented with pairs of student work and are then asked to choose which is better, one or the other. By means of an iterative and adaptive algorithm, a scaled distribution of student work can then be obtained without reference to criteria.
Introduction
Traditional exam script marking began in Cambridge 1792 when, with undergraduate numbers rising, the importance of proper ranking of students was growing. So in 1792 the new Proctor of Examinations, William Farish, introduced marking, a process in which every examiner gives a numerical score to each response by every student, and the overall total mark puts the students in the final rank order. Francis Galton (1869) noted that, in an unidentified year about 1863, the Senior Wrangler scored 7,634 out of a maximum of 17,000, while the Second Wrangler scored 4,123. (The 'Wooden Spoon' scored only 237.)
Prior to 1792, a team of Cambridge examiners convened at 5pm on the last day of examining, reviewed the 19 papers each student had sat – and published their rank order at midnight. Marking solved the problems of numbers and prevented unfair personal bias, and its introduction was a step towards modern objective testing, the format it is best suited to. But the technology of testing that followed, with its major emphasis on reliability and the automatisation of marking, has been an uncomfortable partner for some areas of educational achievement: assessing writing or speaking, and other kinds of performance need something more qualitative and judgemental.
The technique of Adaptive Comparative Judgement is an alternative to marking. It returns to the pre-1792 idea of sorting papers according to their quality, but retains the guarantee of reliability and fairness. It is by far the most rel
Document 2:::
In mathematical psychology and education theory, a knowledge space is a combinatorial structure used to formulate mathematical models describing the progression of a human learner. Knowledge spaces were introduced in 1985 by Jean-Paul Doignon and Jean-Claude Falmagne, and remain in extensive use in the education theory. Modern applications include two computerized tutoring systems, ALEKS and the defunct RATH.
Formally, a knowledge space assumes that a domain of knowledge is a collection of concepts or skills, each of which must be eventually mastered. Not all concepts are interchangeable; some require other concepts as prerequisites. Conversely, competency at one skill may ease the acquisition of another through similarity. A knowledge space marks out which collections of skills are feasible: they can be learned without mastering any other skills. Under reasonable assumptions, the collection of feasible competencies forms the mathematical structure known as an antimatroid.
Researchers and educators usually explore the structure of a discipline's knowledge space as a latent class model.
Motivation
Knowledge Space Theory attempts to address shortcomings of standardized testing when used in educational psychometry. Common tests, such as the SAT and ACT, compress a student's knowledge into a very small range of ordinal ranks, in the process effacing the conceptual dependencies between questions. Consequently, the tests cannot distinguish between true understanding and guesses, nor can they identify a student's particular weaknesses, only the general proportion of skills mastered. The goal of knowledge space theory is to provide a language by which exams can communicate
What the student can do and
What the student is ready to learn.
Model structure
Knowledge Space Theory-based models presume that an educational subject can be modeled as a finite set of concepts, skills, or topics. Each feasible state of knowledge about is then a subset of ; the set of
Document 3:::
Test equating traditionally refers to the statistical process of determining comparable scores on different forms of an exam. It can be accomplished using either classical test theory or item response theory.
In item response theory, equating is the process of placing scores from two or more parallel test forms onto a common score scale. The result is that scores from two different test forms can be compared directly, or treated as though they came from the same test form. When the tests are not parallel, the general process is called linking. It is the process of equating the units and origins of two scales on which the abilities of students have been estimated from results on different tests. The process is analogous to equating degrees Fahrenheit with degrees Celsius by converting measurements from one scale to the other. The determination of comparable scores is a by-product of equating that results from equating the scales obtained from test results.
Purpose
Suppose that Dick and Jane both take a test to become licensed in a certain profession. Because the high stakes (you get to practice the profession if you pass the test) may create a temptation to cheat, the organization that oversees the test creates two forms. If we know that Dick scored 60% on form A and Jane scored 70% on form B, do we know for sure which one has a better grasp of the material? What if form A is composed of very difficult items, while form B is relatively easy? Equating analyses are performed to address this very issue, so that scores are as fair as possible.
Equating in item response theory
In item response theory, person "locations" (measures of some quality being assessed by a test) are estimated on an interval scale; i.e., locations are estimated in relation to a unit and origin. It is common in educational assessment to employ tests in order to assess different groups of students with the intention of establishing a common scale by equating the origins, and when appropri
Document 4:::
Curriculum-based measurement, or CBM, is also referred to as a general outcomes measures (GOMs) of a student's performance in either basic skills or content knowledge.
Early history
CBM began in the mid-1970s with research headed by Stan Deno at the University of Minnesota. Over the course of 10 years, this work led to the establishment of measurement systems in reading, writing, and spelling that were: (a) easy to construct, (b) brief in administration and scoring, (c) had technical adequacy (reliability and various types of validity evidence for use in making educational decisions), and (d) provided alternate forms to allow time series data to be collected on student progress. This focus in the three language arts areas eventually was expanded to include mathematics, though the technical research in this area continues to lag that published in the language arts areas. An even later development was the application of CBM to middle-secondary areas: Espin and colleagues at the University of Minnesota developed a line of research addressing vocabulary and comprehension (with the maze) and by Tindal and colleagues at the University of Oregon developed a line of research on concept-based teaching and learning.
Increasing importance
Early research on the CBM quickly moved from monitoring student progress to its use in screening, normative decision-making, and finally benchmarking. Indeed, with the implementation of the No Child Left Behind Act in 2001, and its focus on large-scale testing and accountability, CBM has become increasingly important as a form of standardized measurement that is highly related to and relevant for understanding student's progress toward and achievement of state standards.
Key feature
Probably the key feature of CBM is its accessibility for classroom application and implementation. It was designed to provide an experimental analysis of the effects from interventions, which includes both instruction and curriculum. This is one of the most imp
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What does the distribution of darts on a dartboard shows the difference between?
A. Aim and precision
B. equilibrium and precision
C. accuracy and precision
D. production and precision
Answer:
|
|
sciq-8114
|
multiple_choice
|
What is the name of the life stage when a child becomes sexually mature?
|
[
"maturity",
"puberty",
"adulthood",
"adolescence"
] |
B
|
Relavent Documents:
Document 0:::
An adult is a human or other animal that has reached full growth. The biological definition of the word means an animal reaching sexual maturity and thus capable of reproduction. In the human context, the term adult has meanings associated with social and legal concepts. In contrast to a non-adult or "minor", a legal adult is a person who has attained the age of majority and is therefore regarded as independent, self-sufficient, and responsible. They may also be regarded as a "major". The typical age of attaining legal adulthood is 18 to 21, although definition may vary by legal rights, country, and psychological development.
Human adulthood encompasses psychological adult development. Definitions of adulthood are often inconsistent and contradictory; a person may be biologically an adult, and have adult behavior, but still be treated as a child if they are under the legal age of majority. Conversely, one may legally be an adult but possess none of the maturity and responsibility that may define an adult character.
In different cultures there are events that relate passing from being a child to becoming an adult or coming of age. This often encompasses the passing a series of tests to demonstrate that a person is prepared for adulthood, or reaching a specified age, sometimes in conjunction with demonstrating preparation. Most modern societies determine legal adulthood based on reaching a legally specified age without requiring a demonstration of physical maturity or preparation for adulthood.
Biological adulthood
Historically and cross-culturally, adulthood has been determined primarily by the start of puberty (the appearance of secondary sex characteristics such as menstruation and the development of breasts in women, ejaculation, the development of facial hair, and a deeper voice in men, and pubic hair in both sexes). In the past, a person usually moved from the status of child directly to the status of adult, often with this shift being marked by some type of
Document 1:::
Sexual maturity is the capability of an organism to reproduce. In humans, it is related to both puberty and adulthood. However, puberty is the process of biological sexual maturation, while the concept of adulthood is generally based on broader cultural definitions.
Most multicellular organisms are unable to sexually reproduce at birth (animals) or germination (e.g. plants): depending on the species, it may be days, weeks, or years until they have developed enough to be able to do so. Also, certain cues may trigger an organism to become sexually mature. They may be external, such as drought (certain plants), or internal, such as percentage of body fat (certain animals). (Such internal cues are not to be confused with hormones, which directly produce sexual maturity – the production/release of those hormones is triggered by such cues.)
Role of reproductive organs
Sexual maturity is brought about by a maturing of the reproductive organs and the production of gametes. It may also be accompanied by a growth spurt or other physical changes which distinguish the immature organism from its adult form. In animals these are termed secondary sex characteristics, and often represent an increase in sexual dimorphism.
After sexual maturity is achieved, some organisms become infertile, or even change their sex. Some organisms are hermaphrodites and may or may not be able to "completely" mature and/or to produce viable offspring. Also, while in many organisms sexual maturity is strongly linked to age, many other factors are involved, and it is possible for some to display most or all of the characteristics of the adult form without being sexually mature. Conversely it is also possible for the "immature" form of an organism to reproduce. This is called progenesis, in which sexual development occurs faster than other physiological development (in contrast, the term neoteny refers to when non-sexual development is slowed – but the result is the same - the retention of juvenile c
Document 2:::
The anal stage is the second stage in Sigmund Freud's theory of psychosexual development, taking place approximately between the ages 18 months and three years. According to Freud, the anus is the primary erogenous zone and pleasure is derived from controlling bladder and bowel movement. The major conflict issue during this stage is toilet training. A fixation at this stage can result in a personality that is too rigid or one that is too disordered.
According to Freud's theory, personality is developed through a series of stages, focused on erogenous areas, throughout childhood.
A healthy personality in adulthood is dependent upon all of these childhood stages being resolved successfully. If issues are not resolved in a stage, then fixation will occur, resulting in an unhealthy personality.
General information
The anal stage, in Freudian psychology, is the period of human development occurring at about one to three years of age. Around this age, the child begins to toilet train, which brings about the child's fascination in the erogenous zone of the anus. The erogenous zone is focused on the bowel and bladder control. Therefore, Freud believed that the libido was mainly focused on controlling the bladder and bowel movements. The anal stage coincides with the start of the child's ability to control their anal sphincter, and therefore their ability to pass or withhold feces at will. If the children during this stage can overcome the conflict it will result in a sense of accomplishment and independence.
Conflict
This is the second stage of Freud's psychosexual stages. This stage represents a conflict with the id, ego, and superego. The child is approached with this conflict with the parent's demands. A successful completion of this stage depends on how the parents interact with the child while toilet training. If a parent praises the child and gives rewards for using the toilet properly and at the right times then the child will successfully go through the stage. H
Document 3:::
A juvenile is an individual organism (especially an animal) that has not yet reached its adult form, sexual maturity or size. Juveniles can look very different from the adult form, particularly in colour, and may not fill the same niche as the adult form. In many organisms the juvenile has a different name from the adult (see List of animal names).
Some organisms reach sexual maturity in a short metamorphosis, such as ecdysis in many insects and some other arthropods. For others, the transition from juvenile to fully mature is a more prolonged process—puberty in humans and other species (like higher primates and whales), for example. In such cases, juveniles during this transformation are sometimes called subadults.
Many invertebrates cease development upon reaching adulthood. The stages of such invertebrates are larvae or nymphs.
In vertebrates and some invertebrates (e.g. spiders), larval forms (e.g. tadpoles) are usually considered a development stage of their own, and "juvenile" refers to a post-larval stage that is not fully grown and not sexually mature. In amniotes, the embryo represents the larval stage. Here, a "juvenile" is an individual in the time between hatching/birth/germination and reaching maturity.
Examples
For animal larval juveniles, see larva
Juvenile birds or bats can be called fledglings
For cat juveniles, see kitten
For dog juveniles, see puppy
For human juvenile life stages, see childhood and adolescence, an intermediary period between the onset of puberty and full physical, psychological, and social adulthood
Document 4:::
Stage-crisis view is a theory of adult development that was established by Daniel Levinson. Although largely influenced by the work of Erik Erikson, Levinson sought to create a broader theory that would encompass all aspects of adult development as opposed to just the psychosocial. This theory is characterized by both definitive eras as well as transition phases, whose purpose is to facilitate a smooth transition out of one era and into the next. According to his theory, various developmental tasks must be mastered as one progresses through each era; pre-adulthood, early adulthood, middle adulthood, and late adulthood. Crises are also experienced throughout the lifecycle and occur when one become burdened by either internal or external factors, such as during the midlife crisis that occurs during the midlife transition from early adulthood to middle adulthood.
Levinson researched both men and women, and found that they typically go through the same cycles, though he suggested that women's cycles were more closely tied to the domestic sphere, or their family life. Due to the use of biased research methods however, the extent to which his results can be generalized remains controversial. Although not widely accepted, his theories entail many implications for both behavioral and cultural psychology.
Stages of adult development
Within his theory, Daniel Levinson explains both the concept of adulthood itself as well as the stages by which development occurs. In order to do so effectively, he must differentiate between the life course and the life cycle. The life course, as he sees it, is the chronological process of living; from the beginning of life until the end. The life cycle on the other hand, is the common sequence of stages that every person undergoes during their life course. Although specific experiences may vary, the life cycle is common to all people and each person progresses through the same set of stages. Each stage, or era, is marked by specific changes
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is the name of the life stage when a child becomes sexually mature?
A. maturity
B. puberty
C. adulthood
D. adolescence
Answer:
|
|
sciq-10517
|
multiple_choice
|
A sprain is a strain or tear in what part of the body?
|
[
"skeleton",
"torso",
"a ligament",
"stomach"
] |
C
|
Relavent Documents:
Document 0:::
Soft tissue is all the tissue in the body that is not hardened by the processes of ossification or calcification such as bones and teeth. Soft tissue connects, surrounds or supports internal organs and bones, and includes muscle, tendons, ligaments, fat, fibrous tissue, lymph and blood vessels, fasciae, and synovial membranes.
It is sometimes defined by what it is not – such as "nonepithelial, extraskeletal mesenchyme exclusive of the reticuloendothelial system and glia".
Composition
The characteristic substances inside the extracellular matrix of soft tissue are the collagen, elastin and ground substance. Normally the soft tissue is very hydrated because of the ground substance. The fibroblasts are the most common cell responsible for the production of soft tissues' fibers and ground substance. Variations of fibroblasts, like chondroblasts, may also produce these substances.
Mechanical characteristics
At small strains, elastin confers stiffness to the tissue and stores most of the strain energy. The collagen fibers are comparatively inextensible and are usually loose (wavy, crimped). With increasing tissue deformation the collagen is gradually stretched in the direction of deformation. When taut, these fibers produce a strong growth in tissue stiffness. The composite behavior is analogous to a nylon stocking, whose rubber band does the role of elastin as the nylon does the role of collagen. In soft tissues, the collagen limits the deformation and protects the tissues from injury.
Human soft tissue is highly deformable, and its mechanical properties vary significantly from one person to another. Impact testing results showed that the stiffness and the damping resistance of a test subject’s tissue are correlated with the mass, velocity, and size of the striking object. Such properties may be useful for forensics investigation when contusions were induced. When a solid object impacts a human soft tissue, the energy of the impact will be absorbed by the tissues
Document 1:::
Epicondylitis is the inflammation of an epicondyle or of adjacent tissues. Epicondyles are on the medial and lateral aspects of the elbow, consisting of the two bony prominences at the distal end of the humerus. These bony projections serve as the attachment point for the forearm musculature. Inflammation to the tendons and muscles at these attachment points can lead to medial and/or lateral epicondylitis. This can occur through a range of factors that overuse the muscles that attach to the epicondyles, such as sports or job-related duties that increase the workload of the forearm musculature and place stress on the elbow. Lateral epicondylitis is also known as “Tennis Elbow” due to its sports related association to tennis athletes, while medial epicondylitis is often referred to as “golfer's elbow.”
Risk factors
In a cross-sectional population-based study among the working population, it was found that psychological distress and bending and straightening of the elbow joint for >1hr per day were associated risk factors to epicondylitis.
Another study revealed the following potential risk factors among the working population:
Force and repetitive motions (handling tools > 1 kg, handling loads >20 kg at least 10 times/day, repetitive movements > 2 h/day) were found to be associated with the occurrence of lateral epicondylitis.
Low job control and low social support were also found to be associated with lateral epicondylitis.
Exposures of force (handling loads >5 kg, handling loads >20 kg at least 10 times/day, high hand grip forces >1 h/day), repetitiveness (repetitive movements for >2 h/day) and vibration (working with vibrating tools > 2 h/day) were associated with medial epicondylitis.
In addition to repetitive activities, obesity and smoking have been implicated as independent risk factors.
Symptoms
Tender to palpation at the medial or lateral epicondyle
Pain or difficulty with wrist flexion or extension
Diminished grip strength
Pain or burning se
Document 2:::
A soft tissue injury is the damage of muscles, ligaments and tendons throughout the body. Common soft tissue injuries usually occur from a sprain, strain, a one-off blow resulting in a contusion or overuse of a particular part of the body. Soft tissue injuries can result in pain, swelling, bruising and loss of function.
Signs and symptoms
Sprains
A sprain is a type of acute injury which results from the stretching or tearing of a ligament. Depending on the severity of the sprain, the movement on the joint can be compromised since ligaments aid in the stability and support of joints. Sprains are commonly seen in vulnerable areas such as the wrists, knees and ankles. They can occur from movements such as falling on an outstretched hand or a twisting of the ankle or foot.
The severity of a sprain can be classified:
Grade 1: Only some of the fibers in the ligament are torn, and the injured site is moderately painful and swollen. Function in the joint will be unaffected for the most part.
Grade 2: Many of the ligament fibers are torn, and pain and swelling is moderate. The functionality of the joint is compromised.
Grade 3: The soft tissue is completely torn, and functionality and strength on the joint is completely compromised. In most cases, surgery is needed to repair the damage.
Strains
A strain is a type of acute injury that occurs to the muscle or tendon. Similar to sprains, it can vary in severity, from a stretching of the muscle or tendon to a complete tear of the tendon from the muscle. Some of the most common places that strains occur are in the foot, back of the leg (hamstring), or back.
Bruising (contusion)
A contusion is the discoloration of the skin, which results from underlying muscle fibers and connective tissue being crushed. This can happen in a variety of ways such as a direct blow to the skin, or a fall taken against a hard surface. The discoloration in the skin is present when blood begins to pool around the injury.
Tendinitis
Tendi
Document 3:::
Musculoskeletal injury refers to damage of muscular or skeletal systems, which is usually due to a strenuous activity and includes damage to skeletal muscles, bones, tendons, joints, ligaments, and other affected soft tissues. In one study, roughly 25% of approximately 6300 adults received a musculoskeletal injury of some sort within 12 months—of which 83% were activity-related. Musculoskeletal injury spans into a large variety of medical specialties including orthopedic surgery (with diseases such as arthritis requiring surgery), sports medicine, emergency medicine (acute presentations of joint and muscular pain) and rheumatology (in rheumatological diseases that affect joints such as rheumatoid arthritis).
Musculoskeletal injuries can affect any part of the human body including; bones, joints, cartilages, ligaments, tendons, muscles, and other soft tissues. Symptoms include mild to severe aches, low back pain, numbness, tingling, atrophy and weakness. These injuries are a result of repetitive motions and actions over a period of time. Tendons connect muscle to bone whereas ligaments connect bone to bone. Tendons and ligaments play an active role in maintain joint stability and controls the limits of joint movements, once injured tendons and ligaments detrimentally impact motor functions. Continuous exercise or movement of a musculoskeletal injury can result in chronic inflammation with progression to permanent damage or disability.
In many cases, during the healing period after a musculoskeletal injury, a period in which the healing area will be completely immobile, a cast-induced muscle atrophy can occur. Routine sessions of physiotherapy after the cast is removed can help return strength in limp muscles or tendons. Alternately, there exist different methods of electrical stimulation of the immobile muscles which can be induced by a device placed underneath a cast, helping prevent atrophies Preventative measures include correcting or modifying one's postures a
Document 4:::
Ulnar collateral ligament injuries can occur during certain activities such as overhead baseball pitching. Acute or chronic disruption of the ulnar collateral ligament result in medial elbow pain, valgus instability, and impaired throwing performance. There are both non-surgical and surgical treatment options.
Signs and symptoms
Pain along the inside of the elbow is the main symptom of this condition. Throwing athletes report it occurs most often during the acceleration phase of throwing. The injury is often associated with an experience of a sharp “pop” in the elbow, followed by pain during a single throw. In addition, swelling and bruising of the elbow, loss of elbow range of motion, and a sudden decrease in throwing velocity are all common symptoms of a UCL injury. If the injury is less severe, pain can alleviate with complete rest.
Causes
The UCL stabilizes the elbow from being abducted during a throwing motion. If intense or repeated bouts of valgus stress occur on the UCL, injury may occur. Damage to the UCL is common among baseball pitchers and javelin throwers because the throwing motion is similar. Physicians believe repetitive movements, especially pitching in baseball, cause UCL injuries. Furthermore, physicians have stated that if an adolescent throws over 85 throws for 8 months or more in a year, or throws when exhausted, the adolescent has a significantly higher risk of UCL injury.
Gridiron football, racquet sports, ice hockey and water polo players have also been treated for damage to the UCL. Specific overhead movements like those that occur during baseball pitching, tennis serving or volleyball spiking increase the risk of UCL injury. During the cocking phase of pitching, the shoulder is horizontally abducted, externally rotated and the elbow is flexed. There is slight stress on the UCL in this position but it increases when the shoulder is further externally rotated in a throw. The greater the stress the more the UCL is stretched causing stra
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
A sprain is a strain or tear in what part of the body?
A. skeleton
B. torso
C. a ligament
D. stomach
Answer:
|
|
sciq-11383
|
multiple_choice
|
What describes the process of vesicles fusing with the plasma membrane and releasing their contents to the outside of the cell?
|
[
"endocytosis",
"autolysis",
"expulsion",
"exocytosis"
] |
D
|
Relavent Documents:
Document 0:::
Porosomes are cup-shaped supramolecular structures in the cell membranes of eukaryotic cells where secretory vesicles transiently dock in the process of vesicle fusion and secretion. The transient fusion of secretory vesicle membrane at a porosome, base via SNARE proteins, results in the formation of a fusion pore or continuity for the release of intravesicular contents from the cell. After secretion is complete, the fusion pore temporarily formed at the base of the porosome is sealed. Porosomes are few nanometers in size and contain many different types of protein, especially chloride and calcium channels, actin, and SNARE proteins that mediate the docking and fusion of the vesicles with the cell membrane. Once the vesicles have docked with the SNARE proteins, they swell, which increases their internal pressure. They then transiently fuse at the base of the porosome, and these pressurized contents are ejected from the cell. Examination of cells following secretion using electron microscopy, demonstrate increased presence of partially empty vesicles following secretion. This suggested that during the secretory process, only a portion of the vesicular contents are able to exit the cell. This could only be possible if the vesicle were to temporarily establish continuity with the cell plasma membrane, expel a portion of its contents, then detach, reseal, and withdraw into the cytosol (endocytose). In this way, the secretory vesicle could be reused for subsequent rounds of exo-endocytosis, until completely empty of its contents.
Porosomes vary in size depending on the cell type. Porosome in the exocrine pancreas and in endocrine and neuroendocrine cells range from 100 nm to 180 nm in diameter while in neurons they range from 10 nm to 15 nm (about 1/10 the size of pancreatic porosomes). When a secretory vesicle containing v-SNARE docks at the porosome base containing t-SNARE, membrane continuity (ring complex) is formed between the two. The size of the t/v-SNARE complex
Document 1:::
In cellular biology, pinocytosis, otherwise known as fluid endocytosis and bulk-phase pinocytosis, is a mode of endocytosis in which small molecules dissolved in extracellular fluid are brought into the cell through an invagination of the cell membrane, resulting in their containment within a small vesicle inside the cell. These pinocytotic vesicles then typically fuse with early endosomes to hydrolyze (break down) the particles.
Pinocytosis is variably subdivided into categories depending on the molecular mechanism and the fate of the internalized molecules.
Function
In humans, this process occurs primarily for absorption of fat droplets. In endocytosis the cell plasma membrane extends and folds around desired extracellular material, forming a pouch that pinches off creating an internalized vesicle. The invaginated pinocytosis vesicles are much smaller than those generated by phagocytosis. The vesicles eventually fuse with the lysosome, whereupon the vesicle contents are digested. Pinocytosis involves a considerable investment of cellular energy in the form of ATP.
Pinocytosis and ATP
Pinocytosis is used primarily for clearing extracellular fluids (ECF) and as part of immune surveillance. In contrast to phagocytosis, it generates very small amounts of ATP from the wastes of alternative substances such as lipids (fat). Unlike receptor-mediated endocytosis, pinocytosis is nonspecific in the substances that it transport: the cell takes in surrounding fluids, including all solutes present.
Etymology and pronunciation
The word pinocytosis () uses combining forms of pino- + cyto- + -osis, all Neo-Latin from Greek, reflecting píno, to drink, and cytosis. The term was proposed by W. H. Lewis in 1931.
Non-specific, adsorptive pinocytosis
Non-specific, adsorptive pinocytosis is a form of endocytosis, a process in which small particles are taken in by a cell by splitting off small vesicles from the cell membrane. Cationic proteins bind to the negative cell surface and
Document 2:::
Microvesicles (ectosomes, or microparticles) are a type of extracellular vesicle (EV) that are released from the cell membrane. In multicellular organisms, microvesicles and other EVs are found both in tissues (in the interstitial space between cells) and in many types of body fluids. Delimited by a phospholipid bilayer, microvesicles can be as small as the smallest EVs (30 nm in diameter) or as large as 1000 nm. They are considered to be larger, on average, than intracellularly-generated EVs known as exosomes. Microvesicles play a role in intercellular communication and can transport molecules such as mRNA, miRNA, and proteins between cells.
Though initially dismissed as cellular debris, microvesicles may reflect the antigenic content of the cell of origin and have a role in cell signaling. Like other EVs, they have been implicated in numerous physiologic processes, including anti-tumor effects, tumor immune suppression, metastasis, tumor-stroma interactions, angiogenesis, and tissue regeneration. Microvesicles may also remove misfolded proteins, cytotoxic agents and metabolic waste from the cell. Changes in microvesicle levels may indicate diseases including cancer.
Formation and contents
Different cells can release microvesicles from the plasma membrane. Sources of microvesicles include megakaryocytes, blood platelets, monocytes, neutrophils, tumor cells and placenta.
Platelets play an important role in maintaining hemostasis: they promote thrombus growth, and thus they prevent loss of blood. Moreover, they enhance immune response, since they express the molecule CD154 (CD40L). Platelets are activated by inflammation, infection, or injury, and after their activation microvesicles containing CD154 are released from platelets. CD154 is a crucial molecule in the development of T cell-dependent humoral immune response. CD154 knockout mice are incapable of producing IgG, IgE, or IgA as a response to antigens. Microvesicles can also transfer prions and molecules CD4
Document 3:::
Trans-endocytosis is the biological process where material created in one cell undergoes endocytosis (enters) into another cell. If the material is large enough, this can be observed using an electron microscope. Trans-endocytosis from neurons to glia has been observed using time-lapse microscopy.
Trans-endocytosis also applies to molecules. For example, this process is involved when a part of the protein Notch is cleaved off and undergoes endocytosis into its neighboring cell. Without Notch trans-endocytosis, there would be too many neurons in a developing embryo. Trans-endocytosis is also involved in cell movement when the protein ephrin is bound by its receptor from a neighboring cell.
Document 4:::
Exocytosis () is a form of active transport and bulk transport in which a cell transports molecules (e.g., neurotransmitters and proteins) out of the cell (exo- + cytosis). As an active transport mechanism, exocytosis requires the use of energy to transport material. Exocytosis and its counterpart, endocytosis, are used by all cells because most chemical substances important to them are large polar molecules that cannot pass through the hydrophobic portion of the cell membrane by passive means. Exocytosis is the process by which a large amount of molecules are released; thus it is a form of bulk transport. Exocytosis occurs via secretory portals at the cell plasma membrane called porosomes. Porosomes are permanent cup-shaped lipoprotein structure at the cell plasma membrane, where secretory vesicles transiently dock and fuse to release intra-vesicular contents from the cell.
In exocytosis, membrane-bound secretory vesicles are carried to the cell membrane, where they dock and fuse at porosomes and their contents (i.e., water-soluble molecules) are secreted into the extracellular environment. This secretion is possible because the vesicle transiently fuses with the plasma membrane. In the context of neurotransmission, neurotransmitters are typically released from synaptic vesicles into the synaptic cleft via exocytosis; however, neurotransmitters can also be released via reverse transport through membrane transport proteins.
Exocytosis is also a mechanism by which cells are able to insert membrane proteins (such as ion channels and cell surface receptors), lipids, and other components into the cell membrane. Vesicles containing these membrane components fully fuse with and become part of the outer cell membrane.
History
The term was proposed by De Duve in 1963.
Types
In eukaryotes there are two types of exocytosis:
1) Ca2+ triggered non-constitutive (i.e., regulated exocytosis) and
2) non-Ca2+ triggered constitutive (i.e., non-regulated).
Ca2+ triggered non-const
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What describes the process of vesicles fusing with the plasma membrane and releasing their contents to the outside of the cell?
A. endocytosis
B. autolysis
C. expulsion
D. exocytosis
Answer:
|
|
sciq-6106
|
multiple_choice
|
What is the second most common skin cancer?
|
[
"basal cell carcinoma",
"precancerous cell carcinoma",
"squamous cell carcinoma",
"melanoma"
] |
C
|
Relavent Documents:
Document 0:::
Cutaneous squamous-cell carcinoma (cSCC), or squamous-cell carcinoma of the skin, also known as squamous-cell skin cancer, is, with basal-cell carcinoma and melanoma, one of the three principal types of skin cancer. cSCC typically presents as a hard lump with a scaly top layer, but it may instead form an ulcer. Onset often occurs over a period of months. Cutaneous squamous-cell carcinoma is more likely to spread to distant areas than basal cell cancer. When confined to the outermost layer of the skin, a pre-invasive, or in situ, form of cSCC is known as Bowen's disease.
The most significant risk factor for cSCC is high lifetime exposure to ultraviolet radiation from the sun. Other risks include prior scars, chronic wounds, actinic keratosis, paler skin that sunburns easily, Bowen's disease, arsenic exposure, radiation therapy, tobacco smoking, poor immune system function, prior basal cell carcinoma, and HPV infection. Risk from UV radiation is related to total exposure, rather than exposure early in life. Tanning beds have become another frequent source of ultraviolet radiation. Risk is also elevated in certain genetic skin disorders, such as xeroderma pigmentosum and certain forms of epidermolysis bullosa. cSCC begins from squamous cells found in the upper layers of the skin. Diagnosis is often based on skin examination, and confirmed by tissue biopsy.
In vivo and in vitro studies have shown that the upregulation of FGFR2, a subset of the fibroblast growth factor receptor (FGFR) immunoglobin family, has a critical role to play in the progression of cSCC cells. Mutations in the TPL2 gene cause over-expression of FGFR2, which activates the mTORC1 and AKT pathways in both primary and metastatic cSCC cell lines. By using a chemical substance, called a "pan FGFR inhibitor", cell migration and cell proliferation in cSCC have been attenuated in vitro.
Avoiding exposure to ultraviolet radiation and the use of sunscreen appear to be effective methods of preventing cutane
Document 1:::
CP-GEP is a non-invasive prediction model for cutaneous melanoma patients that combines clinicopathologic (CP) variables with gene expression profiling (GEP). CP-GEP is able to identify cutaneous melanoma patients at low-risk for nodal metastasis who may forgo the sentinel lymph node biopsy (SLNB) procedure. The CP-GEP model was developed by the Mayo Clinic and SkylineDx BV, and it has been clinically validated in multiple studies.
Clinical relevance
The sentinel lymph node biopsy (SLNB) is the standard of care for detecting nodal metastases in cutaneous melanoma patients and has been the most informative prognostic factor to guide subsequent treatment. However, ~85% of patients undergoing this procedure have no evidence of nodal metastasis. These patients are exposed to the risk of surgical complications. Well-known complications of SLNB include seroma formation, infections, lymphedema and other comorbidities. Because the SLNb procedure is highly complex, involves multiple medical disciplines, and is difficult to standardize, the false-negative rate is relatively high at 15%. Likewise, SLNB results that show minimal tumor cell deposits are difficult to interpret and may falsely indicate high-risk disease. The use of CP-GEP is expected to reduce the number of negative, nontherapeutic SLNB, as it has been specifically developed to identify and deselect patients with a low risk of nodal metastasis (below 10%). Per current clinical guidelines (NCCN, 2022), patients with a risk of having nodal metastases below 10% may choose to forgo SLNB, whereas patients with a nodal metastases risk of greater than 10% are recommended to undergo SLNB surgery. A diagnostic tool (rule-out test) that deselects patients for SLNB is therefore likely to improve clinical care. Better patient selection for SLNB would increase the accuracy of the clinicopathological assessment and reduce the exposure to unnecessary SLNB surgeries, thereby optimizing the allocation of healthcare resources. M
Document 2:::
An atypical teratoid rhabdoid tumor (AT/RT) is a rare tumor usually diagnosed in childhood. Although usually a brain tumor, AT/RT can occur anywhere in the central nervous system (CNS), including the spinal cord. About 60% will be in the posterior cranial fossa (particularly the cerebellum). One review estimated 52% in the posterior fossa, 39% are supratentorial primitive neuroectodermal tumors (sPNET), 5% are in the pineal, 2% are spinal, and 2% are multifocal.
In the United States, three children per 1,000,000 or around 30 new AT/RT cases are diagnosed each year. AT/RT represents around 3% of pediatric cancers of the CNS.
Around 17% of all pediatric cancers involve the CNS, making these cancers the most common childhood solid tumor. The survival rate for CNS tumors is around 60%. Pediatric brain cancer is the second-leading cause of childhood cancer death, just after leukemia. Recent trends suggest that the rate of overall CNS tumor diagnosis is increasing by about 2.7% per year. As diagnostic techniques using genetic markers improve and are used more often, the proportion of AT/RT diagnoses is expected to increase.
AT/RT was only recognized as an entity in 1996 and added to the World Health Organization Brain Tumor Classification in 2000 (Grade IV). The relatively recent classification and rarity has contributed to initial misdiagnosis and nonoptimal therapy. This has led to a historically poor prognosis.
Current research is focusing on using chemotherapy protocols that are effective against rhabdomyosarcoma in combination with surgery and radiation therapy.
Recent studies using multimodal therapy have shown significantly improved survival data. In 2008,
the Dana-Farber Cancer Institute in Boston reported two-year overall survival of 53% and event-free survival of 70% (median age at diagnosis of 26 months).
In 2013, the Medical University of Vienna reported five-year overall survival of 100%, and event-free survival of 89% (median age at diagnosis
Document 3:::
A vascular tumor is a tumor of vascular origin; a soft tissue growth that can be either benign or malignant, formed from blood vessels or lymph vessels. Examples of vascular tumors include hemangiomas, lymphangiomas, hemangioendotheliomas, Kaposi's sarcomas, angiosarcomas, and hemangioblastomas. An angioma refers to any type of benign vascular tumor.
Some vascular tumors can be associated with serious blood-clotting disorders, making correct diagnosis critical.
A vascular tumor may be described in terms of being highly vascularized, or poorly vascularized, referring to the degree of blood supply to the tumor.
Classification
Vascular tumors make up one of the classifications of vascular anomalies. The other grouping is vascular malformations. Vascular tumors can be further subclassified as being benign, borderline or aggressive, and malignant. Vascular tumors are described as proliferative, and vascular malformations as nonproliferative.
Types
A vascular tumor typically grows quickly by the proliferation of endothelial cells. Most are not birth defects.
Benign
The most common type of benign vascular tumors are hemangiomas, most commonly infantile hemangiomas, and less commonly congenital hemangiomas.
Infantile hemangioma
Infantile hemangiomas are the most common type of vascular tumor to affect babies, accounting for 90% of hemangiomas. They are characterised by the abnormal proliferation of endothelial cells and of deviant blood vessel formation or architecture. Hypoxic stress seems to be a major trigger for this. Infantile hemangiomas are easily diagnosed, and little if any aggressive treatment is needed. They are characterised by rapid growth in the first few months, followed by spontaneous regression in early childhood.
Congenital hemangioma
Congenital hemangiomas are present and fully formed at birth, and only account for 2% of the hemangiomas. They do not have the postnatal phase of proliferation common to infantile hemangiomas. There are two main vari
Document 4:::
Carcinoma is a malignancy that develops from epithelial cells. Specifically, a carcinoma is a cancer that begins in a tissue that lines the inner or outer surfaces of the body, and that arises from cells originating in the endodermal, mesodermal or ectodermal germ layer during embryogenesis.
Carcinomas occur when the DNA of a cell is damaged or altered and the cell begins to grow uncontrollably and become malignant. It is from the (itself derived from meaning crab).<ref>Oxford English Dictionary, 3rd edition, s.v.</ref>
Classification
As of 2004, no simple and comprehensive classification system has been devised and accepted within the scientific community. Traditionally, however, malignancies have generally been classified into various types using a combination of criteria, including:
The cell type from which they start; specifically:
Epithelial cells ⇨ carcinoma
Non-hematopoietic mesenchymal cells ⇨ sarcoma
Hematopoietic cells
Bone marrow-derived cells that normally mature in the bloodstream ⇨ leukemia
Bone marrow-derived cells that normally mature in the lymphatic system ⇨ lymphoma
Germ cells ⇨ germinoma
Other criteria that play a role include:
The degree to which the malignant cells resemble their normal, untransformed counterparts
The appearance of the local tissue and stromal architecture
The anatomical location from which tumors arise
Genetic, epigenetic, and molecular features
Histological types
Adenocarcinoma (adeno = gland) Refers to a carcinoma featuring microscopic glandular-related tissue cytology, tissue architecture, and/or gland-related molecular products, e.g., mucin.
Squamous cell carcinoma Refers to a carcinoma with observable features and characteristics indicative of squamous differentiation (intercellular bridges, keratinization, squamous pearls).
Adenosquamous carcinoma Refers to a mixed tumor containing both adenocarcinoma and squamous cell carcinoma, wherein each of these cell types comprise at least 10% of the tumor volu
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is the second most common skin cancer?
A. basal cell carcinoma
B. precancerous cell carcinoma
C. squamous cell carcinoma
D. melanoma
Answer:
|
|
sciq-4860
|
multiple_choice
|
What is the name of a tough, fibrous membrane which protects the outer layer of bone?
|
[
"marrow",
"cartilage",
"the periosteum",
"the epithelium"
] |
C
|
Relavent Documents:
Document 0:::
The perichondrium (from Greek and ) is a layer of dense irregular connective tissue that surrounds the cartilage of developing bone. It consists of two separate layers: an outer fibrous layer and inner chondrogenic layer. The fibrous layer contains fibroblasts, which produce collagenous fibres. The chondrogenic layer remains undifferentiated and can form chondroblasts. Perichondrium can be found around the perimeter of elastic cartilage and hyaline cartilage.
Perichondrium is a type of irregular collagenous ordinary connective tissue, and also functions in the growth and repair of cartilage. Perichondrium contains type I collagen and type XII collagen.
Document 1:::
The fibroelastic coat of the spleen invests the organ, and at the hilum is reflected inward upon the vessels in the form of sheaths. From these sheaths, as well as from the inner surface of the fibroelastic coat, numerous small fibrous bands, the trabeculae of the spleen (or splenic trabeculae), emerge from all directions; these uniting, constitute the frame-work of the spleen.
The spleen therefore consists of a number of small spaces or areolae, formed by the trabeculae; in these areolae is contained the splenic pulp.
See also
Spleen
Trabecula
Document 2:::
H2.00.04.4.01001: Lymphoid tissue
H2.00.05.0.00001: Muscle tissue
H2.00.05.1.00001: Smooth muscle tissue
H2.00.05.2.00001: Striated muscle tissue
H2.00.06.0.00001: Nerve tissue
H2.00.06.1.00001: Neuron
H2.00.06.2.00001: Synapse
H2.00.06.2.00001: Neuroglia
h3.01: Bones
h3.02: Joints
h3.03: Muscles
h3.04: Alimentary system
h3.05: Respiratory system
h3.06: Urinary system
h3.07: Genital system
h3.08:
Document 3:::
In biology, the extracellular matrix (ECM), is a network consisting of extracellular macromolecules and minerals, such as collagen, enzymes, glycoproteins and hydroxyapatite that provide structural and biochemical support to surrounding cells. Because multicellularity evolved independently in different multicellular lineages, the composition of ECM varies between multicellular structures; however, cell adhesion, cell-to-cell communication and differentiation are common functions of the ECM.
The animal extracellular matrix includes the interstitial matrix and the basement membrane. Interstitial matrix is present between various animal cells (i.e., in the intercellular spaces). Gels of polysaccharides and fibrous proteins fill the interstitial space and act as a compression buffer against the stress placed on the ECM. Basement membranes are sheet-like depositions of ECM on which various epithelial cells rest. Each type of connective tissue in animals has a type of ECM: collagen fibers and bone mineral comprise the ECM of bone tissue; reticular fibers and ground substance comprise the ECM of loose connective tissue; and blood plasma is the ECM of blood.
The plant ECM includes cell wall components, like cellulose, in addition to more complex signaling molecules. Some single-celled organisms adopt multicellular biofilms in which the cells are embedded in an ECM composed primarily of extracellular polymeric substances (EPS).
Structure
Components of the ECM are produced intracellularly by resident cells and secreted into the ECM via exocytosis. Once secreted, they then aggregate with the existing matrix. The ECM is composed of an interlocking mesh of fibrous proteins and glycosaminoglycans (GAGs).
Proteoglycans
Glycosaminoglycans (GAGs) are carbohydrate polymers and mostly attached to extracellular matrix proteins to form proteoglycans (hyaluronic acid is a notable exception; see below). Proteoglycans have a net negative charge that attracts positively charged sod
Document 4:::
Outline
h1.00: Cytology
h2.00: General histology
H2.00.01.0.00001: Stem cells
H2.00.02.0.00001: Epithelial tissue
H2.00.02.0.01001: Epithelial cell
H2.00.02.0.02001: Surface epithelium
H2.00.02.0.03001: Glandular epithelium
H2.00.03.0.00001: Connective and supportive tissues
H2.00.03.0.01001: Connective tissue cells
H2.00.03.0.02001: Extracellular matrix
H2.00.03.0.03001: Fibres of connective tissues
H2.00.03.1.00001: Connective tissue proper
H2.00.03.1.01001: Ligaments
H2.00.03.2.00001: Mucoid connective tissue; Gelatinous connective tissue
H2.00.03.3.00001: Reticular tissue
H2.00.03.4.00001: Adipose tissue
H2.00.03.5.00001: Cartilage tissue
H2.00.03.6.00001: Chondroid tissue
H2.00.03.7.00001: Bone tissue; Osseous tissue
H2.00.04.0.00001: Haemotolymphoid complex
H2.00.04.1.00001: Blood cells
H2.00.04.1.01001: Erythrocyte; Red blood cell
H2.00.04.1.02001: Leucocyte; White blood cell
H2.00.04.1.03001: Platelet; Thrombocyte
H2.00.04.2.00001: Plasma
H2.00.04.3.00001: Blood cell production
H2.00.04.4.00001: Postnatal sites of haematopoiesis
H2.00.04.4.01001: Lymphoid tissue
H2.00.05.0.00001: Muscle tissue
H2.00.05.1.00001: Smooth muscle tissue
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is the name of a tough, fibrous membrane which protects the outer layer of bone?
A. marrow
B. cartilage
C. the periosteum
D. the epithelium
Answer:
|
|
sciq-5553
|
multiple_choice
|
Female monotremes share what external opening with reptiles and birds?
|
[
"vagina",
"cloaca",
"urethra",
"anus"
] |
B
|
Relavent Documents:
Document 0:::
Several universities have designed interdisciplinary courses with a focus on human biology at the undergraduate level. There is a wide variation in emphasis ranging from business, social studies, public policy, healthcare and pharmaceutical research.
Americas
Human Biology major at Stanford University, Palo Alto (since 1970)
Stanford's Human Biology Program is an undergraduate major; it integrates the natural and social sciences in the study of human beings. It is interdisciplinary and policy-oriented and was founded in 1970 by a group of Stanford faculty (Professors Dornbusch, Ehrlich, Hamburg, Hastorf, Kennedy, Kretchmer, Lederberg, and Pittendrigh). It is a very popular major and alumni have gone to post-graduate education, medical school, law, business and government.
Human and Social Biology (Caribbean)
Human and Social Biology is a Level 4 & 5 subject in the secondary and post-secondary schools in the Caribbean and is optional for the Caribbean Secondary Education Certification (CSEC) which is equivalent to Ordinary Level (O-Level) under the British school system. The syllabus centers on structure and functioning (anatomy, physiology, biochemistry) of human body and the relevance to human health with Caribbean-specific experience. The syllabus is organized under five main sections: Living organisms and the environment, life processes, heredity and variation, disease and its impact on humans, the impact of human activities on the environment.
Human Biology Program at University of Toronto
The University of Toronto offers an undergraduate program in Human Biology that is jointly offered by the Faculty of Arts & Science and the Faculty of Medicine. The program offers several major and specialist options in: human biology, neuroscience, health & disease, global health, and fundamental genetics and its applications.
Asia
BSc (Honours) Human Biology at All India Institute of Medical Sciences, New Delhi (1980–2002)
BSc (honours) Human Biology at AIIMS (New
Document 1:::
The Science, Technology, Engineering and Mathematics Network or STEMNET is an educational charity in the United Kingdom that seeks to encourage participation at school and college in science and engineering-related subjects (science, technology, engineering, and mathematics) and (eventually) work.
History
It is based at Woolgate Exchange near Moorgate tube station in London and was established in 1996. The chief executive is Kirsten Bodley. The STEMNET offices are housed within the Engineering Council.
Function
Its chief aim is to interest children in science, technology, engineering and mathematics. Primary school children can start to have an interest in these subjects, leading secondary school pupils to choose science A levels, which will lead to a science career. It supports the After School Science and Engineering Clubs at schools. There are also nine regional Science Learning Centres.
STEM ambassadors
To promote STEM subjects and encourage young people to take up jobs in these areas, STEMNET have around 30,000 ambassadors across the UK. these come from a wide selection of the STEM industries and include TV personalities like Rob Bell.
Funding
STEMNET used to receive funding from the Department for Education and Skills. Since June 2007, it receives funding from the Department for Children, Schools and Families and Department for Innovation, Universities and Skills, since STEMNET sits on the chronological dividing point (age 16) of both of the new departments.
See also
The WISE Campaign
Engineering and Physical Sciences Research Council
National Centre for Excellence in Teaching Mathematics
Association for Science Education
Glossary of areas of mathematics
Glossary of astronomy
Glossary of biology
Glossary of chemistry
Glossary of engineering
Glossary of physics
Document 2:::
The Bachelor of Science in Aquatic Resources and Technology (B.Sc. in AQT) (or Bachelor of Aquatic Resource) is an undergraduate degree that prepares students to pursue careers in the public, private, or non-profit sector in areas such as marine science, fisheries science, aquaculture, aquatic resource technology, food science, management, biotechnology and hydrography. Post-baccalaureate training is available in aquatic resource management and related areas.
The Department of Animal Science and Export Agriculture, at the Uva Wellassa University of Badulla, Sri Lanka, has the largest enrollment of undergraduate majors in Aquatic Resources and Technology, with about 200 students as of 2014.
The Council on Education for Aquatic Resources and Technology includes undergraduate AQT degrees in the accreditation review of Aquatic Resources and Technology programs and schools.
See also
Marine Science
Ministry of Fisheries and Aquatic Resources Development
Document 3:::
Female education in STEM refers to child and adult female representation in the educational fields of science, technology, engineering, and mathematics (STEM). In 2017, 33% of students in STEM fields were women.
The organization UNESCO has stated that this gender disparity is due to discrimination, biases, social norms and expectations that influence the quality of education women receive and the subjects they study. UNESCO also believes that having more women in STEM fields is desirable because it would help bring about sustainable development.
Current status of girls and women in STEM education
Overall trends in STEM education
Gender differences in STEM education participation are already visible in early childhood care and education in science- and math-related play, and become more pronounced at higher levels of education. Girls appear to lose interest in STEM subjects with age, particularly between early and late adolescence. This decreased interest affects participation in advanced studies at the secondary level and in higher education. Female students represent 35% of all students enrolled in STEM-related fields of study at this level globally. Differences are also observed by disciplines, with female enrollment lowest in engineering, manufacturing and construction, natural science, mathematics and statistics and ICT fields. Significant regional and country differences in female representation in STEM studies can be observed, though, suggesting the presence of contextual factors affecting girls’ and women's engagement in these fields. Women leave STEM disciplines in disproportionate numbers during their higher education studies, in their transition to the world of work and even in their career cycle.
Learning achievement in STEM education
Data on gender differences in learning achievement present a complex picture, depending on what is measured (subject, knowledge acquisition against knowledge application), the level of education/age of students, and
Document 4:::
Animal science is described as "studying the biology of animals that are under the control of humankind". It can also be described as the production and management of farm animals. Historically, the degree was called animal husbandry and the animals studied were livestock species, like cattle, sheep, pigs, poultry, and horses. Today, courses available look at a broader area, including companion animals, like dogs and cats, and many exotic species. Degrees in Animal Science are offered at a number of colleges and universities. Animal science degrees are often offered at land-grant universities, which will often have on-campus farms to give students hands-on experience with livestock animals.
Education
Professional education in animal science prepares students for careers in areas such as animal breeding, food and fiber production, nutrition, animal agribusiness, animal behavior, and welfare. Courses in a typical Animal Science program may include genetics, microbiology, animal behavior, nutrition, physiology, and reproduction. Courses in support areas, such as genetics, soils, agricultural economics and marketing, legal aspects, and the environment also are offered.
Bachelor degree
At many universities, a Bachelor of Science (BS) degree in Animal Science allows emphasis in certain areas. Typical areas are species-specific or career-specific. Species-specific areas of emphasis prepare students for a career in dairy management, beef management, swine management, sheep or small ruminant management, poultry production, or the horse industry. Other career-specific areas of study include pre-veterinary medicine studies, livestock business and marketing, animal welfare and behavior, animal nutrition science, animal reproduction science, or genetics. Youth programs are also an important part of animal science programs.
Pre-veterinary emphasis
Many schools that offer a degree option in Animal Science also offer a pre-veterinary emphasis such as Iowa State University, th
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Female monotremes share what external opening with reptiles and birds?
A. vagina
B. cloaca
C. urethra
D. anus
Answer:
|
|
sciq-2240
|
multiple_choice
|
What are ocean waves mainly caused by?
|
[
"humidity",
"wind",
"gravity",
"tides"
] |
B
|
Relavent Documents:
Document 0:::
Wind-wave dissipation or "swell dissipation" is process in which a wave generated via a weather system loses its mechanical energy transferred from the atmosphere via wind. Wind waves, as their name suggests, are generated by wind transferring energy from the atmosphere to the ocean's surface, capillary gravity waves play an essential role in this effect, "wind waves" or "swell" are also known as surface gravity waves.
General physics and theory
The process of wind-wave dissipation can be explained by applying energy spectrum theory in a similar manner as for the formation of wind-waves (generally assuming spectral dissipation is a function of wave spectrum). However, although even some of recent innovative improvements for field observations (such as Banner & Babanin et al. ) have contributed to solve the riddles of wave breaking behaviors, unfortunately there hasn't been a clear understanding for exact theories of the wind wave dissipation process still yet because of its non-linear behaviors.
By past and present observations and derived theories, the physics of the ocean-wave dissipation can be categorized by its passing regions along to water depth. In deep water, wave dissipation occurs by the actions of friction or drag forces such as opposite-directed winds or viscous forces generated by turbulent flows—usually nonlinear forces. In shallow water, the behaviors of wave dissipations are mostly types of shore wave breaking (see Types of wave breaking).
Some of simple general descriptions of wind-wave dissipation (defined by Luigi Cavaleri et al. ) were proposed when we consider only ocean surface waves such as wind waves. By means of the simple, the interactions of waves with the vertical structure of the upper layers of the ocean are ignored for simplified theory in many proposed mechanisms.
Sources of wind-wave dissipation
In general understanding, the physics of wave dissipation can be categorized by considering with its dissipation sources, such as 1) wa
Document 1:::
This list of rogue waves compiles incidents of known and likely rogue waves – also known as freak waves, monster waves, killer waves, and extreme waves. These are dangerous and rare ocean surface waves that unexpectedly reach at least twice the height of the tallest waves around them, and are often described by witnesses as "walls of water". They occur in deep water, usually far out at sea, and are a threat even to capital ships , ocean liners and land structures such as lighthouses.
In addition to the incidents listed below, it has also been suggested that these types of waves may be responsible for the loss of several low-flying United States Coast Guard helicopters on search and rescue missions.
Background
Anecdotal evidence from mariners' testimonies and incidents of wave damage to ships have long suggested rogue waves occurred; however, their scientific measurement was positively confirmed only following measurements of the Draupner wave, a rogue wave at the Draupner platform, in the North Sea on 1 January 1995. During this event, minor damage was inflicted on the platform, confirming that the reading was valid.
In modern oceanography, rogue waves are defined not as the biggest possible waves at sea, but instead as extreme sized waves for a given sea state.
Many of these encounters are only reported in the media, and are not examples of open ocean rogue waves. Often a huge wave is loosely and incorrectly denoted as a rogue wave. Extremely large waves offer an explanation for the otherwise-inexplicable disappearance of many ocean-going vessels. However, the claim is contradicted by information held by Lloyd's Register. One of the very few cases where evidence suggests a freak wave incident is the 1978 loss of the freighter . This claim, however, is contradicted by other sources, which maintain that, over a time period from 1969 to 1994 alone, rogue waves were responsible for the complete loss of 22 supertankers, often with their entire crew. In 2007, resear
Document 2:::
In fluid dynamics, dispersion of water waves generally refers to frequency dispersion, which means that waves of different wavelengths travel at different phase speeds. Water waves, in this context, are waves propagating on the water surface, with gravity and surface tension as the restoring forces. As a result, water with a free surface is generally considered to be a dispersive medium.
For a certain water depth, surface gravity waves – i.e. waves occurring at the air–water interface and gravity as the only force restoring it to flatness – propagate faster with increasing wavelength. On the other hand, for a given (fixed) wavelength, gravity waves in deeper water have a larger phase speed than in shallower water. In contrast with the behavior of gravity waves, capillary waves (i.e. only forced by surface tension) propagate faster for shorter wavelengths.
Besides frequency dispersion, water waves also exhibit amplitude dispersion. This is a nonlinear effect, by which waves of larger amplitude have a different phase speed from small-amplitude waves.
Frequency dispersion for surface gravity waves
This section is about frequency dispersion for waves on a fluid layer forced by gravity, and according to linear theory. For surface tension effects on frequency dispersion, see surface tension effects in Airy wave theory and capillary wave.
Wave propagation and dispersion
The simplest propagating wave of unchanging form is a sine wave. A sine wave with water surface elevation η( x, t ) is given by:
where a is the amplitude (in metres) and θ = θ( x, t ) is the phase function (in radians), depending on the horizontal position ( x , in metres) and time ( t , in seconds):
with and
where:
λ is the wavelength (in metres),
T is the period (in seconds),
k is the wavenumber (in radians per metre) and
ω is the angular frequency (in radians per second).
Characteristic phases of a water wave are:
the upward zero-crossing at θ = 0,
the wave crest at θ = ½ π,
th
Document 3:::
The Antarctic Circumpolar Wave (ACW) is a coupled ocean/atmosphere wave that circles the Southern Ocean in approximately eight years at . Since it is a wave-2 phenomenon (there are two ridges and two troughs in a latitude circle) at each fixed point in space a signal with a period of four years is seen. The wave moves eastward with the prevailing currents.
History of the concept
Although the "wave" is seen in temperature, atmospheric pressure, sea ice and ocean height, the variations are hard to see in the raw data and need to be filtered to become apparent. Because the reliable record for the Southern Ocean is short (since the early 1980s) and signal processing is needed to reveal its existence, some climatologists doubt the existence of the wave. Others accept its existence but say that it varies in strength over decades.
The wave was discovered simultaneously by and . Since then, ideas about the wave structure and maintenance mechanisms have changed and grown: by some accounts it is now to be considered as part of a global ENSO wave.
See also
Antarctic Circle
Antarctic Convergence
Document 4:::
Wave loading is most commonly the application of a pulsed or wavelike load to a material or object. This is most commonly used in the analysis of piping, ships, or building structures which experience wind, water, or seismic disturbances.
Examples of wave loading
Offshore storms and pipes: As large waves pass over shallowly buried pipes, water pressure increases above it. As the trough approaches, pressure over the pipe drops and this sudden and repeated variation in pressure can break pipes. The difference in pressure for a wave with wave height of about 10 m would be equivalent to one atmosphere (101.3 kPa or 14.7 psi) pressure variation between crest and trough and repeated fluctuations over pipes in relatively shallow environments could set up resonance vibrations within pipes or structures and cause problems.
Engineering oil platforms: The effects of wave-loading are a serious issue for engineers designing oil platforms, which must contend with the effects of wave loading, and have devised a number of algorithms to do so.
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What are ocean waves mainly caused by?
A. humidity
B. wind
C. gravity
D. tides
Answer:
|
|
sciq-3528
|
multiple_choice
|
What is often the cause of phenotypic differences between identical twins?
|
[
"environment",
"genetics",
"natural selection",
"learned behavior"
] |
A
|
Relavent Documents:
Document 0:::
A twin registry is a database of information about both identical twins and fraternal twins, which is often maintained by an academic institution, such as a university, or by other research institutions.
Investigative use
The use of twins can improve the statistical power of a genetic study by reducing the amount of genetic and/or environmental variability. "Identical twins" (monozygotic (MZ) twins) share virtually all their genes with each other, and "fraternal twins" (dizygotic (DZ) twins), on average, share about 50% of their genes with each other (about the same amount of sharing as non-twin siblings). Both types of twin pairs in twin registries almost always share similar prenatal and early childhood environments as well. By determining what are called "concordance" rates for a disease or trait among identical and fraternal twin pairs, researchers can estimate whether contributing factors for that disease or trait are more likely to be hereditary, environmental, or some combination of these. A concordance rate is a statistical measure of probability - if one twin has a specific trait or disease, what is the probability that the other twin has (or will develop) that same trait or disease. In addition, with structural equation modeling and multivariate analyses of twin data, researchers can offer estimates of the extent to which allelic variants and environment may influence phenotypic traits.
Where maintained
Some twin registries seek to cover all twins in an entire country, including Sweden, Denmark, Norway, Finland, Australia, Sri Lanka and the United Kingdom. The Swedish Twin Registry is the largest twin database in the world, with approximately 85,000 twin pairs.
Other twin registries cover a more limited geographic scope and are maintained by researchers at academic institutions, such as the Michigan State University Twin Registry, a registry of twins produced by researchers at Michigan State University, the Washington State Twin Registry, a registry
Document 1:::
In genetics, concordance is the probability that a pair of individuals will both have a certain characteristic (phenotypic trait) given that one of the pair has the characteristic. Concordance can be measured with concordance rates, reflecting the odds of one person having the trait if the other does. Important clinical examples include the chance of offspring having a certain disease if the mother has it, if the father has it, or if both parents have it. Concordance among siblings is similarly of interest: what are the odds of a subsequent offspring having the disease if an older child does? In research, concordance is often discussed in the context of both members of a pair of twins. Twins are concordant when both have or both lack a given trait. The ideal example of concordance is that of identical twins, because the genome is the same, an equivalence that helps in discovering causation via deconfounding, regarding genetic effects versus epigenetic and environmental effects (nature versus nurture).
In contrast, discordance occurs when a similar trait is not shared by the persons. Studies of twins have shown that genetic traits of monozygotic twins are fully concordant, whereas in dizygotic twins, half of genetic traits are concordant, while the other half are discordant. Discordant rates that are higher than concordant rates express the influence of the environment on twin traits.
Studies
A twin study compares the concordance rate of identical twins to that of fraternal twins. This can help suggest whether a disease or a certain trait has a genetic cause. Controversial uses of twin data have looked at concordance rates for homosexuality and intelligence. Other studies have involved looking at the genetic and environmental factors that can lead to increased LDL in women twins.
Because identical twins are genetically virtually identical, it follows that a genetic pattern carried by one would very likely also be carried by the other. If a characteristic ident
Document 2:::
Family resemblance refers to physical similarities shared between close relatives, especially between parents and children and between siblings. In psychology, the similarities of personality are also observed.
Genetics
Heritability, defined as a measure of family resemblance, causes traits to be genetically passed from parents to offspring (heredity), allowing evolutionarily advantageous traits to persist through generations. Despite sharing parents, siblings do not inherit identical genes, making studies on identical twins (who have identical DNA) especially effective at analyzing the role genetics play in phenotypic similarity. Studies have found that generational resemblance of many phenotypic traits results from the inheritance of multiples genes that collectively influence a trait (additive genetic variance). There is evidence of heritability in personality traits. For example, one study found that approximately half of personality differences in high-school aged fraternal and identical twins were due to genetic variation - and another study suggests that no one personality trait is more heritable than another.
Environment
Family resemblance is also shaped by environmental factors, temperature, light, nutrition, exposure to drugs, the time that different family members spend in shared and non-shared environments, are examples of factors found to influence phenotype. Phenotypes found to be largely environmentally determined in humans include personality, height, and weight. Twin studies have shown that more than half of the variation in a few major aspects of personality are environmentally determined, and that environmental factors even affect traits like immune response and how children handle stress. Additionally, anomalous findings, such as second-degree relatives of alcoholics, showing surprising similarities to them have led some researchers’ attempts in generating better models that account for the environmental impacts on influences like cultural in
Document 3:::
Twin studies are studies conducted on identical or fraternal twins. They aim to reveal the importance of environmental and genetic influences for traits, phenotypes, and disorders. Twin research is considered a key tool in behavioral genetics and in related fields, from biology to psychology. Twin studies are part of the broader methodology used in behavior genetics, which uses all data that are genetically informative – siblings studies, adoption studies, pedigree, etc. These studies have been used to track traits ranging from personal behavior to the presentation of severe mental illnesses such as schizophrenia.
Twins are a valuable source for observation because they allow the study of environmental influence and varying genetic makeup: "identical" or monozygotic (MZ) twins share essentially 100% of their genes, which means that most differences between the twins (such as height, susceptibility to boredom, intelligence, depression, etc.) are due to experiences that one twin has but not the other twin. "Fraternal" or dizygotic (DZ) twins share only about 50% of their genes, the same as any other sibling. Twins also share many aspects of their environment (e.g., uterine environment, parenting style, education, wealth, culture, community) because they are born into the same family. The presence of a given genetic or phenotypic trait in only one member of a pair of identical twins (called discordance) provides a powerful window into environmental effects on such a trait.
Twins are also useful in showing the importance of the unique environment (specific to one twin or the other) when studying trait presentation. Changes in the unique environment can stem from an event or occurrence that has only affected one twin. This could range from a head injury or a birth defect that one twin has sustained while the other remains healthy.
The classical twin design compares the similarity of monozygotic (identical) and dizygotic (fraternal) twins. If identical twins are consi
Document 4:::
Research on the heritability of IQ inquires into the degree of variation in IQ within a population that is due to genetic variation between individuals in that population. There has been significant controversy in the academic community about the heritability of IQ since research on the issue began in the late nineteenth century. Intelligence in the normal range is a polygenic trait, meaning that it is influenced by more than one gene, and in the case of intelligence at least 500 genes. Further, explaining the similarity in IQ of closely related persons requires careful study because environmental factors may be correlated with genetic factors.
Early twin studies of adult individuals have found a heritability of IQ between 57% and 73%, with some recent studies showing heritability for IQ as high as 80%. IQ goes from being weakly correlated with genetics for children, to being strongly correlated with genetics for late teens and adults. The heritability of IQ increases with the child's age and reaches a plateau at 14-16 years old, continuing at that level well into adulthood. However, poor prenatal environment, malnutrition and disease are known to have lifelong deleterious effects.
Although IQ differences between individuals have been shown to have a large hereditary component, it does not follow that disparities in IQ between groups have a genetic basis. The scientific consensus is that genetics does not explain average differences in IQ test performance between racial groups.
Heritability and caveats
Heritability is a statistic used in the fields of breeding and genetics that estimates the degree of variation in a phenotypic trait in a population that is due to genetic variation between individuals in that population. The concept of heritability can be expressed in the form of the following question: "What is the proportion of the variation in a given trait within a population that is not explained by the environment or random chance?"
Estimates of heritabi
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is often the cause of phenotypic differences between identical twins?
A. environment
B. genetics
C. natural selection
D. learned behavior
Answer:
|
|
sciq-329
|
multiple_choice
|
Rapidly produced genetic vairants are found in organisms with what type of generation time?
|
[
"oscillating",
"long",
"short",
"repeating"
] |
C
|
Relavent Documents:
Document 0:::
The Bateson Lecture is an annual genetics lecture held as a part of the John Innes Symposium since 1972, in honour of the first Director of the John Innes Centre, William Bateson.
Past Lecturers
Source: John Innes Centre
1951 Sir Ronald Fisher - "Statistical methods in Genetics"
1953 Julian Huxley - "Polymorphic variation: a problem in genetical natural history"
1955 Sidney C. Harland - "Plant breeding: present position and future perspective"
1957 J.B.S. Haldane - "The theory of evolution before and after Bateson"
1959 Kenneth Mather - "Genetics Pure and Applied"
1972 William Hayes - "Molecular genetics in retrospect"
1974 Guido Pontecorvo - "Alternatives to sex: genetics by means of somatic cells"
1976 Max F. Perutz - "Mechanism of respiratory haemoglobin"
1979 J. Heslop-Harrison - "The forgotten generation: some thoughts on the genetics and physiology of Angiosperm Gametophytes "
1982 Sydney Brenner - "Molecular genetics in prospect"
1984 W.W. Franke - "The cytoskeleton - the insoluble architectural framework of the cell"
1986 Arthur Kornberg - "Enzyme systems initiating replication at the origin of the E. coli chromosome"
1988 Gottfried Schatz - "Interaction between mitochondria and the nucleus"
1990 Christiane Nusslein-Volhard - "Axis determination in the Drosophila embryo"
1992 Frank Stahl - "Genetic recombination: thinking about it in phage and fungi"
1994 Ira Herskowitz - "Violins and orchestras: what a unicellular organism can do"
1996 R.J.P. Williams - "An Introduction to Protein Machines"
1999 Eugene Nester - "DNA and Protein Transfer from Bacteria to Eukaryotes - the Agrobacterium story"
2001 David Botstein - "Extracting biological information from DNA Microarray Data"
2002 Elliot Meyerowitz
2003 Thomas Steitz - "The Macromolecular machines of gene expression"
2008 Sean Carroll - "Endless flies most beautiful: the role of cis-regulatory sequences in the evolution of animal form"
2009 Sir Paul Nurse - "Genetic transmission through
Document 1:::
The Investigative Biology Teaching Laboratories are located at Cornell University on the first floor Comstock Hall. They are well-equipped biology teaching laboratories used to provide hands-on laboratory experience to Cornell undergraduate students. Currently, they are the home of the Investigative Biology Laboratory Course, (BioG1500), and frequently being used by the Cornell Institute for Biology Teachers, the Disturbance Ecology course and Insectapalooza. In the past the Investigative Biology Teaching Laboratories hosted the laboratory portion of the Introductory Biology Course with the course number of Bio103-104 (renumbered to BioG1103-1104).
The Investigative Biology Teaching Laboratories house the Science Communication and Public Engagement Undergraduate Minor.
History
Bio103-104
BioG1103-1104 Biological Sciences Laboratory course was a two-semester, two-credit course. BioG1103 was offered in the spring, while 1104 was offered in the fall.
BioG1500
This course was first offered in Fall 2010. It is a one semester course, offered in the Fall, Spring and Summer for 2 credits. One credit is being awarded for the letter and one credit for the three-hour-long lab, following the SUNY system.
Document 2:::
The online project Virtual Laboratory. Essays and Resources on the Experimentalization of Life, 1830-1930, located at the Max Planck Institute for the History of Science, is dedicated to research in the history of the experimentalization of life. The term experimentalization describes the interaction between the life sciences, the arts, architecture, media and technology within the experimental paradigm, ca. 1830 to 1930. The Virtual Laboratory is a platform that not only presents work on this topic but also acts as a research environment for new studies.
History
In 1977, the first version of the Virtual Laboratory was presented, titled Virtual Laboratory of Physiology. At this time, the main focus lay on the development of technological preconditions of physiological research in the 19th century. Therefore, a database with relevant texts and images was created. In 1998, the concept still used today was created after a series of modifications, followed by the publication of a cd-ROM in 1999. At this time, the focus had been expanded from physiology to the life sciences in general, as well as the arts and literature. As the project had been extended from a sole database to a platform for historiographical research, it was presented at the conference Using the World Wide Web for Historical Research in Science and Technology organized by the Alfred P. Sloan Foundation at Stanford University. In 2000, the project was incorporated into the research project The Experimentalization of Life, funded by the Volkswagen Foundation. This was followed by another presentation at the conference Virtual Research? The impact of new technologies on scientific practices at the ETH Zurich. In 2002, the first version of the Virtual Laboratory went online. Since 2008, the Virtual Laboratory is listed as a journal under the ISSN number 1866-4784.
Structure
The Virtual Laboratory consists of two parts: The archive holds a large number of digitized texts and images as well as data sheets c
Document 3:::
A pre-STEM program is a course of study at any two-year college that prepares a student to transfer to a four-year school to earn a bachelor's degree in a STEM field.
Overview
The concept of a pre-STEM program is being developed to address America's need for more college-trained professionals in science, technology, engineering, and mathematics (STEM). It is an innovation meant to fill a gap at community colleges that do not have 'major' degree paths that students identify with on their way to earning an Associates degree. Students must complete a considerable amount of STEM coursework before transferring from a two-year school to a four-year school and earn a baccalaureate degree in a STEM field. Schools with a pre-STEM program are able to identify those students and support them with STEM-specific academic and career advising, increasing the student's chances of going on to earn a STEM baccalaureate degree in a timely fashion.
With over 50% of America's college-bound students starting their college career at public or private two-year school, and with a very small proportion of students who start college at a two-year school matriculating to and earning STEM degrees from four-year schools, pre-STEM programs have great potential for broadening participation in baccalaureate STEM studies.
Example programs
The effectiveness of pre-STEM programs is being investigated by a consortium of schools in Missouri: Moberly Area Community College, St. Charles Community College, Metropolitan Community College, and Truman State University.
A larger group of schools met at the Belknap Springs Meetings in October 2009 to discuss the challenges and opportunities presented by STEM-focused partnerships between 2-year and 4-year schools. Each program represented a two-year school and a four-year school that were trying to increase the number of people who earn a baccalaureate degree in a STEM area through various means, some of which were pre-STEM programs. Other methods includes
Document 4:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Rapidly produced genetic vairants are found in organisms with what type of generation time?
A. oscillating
B. long
C. short
D. repeating
Answer:
|
|
sciq-6221
|
multiple_choice
|
What is the process by which organisms change in small ways over time?
|
[
"microevolution",
"adaptation",
"spontaneous mutation",
"natural selection"
] |
A
|
Relavent Documents:
Document 0:::
In biology, evolution is the process of change in all forms of life over generations, and evolutionary biology is the study of how evolution occurs. Biological populations evolve through genetic changes that correspond to changes in the organisms' observable traits. Genetic changes include mutations, which are caused by damage or replication errors in organisms' DNA. As the genetic variation of a population drifts randomly over generations, natural selection gradually leads traits to become more or less common based on the relative reproductive success of organisms with those traits.
The age of the Earth is about 4.5 billion years. The earliest undisputed evidence of life on Earth dates from at least 3.5 billion years ago. Evolution does not attempt to explain the origin of life (covered instead by abiogenesis), but it does explain how early lifeforms evolved into the complex ecosystem that we see today. Based on the similarities between all present-day organisms, all life on Earth is assumed to have originated through common descent from a last universal ancestor from which all known species have diverged through the process of evolution.
All individuals have hereditary material in the form of genes received from their parents, which they pass on to any offspring. Among offspring there are variations of genes due to the introduction of new genes via random changes called mutations or via reshuffling of existing genes during sexual reproduction. The offspring differs from the parent in minor random ways. If those differences are helpful, the offspring is more likely to survive and reproduce. This means that more offspring in the next generation will have that helpful difference and individuals will not have equal chances of reproductive success. In this way, traits that result in organisms being better adapted to their living conditions become more common in descendant populations. These differences accumulate resulting in changes within the population. This proce
Document 1:::
Evolutionary biology is the subfield of biology that studies the evolutionary processes (natural selection, common descent, speciation) that produced the diversity of life on Earth. It is also defined as the study of the history of life forms on Earth. Evolution holds that all species are related and gradually change over generations. In a population, the genetic variations affect the phenotypes (physical characteristics) of an organism. These changes in the phenotypes will be an advantage to some organisms, which will then be passed onto their offspring. Some examples of evolution in species over many generations are the peppered moth and flightless birds. In the 1930s, the discipline of evolutionary biology emerged through what Julian Huxley called the modern synthesis of understanding, from previously unrelated fields of biological research, such as genetics and ecology, systematics, and paleontology.
The investigational range of current research has widened to encompass the genetic architecture of adaptation, molecular evolution, and the different forces that contribute to evolution, such as sexual selection, genetic drift, and biogeography. Moreover, the newer field of evolutionary developmental biology ("evo-devo") investigates how embryogenesis is controlled, thus yielding a wider synthesis that integrates developmental biology with the fields of study covered by the earlier evolutionary synthesis.
Subfields
Evolution is the central unifying concept in biology. Biology can be divided into various ways. One way is by the level of biological organization, from molecular to cell, organism to population. Another way is by perceived taxonomic group, with fields such as zoology, botany, and microbiology, reflecting what was once seen as the major divisions of life. A third way is by approaches, such as field biology, theoretical biology, experimental evolution, and paleontology. These alternative ways of dividing up the subject have been combined with evolution
Document 2:::
Developmental systems theory (DST) is an overarching theoretical perspective on biological development, heredity, and evolution. It emphasizes the shared contributions of genes, environment, and epigenetic factors on developmental processes. DST, unlike conventional scientific theories, is not directly used to help make predictions for testing experimental results; instead, it is seen as a collection of philosophical, psychological, and scientific models of development and evolution. As a whole, these models argue the inadequacy of the modern evolutionary synthesis on the roles of genes and natural selection as the principal explanation of living structures. Developmental systems theory embraces a large range of positions that expand biological explanations of organismal development and hold modern evolutionary theory as a misconception of the nature of living processes.
Overview
All versions of developmental systems theory espouse the view that:
All biological processes (including both evolution and development) operate by continually assembling new structures.
Each such structure transcends the structures from which it arose and has its own systematic characteristics, information, functions and laws.
Conversely, each such structure is ultimately irreducible to any lower (or higher) level of structure, and can be described and explained only on its own terms.
Furthermore, the major processes through which life as a whole operates, including evolution, heredity and the development of particular organisms, can only be accounted for by incorporating many more layers of structure and process than the conventional concepts of ‘gene’ and ‘environment’ normally allow for.
In other words, although it does not claim that all structures are equal, development systems theory is fundamentally opposed to reductionism of all kinds. In short, developmental systems theory intends to formulate a perspective which does not presume the causal (or ontological) priority of any p
Document 3:::
Temporal plasticity, also known as fine-grained environmental adaptation, is a type of phenotypic plasticity that involves the phenotypic change of organisms in response to changes in the environment over time. Animals can respond to short-term environmental changes with physiological (reversible) and behavioral changes; plants, which are sedentary, respond to short-term environmental changes with both physiological and developmental (non-reversible) changes.
Temporal plasticity takes place over a time scale of minutes, days, or seasons, and in environments that are both variable and predictable within the lifespan of an individual. Temporal plasticity is considered adaptive if the phenotypic response results in increased fitness. Non-reversible phenotypic changes can be observed in metameric organisms such as plants that depend on the environmental condition(s) each metamer was developed under. Under some circumstances early exposure to specific stressors can affect how an individual plant is capable of responding to future environmental changes (Metaplasticity).
Reversible plasticity
A reversible change is defined as one that is expressed in response to an environmental stressor but returns to a normal state after the stress is no longer present. Reversible changes are more likely to be adaptive for an organism when the stress driving the change is temporary and the organism is likely to be exposed to it again within its lifetime. Reversible plasticity often involves changes in physiology or behavior. Perennial plants, which often experience recurring stresses in their environment due to lack of mobility, benefit greatly from reversible physiological plasticity such as changes in resource uptake and allocation. When essential nutrients are low, root and leaf resorption rates can increase, persisting at a high rate until there are more nutrients available in the soil and resorption rates can return back to their normal state.
Irreversible plasticity
Irrevers
Document 4:::
Life history theory is an analytical framework designed to study the diversity of life history strategies used by different organisms throughout the world, as well as the causes and results of the variation in their life cycles. It is a theory of biological evolution that seeks to explain aspects of organisms' anatomy and behavior by reference to the way that their life histories—including their reproductive development and behaviors, post-reproductive behaviors, and lifespan (length of time alive)—have been shaped by natural selection. A life history strategy is the "age- and stage-specific patterns" and timing of events that make up an organism's life, such as birth, weaning, maturation, death, etc. These events, notably juvenile development, age of sexual maturity, first reproduction, number of offspring and level of parental investment, senescence and death, depend on the physical and ecological environment of the organism.
The theory was developed in the 1950s and is used to answer questions about topics such as organism size, age of maturation, number of offspring, life span, and many others. In order to study these topics, life history strategies must be identified, and then models are constructed to study their effects. Finally, predictions about the importance and role of the strategies are made, and these predictions are used to understand how evolution affects the ordering and length of life history events in an organism's life, particularly the lifespan and period of reproduction. Life history theory draws on an evolutionary foundation, and studies the effects of natural selection on organisms, both throughout their lifetime and across generations. It also uses measures of evolutionary fitness to determine if organisms are able to maximize or optimize this fitness, by allocating resources to a range of different demands throughout the organism's life. It serves as a method to investigate further the "many layers of complexity of organisms and their worl
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is the process by which organisms change in small ways over time?
A. microevolution
B. adaptation
C. spontaneous mutation
D. natural selection
Answer:
|
|
sciq-2046
|
multiple_choice
|
What is an electrically neutral compound with positive and negative ions?
|
[
"isotope",
"neutralized molecule",
"ionic compound",
"correlated compound"
] |
C
|
Relavent Documents:
Document 0:::
An ion () is an atom or molecule with a net electrical charge. The charge of an electron is considered to be negative by convention and this charge is equal and opposite to the charge of a proton, which is considered to be positive by convention. The net charge of an ion is not zero because its total number of electrons is unequal to its total number of protons.
A cation is a positively charged ion with fewer electrons than protons while an anion is a negatively charged ion with more electrons than protons. Opposite electric charges are pulled towards one another by electrostatic force, so cations and anions attract each other and readily form ionic compounds.
Ions consisting of only a single atom are termed atomic or monatomic ions, while two or more atoms form molecular ions or polyatomic ions. In the case of physical ionization in a fluid (gas or liquid), "ion pairs" are created by spontaneous molecule collisions, where each generated pair consists of a free electron and a positive ion. Ions are also created by chemical interactions, such as the dissolution of a salt in liquids, or by other means, such as passing a direct current through a conducting solution, dissolving an anode via ionization.
History of discovery
The word ion was coined from Greek neuter present participle of ienai (), meaning "to go". A cation is something that moves down ( pronounced kato, meaning "down") and an anion is something that moves up (, meaning "up"). They are so called because ions move toward the electrode of opposite charge. This term was introduced (after a suggestion by the English polymath William Whewell) by English physicist and chemist Michael Faraday in 1834 for the then-unknown species that goes from one electrode to the other through an aqueous medium. Faraday did not know the nature of these species, but he knew that since metals dissolved into and entered a solution at one electrode and new metal came forth from a solution at the other electrode; that some kind of
Document 1:::
Advanced Placement (AP) Physics C: Electricity and Magnetism (also known as AP Physics C: E&M or AP E&M) is an introductory physics course administered by the College Board as part of its Advanced Placement program. It is intended to proxy a second-semester calculus-based university course in electricity and magnetism. The content of Physics C: E&M overlaps with that of AP Physics 2, but Physics 2 is algebra-based and covers other topics outside of electromagnetism, while Physics C is calculus-based and only covers electromagnetism. Physics C: E&M may be combined with its mechanics counterpart to form a year-long course that prepares for both exams.
Course content
E&M is equivalent to an introductory college course in electricity and magnetism for physics or engineering majors. The course modules are:
Electrostatics
Conductors, capacitors, and dielectrics
Electric circuits
Magnetic fields
Electromagnetism.
Methods of calculus are used wherever appropriate in formulating physical principles and in applying them to physical problems. Therefore, students should have completed or be concurrently enrolled in a calculus class.
AP test
The course culminates in an optional exam for which high-performing students may receive some credit towards their college coursework, depending on the institution.
Registration
The AP examination for AP Physics C: Electricity and Magnetism is separate from the AP examination for AP Physics C: Mechanics. Before 2006, test-takers paid only once and were given the choice of taking either one or two parts of the Physics C test.
Format
The exam is typically administered on a Monday afternoon in May. The exam is configured in two categories: a 35-question multiple choice section and a 3-question free response section. Test takers are allowed to use an approved calculator during the entire exam. The test is weighted such that each section is worth half of the final score. This and AP Physics C: Mechanics are the shortest AP exams, with
Document 2:::
An electrolyte is a medium containing ions that is electrically conducting through the movement of those ions, but not conducting electrons. This includes most soluble salts, acids, and bases dissolved in a polar solvent, such as water. Upon dissolving, the substance separates into cations and anions, which disperse uniformly throughout the solvent. Solid-state electrolytes also exist. In medicine and sometimes in chemistry, the term electrolyte refers to the substance that is dissolved.
Electrically, such a solution is neutral. If an electric potential is applied to such a solution, the cations of the solution are drawn to the electrode that has an abundance of electrons, while the anions are drawn to the electrode that has a deficit of electrons. The movement of anions and cations in opposite directions within the solution amounts to a current. Some gases, such as hydrogen chloride (HCl), under conditions of high temperature or low pressure can also function as electrolytes. Electrolyte solutions can also result from the dissolution of some biological (e.g., DNA, polypeptides) or synthetic polymers (e.g., polystyrene sulfonate), termed "polyelectrolytes", which contain charged functional groups. A substance that dissociates into ions in solution or in the melt acquires the capacity to conduct electricity. Sodium, potassium, chloride, calcium, magnesium, and phosphate in a liquid phase are examples of electrolytes.
In medicine, electrolyte replacement is needed when a person has prolonged vomiting or diarrhea, and as a response to sweating due to strenuous athletic activity. Commercial electrolyte solutions are available, particularly for sick children (such as oral rehydration solution, Suero Oral, or Pedialyte) and athletes (sports drinks). Electrolyte monitoring is important in the treatment of anorexia and bulimia.
In science, electrolytes are one of the main components of electrochemical cells.
In clinical medicine, mentions of electrolytes usually refer m
Document 3:::
In chemistry, an ionophore () is a chemical species that reversibly binds ions. Many ionophores are lipid-soluble entities that transport ions across the cell membrane. Ionophores catalyze ion transport across hydrophobic membranes, such as liquid polymeric membranes (carrier-based ion selective electrodes) or lipid bilayers found in the living cells or synthetic vesicles (liposomes). Structurally, an ionophore contains a hydrophilic center and a hydrophobic portion that interacts with the membrane.
Some ionophores are synthesized by microorganisms to import ions into their cells. Synthetic ion carriers have also been prepared. Ionophores selective for cations and anions have found many applications in analysis. These compounds have also shown to have various biological effects and a synergistic effect when combined with the ion they bind.
Classification
Biological activities of metal ion-binding compounds can be changed in response to the increment of the metal concentration, and based on the latter compounds can be classified as "metal ionophores", "metal chelators" or "metal shuttles". If the biological effect is augmented by increasing the metal concentration, it is classified as a "metal ionophore". If the biological effect is decreased or reversed by increasing the metal concentration, it is classified as a "metal chelator". If the biological effect is not affected by increasing the metal concentration, and the compound-metal complex enters the cell, it is classified as a "metal shuttle". The term ionophore (from Greek ion carrier or ion bearer) was proposed by Berton Pressman in 1967 when he and his colleagues were investigating the antibiotic mechanisms of valinomycin and nigericin.
Many ionophores are produced naturally by a variety of microbes, fungi and plants, and act as a defense against competing or pathogenic species. Multiple synthetic membrane-spanning ionophores have also been synthesized.
The two broad classifications of ionophores synthesiz
Document 4:::
The degree of ionization (also known as ionization yield in the literature) refers to the proportion of neutral particles, such as those in a gas or aqueous solution, that are ionized. For electrolytes, it could be understood as a capacity of acid/base to ionize itself. A low degree of ionization is sometimes called partially ionized (also weakly ionized), and a high degree of ionization as fully ionized. However, fully ionized can also mean that an ion has no electrons left.
Ionization refers to the process whereby an atom or molecule loses one or several electrons from its atomic orbital, or conversely gains an additional one, from an incoming free electron (electron attachment). In both cases, the atom or molecule ceases to be a neutral particle and becomes a charge carrier. If the species has lost one or several electrons, it becomes positively charged and is called a positive ion, or cation. On the contrary, if the species has gained one or several additional electrons, it becomes negatively charged and is called a negative ion, or anion. Individual free electrons and ions in a plasma have very short lives typically inferior to the microsecond, as ionization and recombination, excitation and relaxation are collective continuous processes.
Chemistry usage
The degree of dissociation α (also known as degree of ionization), is a way of representing the strength of an acid. It is defined as the ratio of the number of ionized molecules and the number of molecules dissolved in water. It can be represented as a decimal number or as a percentage. One can classify strong acids as those having ionization degrees above 30%, weak acids as those with α below 30%, and the rest as moderate acids, at a specified molar concentration.
Physics usage
In plasma, the degree of ionization refers to the proportion of neutral particles that are ionized:
where is the ion density and the neutral density (in particles per cubic meter). It is a dimensionless number, sometimes expres
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is an electrically neutral compound with positive and negative ions?
A. isotope
B. neutralized molecule
C. ionic compound
D. correlated compound
Answer:
|
|
ai2_arc-303
|
multiple_choice
|
Where is the smallest negatively-charged particle found in an atom?
|
[
"nucleus",
"orbits in the nucleus",
"orbits around the nucleus",
"between the proton and neutron"
] |
C
|
Relavent Documents:
Document 0:::
The objective of the Thomson problem is to determine the minimum electrostatic potential energy configuration of electrons constrained to the surface of a unit sphere that repel each other with a force given by Coulomb's law. The physicist J. J. Thomson posed the problem in 1904 after proposing an atomic model, later called the plum pudding model, based on his knowledge of the existence of negatively charged electrons within neutrally-charged atoms.
Related problems include the study of the geometry of the minimum energy configuration and the study of the large behavior of the minimum energy.
Mathematical statement
The electrostatic interaction energy occurring between each pair of electrons of equal charges (, with the elementary charge of an electron) is given by Coulomb's law,
where is the electric constant and is the distance between each pair of electrons located at points on the sphere defined by vectors and , respectively.
Simplified units of and (the Coulomb constant) are used without loss of generality. Then,
The total electrostatic potential energy of each N-electron configuration may then be expressed as the sum of all pair-wise interaction energies
The global minimization of over all possible configurations of N distinct points is typically found by numerical minimization algorithms.
Thomson's problem is related to the 7th of the eighteen unsolved mathematics problems proposed by the mathematician Steve Smale — "Distribution of points on the 2-sphere".
The main difference is that in Smale's problem the function to minimise is not the electrostatic potential but a logarithmic potential given by A second difference is that Smale's question is about the asymptotic behaviour of the total potential when the number N of points goes to infinity, not for concrete values of N.
Example
The solution of the Thomson problem for two electrons is obtained when both electrons are as far apart as possible on opposite sides of the origin, , or
K
Document 1:::
The electric dipole moment is a measure of the separation of positive and negative electrical charges within a system, that is, a measure of the system's overall polarity. The SI unit for electric dipole moment is the coulomb-meter (C⋅m). The debye (D) is another unit of measurement used in atomic physics and chemistry.
Theoretically, an electric dipole is defined by the first-order term of the multipole expansion; it consists of two equal and opposite charges that are infinitesimally close together, although real dipoles have separated charge.
Elementary definition
Often in physics the dimensions of a massive object can be ignored and can be treated as a pointlike object, i.e. a point particle. Point particles with electric charge are referred to as point charges. Two point charges, one with charge and the other one with charge separated by a distance , constitute an electric dipole (a simple case of an electric multipole). For this case, the electric dipole moment has a magnitude and is directed from the negative charge to the positive one. Some authors may split in half and use since this quantity is the distance between either charge and the center of the dipole, leading to a factor of two in the definition.
A stronger mathematical definition is to use vector algebra, since a quantity with magnitude and direction, like the dipole moment of two point charges, can be expressed in vector form where is the displacement vector pointing from the negative charge to the positive charge. The electric dipole moment vector also points from the negative charge to the positive charge. With this definition the dipole direction tends to align itself with an external electric field (and note that the electric flux lines produced by the charges of the dipole itself, which point from positive charge to negative charge then tend to oppose the flux lines of the external field). Note that this sign convention is used in physics, while the opposite sign convention for th
Document 2:::
The subatomic scale is the domain of physical size that encompasses objects smaller than an atom. It is the scale at which the atomic constituents, such as the nucleus containing protons and neutrons, and the electrons in their orbitals, become apparent.
The subatomic scale includes the many thousands of times smaller subnuclear scale, which is the scale of physical size at which constituents of the protons and neutrons - particularly quarks - become apparent.
See also
Astronomical scale the opposite end of the spectrum
Subatomic particles
Document 3:::
Secondary electrons are electrons generated as ionization products. They are called 'secondary' because they are generated by other radiation (the primary radiation). This radiation can be in the form of ions, electrons, or photons with sufficiently high energy, i.e. exceeding the ionization potential. Photoelectrons can be considered an example of secondary electrons where the primary radiation are photons; in some discussions photoelectrons with higher energy (>50 eV) are still considered "primary" while the electrons freed by the photoelectrons are "secondary".
Applications
Secondary electrons are also the main means of viewing images in the scanning electron microscope (SEM). The range of secondary electrons depends on the energy. Plotting the inelastic mean free path as a function of energy often shows characteristics of the "universal curve" familiar to electron spectroscopists and surface analysts. This distance is on the order of a few nanometers in metals and tens of nanometers in insulators. This small distance allows such fine resolution to be achieved in the SEM.
For SiO2, for a primary electron energy of 100 eV, the secondary electron range is up to 20 nm from the point of incidence.
See also
Delta ray
Everhart-Thornley detector
Document 4:::
An atom is a particle that consists of a nucleus of protons and neutrons surrounded by an electromagnetically-bound cloud of electrons. The atom is the basic particle of the chemical elements, and the chemical elements are distinguished from each other by the number of protons that are in their atoms. For example, any atom that contains 11 protons is sodium, and any atom that contains 29 protons is copper. The number of neutrons defines the isotope of the element.
Atoms are extremely small, typically around 100 picometers across. A human hair is about a million carbon atoms wide. This is smaller than the shortest wavelength of visible light, which means humans cannot see atoms with conventional microscopes. Atoms are so small that accurately predicting their behavior using classical physics is not possible due to quantum effects.
More than 99.94% of an atom's mass is in the nucleus. Each proton has a positive electric charge, while each electron has a negative charge, and the neutrons, if any are present, have no electric charge. If the numbers of protons and electrons are equal, as they normally are, then the atom is electrically neutral. If an atom has more electrons than protons, then it has an overall negative charge, and is called a negative ion (or anion). Conversely, if it has more protons than electrons, it has a positive charge, and is called a positive ion (or cation).
The electrons of an atom are attracted to the protons in an atomic nucleus by the electromagnetic force. The protons and neutrons in the nucleus are attracted to each other by the nuclear force. This force is usually stronger than the electromagnetic force that repels the positively charged protons from one another. Under certain circumstances, the repelling electromagnetic force becomes stronger than the nuclear force. In this case, the nucleus splits and leaves behind different elements. This is a form of nuclear decay.
Atoms can attach to one or more other atoms by chemical bonds to
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Where is the smallest negatively-charged particle found in an atom?
A. nucleus
B. orbits in the nucleus
C. orbits around the nucleus
D. between the proton and neutron
Answer:
|
|
sciq-2258
|
multiple_choice
|
Magnetic force is always perpendicular to what?
|
[
"sunlight",
"length",
"velocity",
"speed"
] |
C
|
Relavent Documents:
Document 0:::
Advanced Placement (AP) Physics C: Electricity and Magnetism (also known as AP Physics C: E&M or AP E&M) is an introductory physics course administered by the College Board as part of its Advanced Placement program. It is intended to proxy a second-semester calculus-based university course in electricity and magnetism. The content of Physics C: E&M overlaps with that of AP Physics 2, but Physics 2 is algebra-based and covers other topics outside of electromagnetism, while Physics C is calculus-based and only covers electromagnetism. Physics C: E&M may be combined with its mechanics counterpart to form a year-long course that prepares for both exams.
Course content
E&M is equivalent to an introductory college course in electricity and magnetism for physics or engineering majors. The course modules are:
Electrostatics
Conductors, capacitors, and dielectrics
Electric circuits
Magnetic fields
Electromagnetism.
Methods of calculus are used wherever appropriate in formulating physical principles and in applying them to physical problems. Therefore, students should have completed or be concurrently enrolled in a calculus class.
AP test
The course culminates in an optional exam for which high-performing students may receive some credit towards their college coursework, depending on the institution.
Registration
The AP examination for AP Physics C: Electricity and Magnetism is separate from the AP examination for AP Physics C: Mechanics. Before 2006, test-takers paid only once and were given the choice of taking either one or two parts of the Physics C test.
Format
The exam is typically administered on a Monday afternoon in May. The exam is configured in two categories: a 35-question multiple choice section and a 3-question free response section. Test takers are allowed to use an approved calculator during the entire exam. The test is weighted such that each section is worth half of the final score. This and AP Physics C: Mechanics are the shortest AP exams, with
Document 1:::
In physics, the magnetomotive force (abbreviated mmf or MMF, symbol ) is a quantity appearing in the equation for the magnetic flux in a magnetic circuit, Hopkinson's law. It is the property of certain substances or phenomena that give rise to magnetic fields:
where is the magnetic flux and is the reluctance of the circuit. It can be seen that the magnetomotive force plays a role in this equation analogous to the voltage in Ohm's law, , since it is the cause of magnetic flux in a magnetic circuit:
where is the number of turns in the coil and is the electric current through the circuit.
where is the magnetic flux and is the magnetic reluctance
where is the magnetizing force (the strength of the magnetizing field) and is the mean length of a solenoid or the circumference of a toroid.
Units
The SI unit of mmf is the ampere, the same as the unit of current (analogously the units of emf and voltage are both the volt). Informally, and frequently, this unit is stated as the ampere-turn to avoid confusion with current. This was the unit name in the MKS system. Occasionally, the cgs system unit of the gilbert may also be encountered.
History
The term magnetomotive force was coined by Henry Augustus Rowland in 1880. Rowland intended this to indicate a direct analogy with electromotive force. The idea of a magnetic analogy to electromotive force can be found much earlier in the work of Michael Faraday (1791–1867) and it is hinted at by James Clerk Maxwell (1831–1879). However, Rowland coined the term and was the first to make explicit an Ohm's law for magnetic circuits in 1873.
Ohm's law for magnetic circuits is sometimes referred to as Hopkinson's law rather than Rowland's law as some authors attribute the law to John Hopkinson instead of Rowland. According to a review of magnetic circuit analysis methods this is an incorrect attribution originating from an 1885 paper by Hopkinson. Furthermore, Hopkinson actually cites Rowland's 1873 paper in th
Document 2:::
Advanced Placement (AP) Physics C: Mechanics (also known as AP Mechanics) is an introductory physics course administered by the College Board as part of its Advanced Placement program. It is intended to proxy a one-semester calculus-based university course in mechanics. The content of Physics C: Mechanics overlaps with that of AP Physics 1, but Physics 1 is algebra-based, while Physics C is calculus-based. Physics C: Mechanics may be combined with its electricity and magnetism counterpart to form a year-long course that prepares for both exams.
Course content
Intended to be equivalent to an introductory college course in mechanics for physics or engineering majors, the course modules are:
Kinematics
Newton's laws of motion
Work, energy and power
Systems of particles and linear momentum
Circular motion and rotation
Oscillations and gravitation.
Methods of calculus are used wherever appropriate in formulating physical principles and in applying them to physical problems. Therefore, students should have completed or be concurrently enrolled in a Calculus I class.
This course is often compared to AP Physics 1: Algebra Based for its similar course material involving kinematics, work, motion, forces, rotation, and oscillations. However, AP Physics 1: Algebra Based lacks concepts found in Calculus I, like derivatives or integrals.
This course may be combined with AP Physics C: Electricity and Magnetism to make a unified Physics C course that prepares for both exams.
AP test
The course culminates in an optional exam for which high-performing students may receive some credit towards their college coursework, depending on the institution.
Registration
The AP examination for AP Physics C: Mechanics is separate from the AP examination for AP Physics C: Electricity and Magnetism. Before 2006, test-takers paid only once and were given the choice of taking either one or two parts of the Physics C test.
Format
The exam is typically administered on a Monday aftern
Document 3:::
There are four Advanced Placement (AP) Physics courses administered by the College Board as part of its Advanced Placement program: the algebra-based Physics 1 and Physics 2 and the calculus-based Physics C: Mechanics and Physics C: Electricity and Magnetism. All are intended to be at the college level. Each AP Physics course has an exam for which high-performing students may receive credit toward their college coursework.
AP Physics 1 and 2
AP Physics 1 and AP Physics 2 were introduced in 2015, replacing AP Physics B. The courses were designed to emphasize critical thinking and reasoning as well as learning through inquiry. They are algebra-based and do not require any calculus knowledge.
AP Physics 1
AP Physics 1 covers Newtonian mechanics, including:
Unit 1: Kinematics
Unit 2: Dynamics
Unit 3: Circular Motion and Gravitation
Unit 4: Energy
Unit 5: Momentum
Unit 6: Simple Harmonic Motion
Unit 7: Torque and Rotational Motion
Until 2020, the course also covered topics in electricity (including Coulomb's Law and resistive DC circuits), mechanical waves, and sound. These units were removed because they are included in AP Physics 2.
AP Physics 2
AP Physics 2 covers the following topics:
Unit 1: Fluids
Unit 2: Thermodynamics
Unit 3: Electric Force, Field, and Potential
Unit 4: Electric Circuits
Unit 5: Magnetism and Electromagnetic Induction
Unit 6: Geometric and Physical Optics
Unit 7: Quantum, Atomic, and Nuclear Physics
AP Physics C
From 1969 to 1972, AP Physics C was a single course with a single exam that covered all standard introductory university physics topics, including mechanics, fluids, electricity and magnetism, optics, and modern physics. In 1973, the College Board split the course into AP Physics C: Mechanics and AP Physics C: Electricity and Magnetism. The exam was also split into two separate 90-minute tests, each equivalent to a semester-length calculus-based college course. Until 2006, both exams could be taken for a single
Document 4:::
Magnetic deviation is the error induced in a compass by local magnetic fields, which must be allowed for, along with magnetic declination, if accurate bearings are to be calculated. (More loosely, "magnetic deviation" is used by some to mean the same as "magnetic declination". This article is about the former meaning.)
Compass readings
Compasses are used to determine the direction of true North. However, the compass reading must be corrected for two effects. The first is magnetic declination or variation—the angular difference between magnetic North (the local direction of the Earth's magnetic field) and true North. The second is magnetic deviation—the angular difference between magnetic North and the compass needle due to nearby sources of interference such as magnetically permeable bodies, or other magnetic fields within the field of influence.
Sources
In navigation manuals, magnetic deviation refers specifically to compass error caused by magnetized iron within a ship or aircraft. This iron has a mixture of permanent magnetization and an induced (temporary) magnetization that is induced by the Earth's magnetic field. Because the latter depends on the orientation of the craft relative to the Earth's field, it can be difficult to analyze and correct for it.
The deviation errors caused by magnetism in the ship's structure are minimised by precisely positioning small magnets and iron compensators close to the compass. To compensate for the induced magnetization, two magnetically soft iron spheres are placed on side arms. However, because the magnetic "signature" of every ship changes slowly with location, and with time, it is necessary to adjust the compensating magnets, periodically, to keep the deviation errors to a practical minimum. Magnetic compass adjustment and correction is one of the subjects in the examination curriculum for a shipmaster's certificate of competency.
The sources of magnetic deviation vary from compass to compass or vehicle to vehicle. H
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Magnetic force is always perpendicular to what?
A. sunlight
B. length
C. velocity
D. speed
Answer:
|
|
sciq-104
|
multiple_choice
|
Dialysis is a treatment for failure of what organs?
|
[
"liver",
"kidneys",
"spleen",
"lungs"
] |
B
|
Relavent Documents:
Document 0:::
Artificial kidney is often a synonym for hemodialysis, but may also refer to the other renal replacement therapies (with exclusion of kidney transplantation) that are in use and/or in development. This article deals mainly with bioengineered kidneys/bioartificial kidneys that are grown from renal cell lines/renal tissue.
The first successful artificial kidney was developed by Willem Kolff in the Netherlands during the early 1940s: Kolff was the first to construct a working dialyzer in 1943.
Medical uses
Kidney failure
Kidneys are paired vital organs located behind the abdominal cavity at the bottom of the ribcage corresponding to the levels T12-L3 of the spine vertebrae. They perform about a dozen physiologic functions and are fairly easily damaged. Some of these functions include filtration and excretion of metabolic waste products, regulation of necessary electrolytes and fluids and stimulation of red blood cell-production. These organs routinely filter about 100 to 140 liters of blood a day to produce 1 to 2 liters of urine, composed of wastes and excess fluid.
Kidney failure results in the slow accumulation of nitrogenous wastes, salts, water, and disruption of the body's normal pH balance. This failure typically occurs over a long period of time, and when the patient's renal function declines enough over the course of the disease, is commonly known as end stage renal disease (ESRD; which is also known as Level 5 or 6 kidney disease, depending on whether dialysis or renal replacement therapy is used). Detecting kidney disease before the kidneys start to shut down is uncommon, with high blood pressure and decreased appetite being symptoms that indicate a problem. Diabetes and high blood pressure are seen as the 2 most common causes of kidney failure. Experts predict that the demand for dialysis will increase as the prevalence of diabetes increases. Until the Second World War, kidney failure generally meant death for the patient. Several insights into kidney f
Document 1:::
An artificial organ is a human made organ device or tissue that is implanted or integrated into a human — interfacing with living tissue — to replace a natural organ, to duplicate or augment a specific function or functions so the patient may return to a normal life as soon as possible. The replaced function does not have to be related to life support, but it often is. For example, replacement bones and joints, such as those found in hip replacements, could also be considered artificial organs.
Implied by definition, is that the device must not be continuously tethered to a stationary power supply or other stationary resources such as filters or chemical processing units. (Periodic rapid recharging of batteries, refilling of chemicals, and/or cleaning/replacing of filters would exclude a device from being called an artificial organ.) Thus, a dialysis machine, while a very successful and critically important life support device that almost completely replaces the duties of a kidney, is not an artificial organ.
Purpose
Constructing and installing artificial organs, an extremely research-intensive and expensive process initially, may entail many years of ongoing maintenance services not needed by a natural organ:
providing life support to prevent imminent death while awaiting a transplant (e.g. artificial heart);
dramatically improving the patient's ability for self care (e.g. artificial limb);
improving the patient's ability to interact socially (e.g. cochlear implant); or
improving a patient's quality of life through cosmetic restoration after cancer surgery or an accident.
The use of any artificial organ by humans is almost always preceded by extensive experiments with animals. Initial testing in humans is frequently limited to those either already facing death or who have exhausted every other treatment possibility.
Examples
Artificial limbs
Artificial arms and legs, or prosthetics, are intended to restore a degree of normal function to amputees. Mechan
Document 2:::
Renal pathology is a subspecialty of anatomic pathology that deals with the diagnosis and characterization of medical diseases (non-tumor) of the kidneys. In the academic setting, renal pathologists work closely with nephrologists and transplant surgeons, who typically obtain diagnostic specimens via percutaneous renal biopsy. The renal pathologist must synthesize findings from light microscopy, electron microscopy, and immunofluorescence to obtain a definitive diagnosis. Medical renal diseases may affect the glomerulus, the tubules and interstitium, the vessels, or a combination of these compartments.
External links
http://www.renalpathsoc.org/
Renal Pathology Tutorial written by J. Charles Jennette
Pathologist Guide
Anatomical pathology
Document 3:::
The excretory system is a passive biological system that removes excess, unnecessary materials from the body fluids of an organism, so as to help maintain internal chemical homeostasis and prevent damage to the body. The dual function of excretory systems is the elimination of the waste products of metabolism and to drain the body of used up and broken down components in a liquid and gaseous state. In humans and other amniotes (mammals, birds and reptiles) most of these substances leave the body as urine and to some degree exhalation, mammals also expel them through sweating.
Only the organs specifically used for the excretion are considered a part of the excretory system. In the narrow sense, the term refers to the urinary system. However, as excretion involves several functions that are only superficially related, it is not usually used in more formal classifications of anatomy or function.
As most healthy functioning organs produce metabolic and other wastes, the entire organism depends on the function of the system. Breaking down of one of more of the systems is a serious health condition, for example kidney failure.
Systems
Urinary system
The kidneys are large, bean-shaped organs which are present on each side of the vertebral column in the abdominal cavity. Humans have two kidneys and each kidney is supplied with blood from the renal artery. The kidneys remove from the blood the nitrogenous wastes such as urea, as well as salts and excess water, and excrete them in the form of urine. This is done with the help of millions of nephrons present in the kidney. The filtrated blood is carried away from the kidneys by the renal vein (or kidney vein). The urine from the kidney is collected by the ureter (or excretory tubes), one from each kidney, and is passed to the urinary bladder. The urinary bladder collects and stores the urine until urination. The urine collected in the bladder is passed into the external environment from the body through an opening called
Document 4:::
Assessment of kidney function occurs in different ways, using the presence of symptoms and signs, as well as measurements using urine tests, blood tests, and medical imaging.
Functions of a healthy kidney include maintaining a person's fluid balance, maintaining an acid-base balance; regulating electrolytes including sodium, potassium, and other electrolytes; clearing toxins; regulating blood pressure; and regulating hormones, such as erythropoietin; and activation of vitamin D.
Description
The functions of the kidney include maintenance of acid-base balance; regulation of fluid balance; regulation of sodium, potassium, and other electrolytes; clearance of toxins; absorption of glucose, amino acids, and other small molecules; regulation of blood pressure; production of various hormones, such as erythropoietin; and activation of vitamin D.
The GFR is regarded as the best overall measure of the kidney's ability to carry out these numerous functions. An estimate of the GFR is used clinically to determine the degree of kidney impairment and to track the progression of the disease. The GFR, however, does not reveal the source of the kidney disease. This is accomplished by urinalysis, measurement of urine protein excretion, kidney imaging, and, if necessary, kidney biopsy.
Much of renal physiology is studied at the level of the nephron the smallest functional unit of the kidney. Each nephron begins with a filtration component that filters the blood entering the kidney. This filtrate then flows along the length of the nephron, which is a tubular structure lined by a single layer of specialized cells and surrounded by capillaries. The major functions of these lining cells are the reabsorption of water and small molecules from the filtrate into the blood, and the secretion of wastes from the blood into the urine.
Proper function of the kidney requires that it receives and adequately filters blood. This is performed at the microscopic level by many hundreds of thousa
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Dialysis is a treatment for failure of what organs?
A. liver
B. kidneys
C. spleen
D. lungs
Answer:
|
|
sciq-8191
|
multiple_choice
|
Each silicon atom shares how many valence electrons with other silicon atoms in a crystal?
|
[
"four",
"three",
"five",
"ten"
] |
A
|
Relavent Documents:
Document 0:::
A silicon–oxygen bond ( bond) is a chemical bond between silicon and oxygen atoms that can be found in many inorganic and organic compounds. In a silicon–oxygen bond, electrons are shared unequally between the two atoms, with oxygen taking the larger share due to its greater electronegativity. This polarisation means Si–O bonds show characteristics of both covalent and ionic bonds. Compounds containing silicon–oxygen bonds include materials of major geological and industrial significance such as silica, silicate minerals and silicone polymers like polydimethylsiloxane.
Bond polarity, length and strength
On the Pauling electronegativity scale, silicon has an electronegativity of 1.90 and oxygen 3.44. The electronegativity difference between the elements is therefore 1.54. Because of this moderately large difference in electronegativities, the bond is polar but not fully ionic. Carbon has an electronegativity of 2.55 so carbon–oxygen bonds have an electronegativity difference of 0.89 and are less polar than silicon–oxygen bonds. Silicon–oxygen bonds are therefore covalent and polar, with a partial positive charge on silicon and a partial negative charge on oxygen: Siδ+—Oδ−.
Silicon–oxygen single bonds are longer (1.6 vs 1.4 Å) but stronger (452 vs. about 360 kJ mol−1) than carbon–oxygen single bonds. However, silicon–oxygen double bonds are weaker than carbon–oxygen double bonds (590 vs. 715 kJ mol−1) due to a better overlap of p orbitals forming a stronger pi bond in the latter. This is an example of the double bond rule. For these reasons, carbon dioxide is a molecular gas containing two C=O double bonds per carbon atom whereas silicon dioxide is a polymeric solid containing four Si–O single bonds per silicon atom; molecular SiO2 containing two Si=O double bonds would polymerise. Other compounds containing Si=O double bonds are normally very reactive and unstable with respect to polymerisation or oligomerization. Silanones oligomerise to siloxanes unless they ar
Document 1:::
Steudel R 2020, Chemistry of the Non-metals: Syntheses - Structures - Bonding - Applications, in collaboration with D Scheschkewitz, Berlin, Walter de Gruyter, . ▲
An updated translation of the 5th German edition of 2013, incorporating the literature up to Spring 2019. Twenty-three nonmetals, including B, Si, Ge, As, Se, Te, and At but not Sb (nor Po). The nonmetals are identified on the basis of their electrical conductivity at absolute zero putatively being close to zero, rather than finite as in the case of metals. That does not work for As however, which has the electronic structure of a semimetal (like Sb).
Halka M & Nordstrom B 2010, "Nonmetals", Facts on File, New York,
A reading level 9+ book covering H, C, N, O, P, S, Se. Complementary books by the same authors examine (a) the post-transition metals (Al, Ga, In, Tl, Sn, Pb and Bi) and metalloids (B, Si, Ge, As, Sb, Te and Po); and (b) the halogens and noble gases.
Woolins JD 1988, Non-Metal Rings, Cages and Clusters, John Wiley & Sons, Chichester, .
A more advanced text that covers H; B; C, Si, Ge; N, P, As, Sb; O, S, Se and Te.
Steudel R 1977, Chemistry of the Non-metals: With an Introduction to Atomic Structure and Chemical Bonding, English edition by FC Nachod & JJ Zuckerman, Berlin, Walter de Gruyter, . ▲
Twenty-four nonmetals, including B, Si, Ge, As, Se, Te, Po and At.
Powell P & Timms PL 1974, The Chemistry of the Non-metals, Chapman & Hall, London, . ▲
Twenty-two nonmetals including B, Si, Ge, As and Te. Tin and antimony are shown as being intermediate between metals and nonmetals; they are later shown as either metals or nonmetals. Astatine is counted as a metal.
Document 2:::
The "Centre d’Élaboration de Matériaux et d’Etudes Structurales" (CEMES) is a CNRS laboratory located in Toulouse, France.
CEMES is a public fundamental research laboratory specializing in solid state physics, nanosciences, molecular chemistry and materials science. Its activities cover a spectrum from synthesizing (nano)materials and molecular systems to study and modelling of their structures and physical properties (optical, mechanical, electronic and magnetic) and their integration in devices.
An important part of the laboratory’s experimental activity is studying and manipulating individual objects whose characteristic sizes are at the nanometric or atomic scales. Most of this experimental work uses state-of-the-art instrumentation supported by instrumental and methodological developments in the laboratory’s fields of transmission electron microscopy, near-field microscopy and optical spectroscopy. These research and development themes integrate modelling and theoretical studies carried out at different scales within the laboratory.
CEMES is a CNRS laboratory associated with Toulouse III - Paul Sabatier University and Institut National des Sciences Appliquées de Toulouse (INSA). Created in 1988, CEMES followed the previous Laboratoire d’Optique Electronique (LOE) created in 1957 by Prof. Gaston Dupouy.
In contact with the academic community, CEMES is involved in training given at the university at every level: Bachelor, Master and Doctorate.
CEMES currently hosts about 160 people. In 2022 this included: 40 full-time CNRS researchers, 26 University professors or assistant professors, 36 engineers, technicians and administrative staff, and 12 postdocs, 31 Ph.D. students, and many undergraduate students.
Objectives
Studying structures and properties of nanomaterials and nanostructures at the atomic scale
Establishing relationships between nano and microstructures and the physical properties of various kinds of materials and nanomaterials
Inventing and d
Document 3:::
Stannide ions,
Some examples of stannide Zintl ions are listed below. Some of them contain 2-centre 2-electron bonds (2c-2e), others are "electron deficient" and bonding sometimes can be described using polyhedral skeletal electron pair theory (Wade's rules) where the number of valence electrons contributed by each tin atom is considered to be 2 (the s electrons do not contribute). There are some examples of silicide and plumbide ions with similar structures, for example tetrahedral , the chain anion (Si2−)n, and .
Sn4− found for example in Mg2Sn.
, tetrahedral with 2c-2e bonds e.g. in CsSn.
, tetrahedral closo-cluster with 10 electrons (2n + 2).
(Sn2−)n zig-zag chain polymeric anion with 2c-2e bonds found for example in BaSn.
closo-
Document 4:::
Dr. Rudolf Maria "Ruud" Tromp (born 3 September 1954, Alkmaar) is a Dutch American scientist at IBM Research Division, Thomas J. Watson Research Center and a Physics Professor at Leiden University.
Education
He attended Petrus Canisius College The Lyceum (Alkmaar).
1982 Ph.D. in physics from the University of Utrecht (The Netherlands)
1982, completed a thesis on medium-energy ion scattering (MEIS) studies of the structure of silicon surfaces.
Biography
In 1983 he joined IBM at the Thomas J. Watson Research Center, where his scanning tunneling microscopy studies revealed the Si(001) dimer structure for the first time, as well as the spatial distribution of the Si(111) (7 × 7) electronic surface states and their relation to the underlying atomic structure. Using MEIS he co-invented “surfactant-mediated epitaxial growth,” a technique that allows much-improved control over the morphology of epitaxial films and superlattices.
More recently, his studies have focused on the dynamics of surface and interface processes such as phase transitions, chemisorption and etching, epitaxial growth, and aspects of nanotechnology. Ultrahigh-vacuum transmission electron microscopy and low-energy electron microscopy allow detailed, real-time, in situ observations of such processes with high spatial resolution. Using those techniques, the studies have shed new light on the thermodynamics of epitaxial growth, the dynamic evolution of the surface morphology of epitaxial films, the self-assembly of quantum dots, the spatiotemporal character of first- and second-order phase transitions at surfaces, etc.
He has also developed a novel Low Energy Electron Microscopy instrument, including energy filtering and aberration correction.
Awards
1981: Wayne B. Nottingham Prize of the Physical Electronics Conference
1995: Materials Research Society Medal
2003: American Physical Society Davisson-Germer Prize
2004: American Vacuum Society Medard W. Welch Award
2009: Materials Research Society (MR
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Each silicon atom shares how many valence electrons with other silicon atoms in a crystal?
A. four
B. three
C. five
D. ten
Answer:
|
|
sciq-5453
|
multiple_choice
|
Some animals prepare for the long winters by storing food and going dormant, a behavior called what?
|
[
"hibernation",
"gestation",
"resting",
"pollination"
] |
A
|
Relavent Documents:
Document 0:::
Dormancy is a period in an organism's life cycle when growth, development, and (in animals) physical activity are temporarily stopped. This minimizes metabolic activity and therefore helps an organism to conserve energy. Dormancy tends to be closely associated with environmental conditions. Organisms can synchronize entry to a dormant phase with their environment through predictive or consequential means. Predictive dormancy occurs when an organism enters a dormant phase before the onset of adverse conditions. For example, photoperiod and decreasing temperature are used by many plants to predict the onset of winter. Consequential dormancy occurs when organisms enter a dormant phase after adverse conditions have arisen. This is commonly found in areas with an unpredictable climate. While very sudden changes in conditions may lead to a high mortality rate among animals relying on consequential dormancy, its use can be advantageous, as organisms remain active longer and are therefore able to make greater use of available resources.
Animals
Hibernation
Hibernation is a mechanism used by many mammals to reduce energy expenditure and survive food shortages over the winter. Hibernation may be predictive or consequential. An animal prepares for hibernation by building up a thick layer of body fat during late summer and autumn that will provide it with energy during the dormant period. During hibernation, the animal undergoes many physiological changes, including decreased heart rate (by as much as 95%) and decreased body temperature. In addition to shivering, some hibernating animals also produce body heat by non-shivering thermogenesis to avoid freezing. Non-shivering thermogenesis is a regulated process in which the proton gradient generated by electron transport in mitochondria is used to produce heat instead of ATP in brown adipose tissue. Animals that hibernate include bats, ground squirrels and other rodents, mouse lemurs, the European hedgehog and other insectivo
Document 1:::
Heterothermy or heterothermia (from Greek ἕτερος heteros "other" and θέρμη thermē "heat") is a physiological term for animals that vary between self-regulating their body temperature, and allowing the surrounding environment to affect it. In other words, they exhibit characteristics of both poikilothermy and homeothermy.
Definition
Heterothermic animals are those that can switch between poikilothermic and homeothermic strategies. These changes in strategies typically occur on a daily basis or on an annual basis. More often than not, it is used as a way to dissociate the fluctuating metabolic rates seen in some small mammals and birds (e.g. bats and hummingbirds), from those of traditional cold blooded animals. In many bat species, body temperature and metabolic rate are elevated only during activity. When at rest, these animals reduce their metabolisms drastically, which results in their body temperature dropping to that of the surrounding environment. This makes them homeothermic when active, and poikilothermic when at rest. This phenomenon has been termed 'daily torpor' and was intensively studied in the Djungarian hamster. During the hibernation season, this animal shows strongly reduced metabolism each day during the rest phase while it reverts to endothermic metabolism during its active phase, leading to normal euthermic body temperatures (around 38 °C).
Larger mammals (e.g. ground squirrels) and bats show multi-day torpor bouts during hibernation (up to several weeks) in winter. During these multi-day torpor bouts, body temperature drops to ~1 °C above ambient temperature and metabolism may drop to about 1% of the normal endothermic metabolic rate. Even in these deep hibernators, the long periods of torpor is interrupted by bouts of endothermic metabolism, called arousals (typically lasting between 4–20 hours). These metabolic arousals cause body temperature to return to euthermic levels 35-37 °C. Most of the energy spent during hibernation is spent in arous
Document 2:::
Diurnality is a form of plant and animal behavior characterized by activity during daytime, with a period of sleeping or other inactivity at night. The common adjective used for daytime activity is "diurnal". The timing of activity by an animal depends on a variety of environmental factors such as the temperature, the ability to gather food by sight, the risk of predation, and the time of year. Diurnality is a cycle of activity within a 24-hour period; cyclic activities called circadian rhythms are endogenous cycles not dependent on external cues or environmental factors except for a zeitgeber. Animals active during twilight are crepuscular, those active during the night are nocturnal and animals active at sporadic times during both night and day are cathemeral.
Plants that open their flowers during the daytime are described as diurnal, while those that bloom during nighttime are nocturnal. The timing of flower opening is often related to the time at which preferred pollinators are foraging. For example, sunflowers open during the day to attract bees, whereas the night-blooming cereus opens at night to attract large sphinx moths.
In animals
Many types of animals are classified as being diurnal, meaning they are active during the day time and inactive or have periods of rest during the night time. Commonly classified diurnal animals include mammals, birds, and reptiles. Most primates are diurnal, including humans. Scientifically classifying diurnality within animals can be a challenge, apart from the obvious increased activity levels during the day time light.
Evolution of diurnality
Initially, most animals were diurnal, but adaptations that allowed some animals to become nocturnal is what helped contribute to the success of many, especially mammals. This evolutionary movement to nocturnality allowed them to better avoid predators and gain resources with less competition from other animals. This did come with some adaptations that mammals live with today. Visi
Document 3:::
Hoarding or caching in animal behavior is the storage of food in locations hidden from the sight of both conspecifics (animals of the same or closely related species) and members of other species. Most commonly, the function of hoarding or caching is to store food in times of surplus for times when food is less plentiful. However, there is evidence that some amount of caching or hoarding is done in order to ripen the food, called ripening caching. The term hoarding is most typically used for rodents, whereas caching is more commonly used in reference to birds, but the behaviors in both animal groups are quite similar.
Hoarding is done either on a long-term basiscached on a seasonal cycle, with food to be consumed months down the lineor on a short-term basis, in which case the food will be consumed over a period of one or several days.
Some common animals that cache their food are rodents such as hamsters and squirrels, and many different bird species, such as rooks and woodpeckers. The western scrub jay is noted for its particular skill at caching. There are two types of caching behavior: larder hoarding, where a species creates a few large caches which it often defends, and scatter hoarding, where a species will create multiple caches, often with each individual food item stored in a unique place. Both types of caching have their advantage.
Function
Caching behavior is typically a way to save excess edible food for later consumption—either soon to be eaten food, such as when a jaguar hangs partially eaten prey from a tree to be eaten within a few days, or long term, where the food is hidden and retrieved many months later. Caching is a common adaptation to seasonal changes in food availability. In regions where winters are harsh, food availability typically becomes low, and caching food during the times of high food availability in the warmer months provides a significant survival advantage. For species that hoard perishable food weather can significantly aff
Document 4:::
Aestivation ( (summer); also spelled estivation in American English) is a state of animal dormancy, similar to hibernation, although taking place in the summer rather than the winter. Aestivation is characterized by inactivity and a lowered metabolic rate, that is entered in response to high temperatures and arid conditions. It takes place during times of heat and dryness, which are often the summer months.
Invertebrate and vertebrate animals are known to enter this state to avoid damage from high temperatures and the risk of desiccation. Both terrestrial and aquatic animals undergo aestivation. Fossil records suggest that aestivation may have evolved several hundred million years ago.
Physiology
Organisms that aestivate appear to be in a fairly "light" state of dormancy, as their physiological state can be rapidly reversed, and the organism can quickly return to a normal state. A study done on Otala lactea, a snail native to parts of Europe and Northern Africa, shows that they can wake from their dormant state within ten minutes of being introduced to a wetter environment.
The primary physiological and biochemical concerns for an aestivating animal are to conserve energy, retain water in the body, ration the use of stored energy, handle the nitrogenous end products, and stabilize bodily organs, cells, and macromolecules. This can be quite a task as hot temperatures and arid conditions may last for months, in some cases for years. The depression of metabolic rate during aestivation causes a reduction in macromolecule synthesis and degradation. To stabilise the macromolecules, aestivators will enhance antioxidant defenses and elevate chaperone proteins. This is a widely used strategy across all forms of hypometabolism. These physiological and biochemical concerns appear to be the core elements of hypometabolism throughout the animal kingdom. In other words, animals which aestivate appear to go through nearly the same physiological processes as animals that hibern
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Some animals prepare for the long winters by storing food and going dormant, a behavior called what?
A. hibernation
B. gestation
C. resting
D. pollination
Answer:
|
|
sciq-10168
|
multiple_choice
|
What is the process by which plants make their own food?
|
[
"respiration",
"photosynthesis",
"mitochondrial process",
"ATP cycle"
] |
B
|
Relavent Documents:
Document 0:::
What a Plant Knows is a popular science book by Daniel Chamovitz, originally published in 2012, discussing the sensory system of plants. A revised edition was published in 2017.
Release details / Editions / Publication
Hardcover edition, 2012
Paperback version, 2013
Revised edition, 2017
What a Plant Knows has been translated and published in a number of languages.
Document 1:::
Maintenance respiration (or maintenance energy) refers to metabolism occurring in an organism that is needed to maintain that organism in a healthy, living state. Maintenance respiration contrasts with growth respiration, which is responsible for the synthesis of new structures in growth, nutrient uptake, nitrogen (N) reduction and phloem loading, whereas maintenance respiration is associated with protein and membrane turnover and maintenance of ion concentrations and gradients.
In plants
Maintenance respiration in plants refers to the amount of cellular respiration, measured by the carbon dioxide (CO2) released or oxygen (O2) consumed, during the generation of usable energy (mainly ATP, NADPH, and NADH) and metabolic intermediates used for (i) resynthesis of compounds that undergo renewal (turnover) in the normal process of metabolism (examples are enzymatic proteins, ribonucleic acids, and membrane lipids); (ii) maintenance of chemical gradients of ions and metabolites across cellular membranes that are necessary for cellular integrity and plant health; and (iii) operation of metabolic processes involved in physiological adjustment (i.e., acclimation) to a change in the plant's environment. The metabolic costs of the repair of injury from biotic or abiotic stress may also be considered a part of maintenance respiration.
Maintenance respiration is essential for biological health and growth of plants. It is estimated that about half of the respiration carried out by terrestrial plants during their lifetime is for the support of maintenance processes. Because typically more than half of global terrestrial plant photosynthesis (or gross primary production) is used for plant respiration, more than one quarter of global terrestrial plant photosynthesis is presumably consumed in maintenance respiration.
Maintenance respiration is a key component of most physiologically based mathematical models of plant growth, including models of crop growth and yield and models of
Document 2:::
Plants are the eukaryotes that form the kingdom Plantae; they are predominantly photosynthetic. This means that they obtain their energy from sunlight, using chloroplasts derived from endosymbiosis with cyanobacteria to produce sugars from carbon dioxide and water, using the green pigment chlorophyll. Exceptions are parasitic plants that have lost the genes for chlorophyll and photosynthesis, and obtain their energy from other plants or fungi.
Historically, as in Aristotle's biology, the plant kingdom encompassed all living things that were not animals, and included algae and fungi. Definitions have narrowed since then; current definitions exclude the fungi and some of the algae. By the definition used in this article, plants form the clade Viridiplantae (green plants), which consists of the green algae and the embryophytes or land plants (hornworts, liverworts, mosses, lycophytes, ferns, conifers and other gymnosperms, and flowering plants). A definition based on genomes includes the Viridiplantae, along with the red algae and the glaucophytes, in the clade Archaeplastida.
There are about 380,000 known species of plants, of which the majority, some 260,000, produce seeds. They range in size from single cells to the tallest trees. Green plants provide a substantial proportion of the world's molecular oxygen; the sugars they create supply the energy for most of Earth's ecosystems; other organisms, including animals, either consume plants directly or rely on organisms which do so.
Grain, fruit, and vegetables are basic human foods and have been domesticated for millennia. People use plants for many purposes, such as building materials, ornaments, writing materials, and, in great variety, for medicines. The scientific study of plants is known as botany, a branch of biology.
Definition
Taxonomic history
All living things were traditionally placed into one of two groups, plants and animals. This classification dates from Aristotle (384–322 BC), who distinguished d
Document 3:::
Plant Physiology is a monthly peer-reviewed scientific journal that covers research on physiology, biochemistry, cellular and molecular biology, genetics, biophysics, and environmental biology of plants. The journal has been published since 1926 by the American Society of Plant Biologists. The current editor-in-chief is Yunde Zhao (University of California San Diego. According to the Journal Citation Reports, the journal has a 2021 impact factor of 8.005.
Document 4:::
A plantoid is a robot or synthetic organism designed to look, act and grow like a plant. The concept was first scientifically published in 2010 (although models of comparable systems controlled by neural networks date back to 2003) and has so far remained largely theoretical. Plantoids imitate plants through appearances and mimicking behaviors and internal processes (which function to keep the plant alive or to ensure its survival). A prototype for the European Commission is now in development by сonsortium of the following scientists: Dario Floreano, Barbara Mazzolai, Josep Samitier, Stefano Mancuso.
A plantoid incorporates an inherently distributed architecture consisting of autonomous and specialized modules. Modules can be modeled on plant parts such as the root cap and communicate to form a simple swarm intelligence. This kind of system may display great robustness and resilience. It is conjectured to be capable of energy harvesting and management, collective environmental awareness and many other functions.
In science fiction, while human-like robots (androids) are fairly frequent and animal-like biomorphic robots turn up occasionally, plantoids are quite rare. Exceptions occur in the novel Hearts, Hands and Voices (1992, US: The Broken Land) by Ian McDonald and the TV series Jikuu Senshi Spielban.
Systems and Processes
Like plants, plantoids position its roots and appendages (projecting parts of the plantoid) towards beneficial conditions that stimulate growth (i.e sunlight, ideal temperatures, areas with larger water concentration) and away from factors that bar growth. This occurs through a combination of information from its sensors and the plantoid reacting accordingly.
Sensors
The use of soft tactical sensors (devices that gather information based on the surrounding physical environment) allows the plantoid to navigate its way through its environment. These sensors relay information to the plantoid and produce signals, similar to how a computer ca
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is the process by which plants make their own food?
A. respiration
B. photosynthesis
C. mitochondrial process
D. ATP cycle
Answer:
|
|
sciq-7339
|
multiple_choice
|
Wearing what will reduce the amplitude of sound waves entering the ears?
|
[
"hearing protectors",
"hardhat",
"megaphone",
"goggles"
] |
A
|
Relavent Documents:
Document 0:::
Hearing protector fit-testing, also known as field attenuation estimation system (FAES), measures the degree of noise attenuation when using the hearing protection device by a particular worker in the enterprise. Such measurements are necessary because noise attenuation is very variable among workers, and it can be close to zero for a significant proportion of them. Individual anatomical differences and varying ability to correctly apply HPD do not allow predicting noise attenuation in workers based on laboratory measurements with satisfactory accuracy. This is typically carried out using one of the available fit-testing hardware and software systems. The effectiveness is typically measured as a personal attenuation rating (PAR) which is subtracted from the known noise exposure to estimate the total noise exposure a single person has when wearing the tested hearing protection device (HPD). The Occupational Safety and Health Administration, National Hearing Conservation Association and National Institute for Occupational Safety and Health recommend it for all workers used HPD as a best practice, and describes existing testing methods and how to incorporate them in hearing conservation programs.
Hearing protection devices such as earplugs or earmuffs must be worn correctly for the wearer to be protected from noise. Correct use of hearing protection includes:
Choosing the most appropriate hearing protection device, both with appropriate level of attenuation and appropriate fit for the individual. Ideally, the device should limit the sound intensity that reaches the ear to levels below 85 dBA. If the attenuation does not limit the noise levels to that level, other alternatives should be sought. If the attenuation is greater than that, it can also interfere with the HPD use.
Wearing or inserting the hearing protection device correctly so it seals the wearer's ear canal, using the "roll-pull-hold" method for foam earplugs, and ensuring earmuffs create an unbroken s
Document 1:::
Acoustical engineering (also known as acoustic engineering) is the branch of engineering dealing with sound and vibration. It includes the application of acoustics, the science of sound and vibration, in technology. Acoustical engineers are typically concerned with the design, analysis and control of sound.
One goal of acoustical engineering can be the reduction of unwanted noise, which is referred to as noise control. Unwanted noise can have significant impacts on animal and human health and well-being, reduce attainment by students in schools, and cause hearing loss. Noise control principles are implemented into technology and design in a variety of ways, including control by redesigning sound sources, the design of noise barriers, sound absorbers, suppressors, and buffer zones, and the use of hearing protection (earmuffs or earplugs).
Besides noise control, acoustical engineering also covers positive uses of sound, such as the use of ultrasound in medicine, programming digital synthesizers, designing concert halls to enhance the sound of orchestras and specifying railway station sound systems so that announcements are intelligible.
Acoustic engineer (professional)
Acoustic engineers usually possess a bachelor's degree or higher qualification in acoustics, physics or another engineering discipline. Practicing as an acoustic engineer usually requires a bachelor's degree with significant scientific and mathematical content. Acoustic engineers might work in acoustic consultancy, specializing in particular fields, such as architectural acoustics, environmental noise or vibration control. In other industries, acoustic engineers might: design automobile sound systems; investigate human response to sounds, such as urban soundscapes and domestic appliances; develop audio signal processing software for mixing desks, and design loudspeakers and microphones for mobile phones. Acousticians are also involved in researching and understanding sound scientifically. Some pos
Document 2:::
A head shadow (or acoustic shadow) is a region of reduced amplitude of a sound because it is obstructed by the head. It is an example of diffraction.
Sound may have to travel through and around the head in order to reach an ear. The obstruction caused by the head can account for attenuation (reduced amplitude) of overall intensity as well as cause a filtering effect. The filtering effects of head shadowing are an essential element of sound localisation—the brain weighs the relative amplitude, timbre, and phase of a sound heard by the two ears and uses the difference to interpret directional information.
The shadowed ear, the ear further from the sound source, receives sound slightly later (up to approximately 0.7 ms later) than the unshadowed ear, and the timbre, or frequency spectrum, of the shadowed sound wave is different because of the obstruction of the head.
The head shadow causes particular difficulty in sound localisation in people suffering from unilateral hearing loss. It is a factor to consider when correcting hearing loss with directional hearing aids.
See also
Interaural intensity difference
Hearing
Document 3:::
An audio engineer (also known as a sound engineer or recording engineer) helps to produce a recording or a live performance, balancing and adjusting sound sources using equalization, dynamics processing and audio effects, mixing, reproduction, and reinforcement of sound. Audio engineers work on the "technical aspect of recording—the placing of microphones, pre-amp knobs, the setting of levels. The physical recording of any project is done by an engineer... the nuts and bolts."
Sound engineering is increasingly seen as a creative profession where musical instruments and technology are used to produce sound for film, radio, television, music and video games. Audio engineers also set up, sound check and do live sound mixing using a mixing console and a sound reinforcement system for music concerts, theatre, sports games and corporate events.
Alternatively, audio engineer can refer to a scientist or professional engineer who holds an engineering degree and who designs, develops and builds audio or musical technology working under terms such as acoustical engineering, electronic/electrical engineering or (musical) signal processing.
Research and development
Research and development audio engineers invent new technologies, audio software, equipment and techniques, to enhance the process and art of audio engineering. They might design acoustical simulations of rooms, shape algorithms for audio signal processing, specify the requirements for public address systems, carry out research on audible sound for video game console manufacturers, and other advanced fields of audio engineering. They might also be referred to as acoustic engineers.
Education
Audio engineers working in research and development may come from backgrounds such as acoustics, computer science, broadcast engineering, physics, acoustical engineering, electrical engineering and electronics. Audio engineering courses at university or college fall into two rough categories: (i) training in the creative us
Document 4:::
Headphones are a pair of small loudspeaker drivers worn on or around the head over a user's ears. They are electroacoustic transducers, which convert an electrical signal to a corresponding sound. Headphones let a single user listen to an audio source privately, in contrast to a loudspeaker, which emits sound into the open air for anyone nearby to hear. Headphones are also known as earphones or, colloquially, cans. Circumaural ('around the ear') and supra-aural ('over the ear') headphones use a band over the top of the head to hold the speakers in place. Another type, known as earbuds or earpieces consist of individual units that plug into the user's ear canal. A third type are bone conduction headphones, which typically wrap around the back of the head and rest in front of the ear canal, leaving the ear canal open. In the context of telecommunication, a headset is a combination of headphone and microphone.
Headphones connect to a signal source such as an audio amplifier, radio, CD player, portable media player, mobile phone, video game console, or electronic musical instrument, either directly using a cord, or using wireless technology such as Bluetooth, DECT or FM radio. The first headphones were developed in the late 19th century for use by telephone operators, to keep their hands free. Initially the audio quality was mediocre and a step forward was the invention of high fidelity headphones.
Headphones exhibit a range of different audio reproduction quality capabilities. Headsets designed for telephone use typically cannot reproduce sound with the high fidelity of expensive units designed for music listening by audiophiles. Headphones that use cables typically have either a or phone jack for plugging the headphones into the audio source. Some stereo earbuds are wireless, using Bluetooth connectivity to transmit the audio signal by radio waves from source devices like cellphones and digital players. As a result of the Walkman effect, beginning in the 1980s, he
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Wearing what will reduce the amplitude of sound waves entering the ears?
A. hearing protectors
B. hardhat
C. megaphone
D. goggles
Answer:
|
|
sciq-5255
|
multiple_choice
|
In some ways, a collection of gas molecules represents the simplest form of what?
|
[
"matter",
"energy",
"life",
"protein"
] |
A
|
Relavent Documents:
Document 0:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 1:::
This is an index of lists of molecules (i.e. by year, number of atoms, etc.). Millions of molecules have existed in the universe since before the formation of Earth. Three of them, carbon dioxide, water and oxygen were necessary for the growth of life. Although humanity had always been surrounded by these substances, it has not always known what they were composed of.
By century
The following is an index of list of molecules organized by time of discovery of their molecular formula or their specific molecule in case of isomers:
List of compounds
By number of carbon atoms in the molecule
List of compounds with carbon number 1
List of compounds with carbon number 2
List of compounds with carbon number 3
List of compounds with carbon number 4
List of compounds with carbon number 5
List of compounds with carbon number 6
List of compounds with carbon number 7
List of compounds with carbon number 8
List of compounds with carbon number 9
List of compounds with carbon number 10
List of compounds with carbon number 11
List of compounds with carbon number 12
List of compounds with carbon number 13
List of compounds with carbon number 14
List of compounds with carbon number 15
List of compounds with carbon number 16
List of compounds with carbon number 17
List of compounds with carbon number 18
List of compounds with carbon number 19
List of compounds with carbon number 20
List of compounds with carbon number 21
List of compounds with carbon number 22
List of compounds with carbon number 23
List of compounds with carbon number 24
List of compounds with carbon numbers 25-29
List of compounds with carbon numbers 30-39
List of compounds with carbon numbers 40-49
List of compounds with carbon numbers 50+
Other lists
List of interstellar and circumstellar molecules
List of gases
List of molecules with unusual names
See also
Molecule
Empirical formula
Chemical formula
Chemical structure
Chemical compound
Chemical bond
Coordination complex
L
Document 2:::
In mathematical psychology and education theory, a knowledge space is a combinatorial structure used to formulate mathematical models describing the progression of a human learner. Knowledge spaces were introduced in 1985 by Jean-Paul Doignon and Jean-Claude Falmagne, and remain in extensive use in the education theory. Modern applications include two computerized tutoring systems, ALEKS and the defunct RATH.
Formally, a knowledge space assumes that a domain of knowledge is a collection of concepts or skills, each of which must be eventually mastered. Not all concepts are interchangeable; some require other concepts as prerequisites. Conversely, competency at one skill may ease the acquisition of another through similarity. A knowledge space marks out which collections of skills are feasible: they can be learned without mastering any other skills. Under reasonable assumptions, the collection of feasible competencies forms the mathematical structure known as an antimatroid.
Researchers and educators usually explore the structure of a discipline's knowledge space as a latent class model.
Motivation
Knowledge Space Theory attempts to address shortcomings of standardized testing when used in educational psychometry. Common tests, such as the SAT and ACT, compress a student's knowledge into a very small range of ordinal ranks, in the process effacing the conceptual dependencies between questions. Consequently, the tests cannot distinguish between true understanding and guesses, nor can they identify a student's particular weaknesses, only the general proportion of skills mastered. The goal of knowledge space theory is to provide a language by which exams can communicate
What the student can do and
What the student is ready to learn.
Model structure
Knowledge Space Theory-based models presume that an educational subject can be modeled as a finite set of concepts, skills, or topics. Each feasible state of knowledge about is then a subset of ; the set of
Document 3:::
GRE Subject Biochemistry, Cell and Molecular Biology was a standardized exam provided by ETS (Educational Testing Service) that was discontinued in December 2016. It is a paper-based exam and there are no computer-based versions of it. ETS places this exam three times per year: once in April, once in October and once in November. Some graduate programs in the United States recommend taking this exam, while others require this exam score as a part of the application to their graduate programs. ETS sends a bulletin with a sample practice test to each candidate after registration for the exam. There are 180 questions within the biochemistry subject test.
Scores are scaled and then reported as a number between 200 and 990; however, in recent versions of the test, the maximum and minimum reported scores have been 760 (corresponding to the 99 percentile) and 320 (1 percentile) respectively. The mean score for all test takers from July, 2009, to July, 2012, was 526 with a standard deviation of 95.
After learning that test content from editions of the GRE® Biochemistry, Cell and Molecular Biology (BCM) Test has been compromised in Israel, ETS made the decision not to administer this test worldwide in 2016–17.
Content specification
Since many students who apply to graduate programs in biochemistry do so during the first half of their fourth year, the scope of most questions is largely that of the first three years of a standard American undergraduate biochemistry curriculum. A sampling of test item content is given below:
Biochemistry (36%)
A Chemical and Physical Foundations
Thermodynamics and kinetics
Redox states
Water, pH, acid-base reactions and buffers
Solutions and equilibria
Solute-solvent interactions
Chemical interactions and bonding
Chemical reaction mechanisms
B Structural Biology: Structure, Assembly, Organization and Dynamics
Small molecules
Macromolecules (e.g., nucleic acids, polysaccharides, proteins and complex lipids)
Supramolecular complexes (e.g.
Document 4:::
Physics First is an educational program in the United States, that teaches a basic physics course in the ninth grade (usually 14-year-olds), rather than the biology course which is more standard in public schools. This course relies on the limited math skills that the students have from pre-algebra and algebra I. With these skills students study a broad subset of the introductory physics canon with an emphasis on topics which can be experienced kinesthetically or without deep mathematical reasoning. Furthermore, teaching physics first is better suited for English Language Learners, who would be overwhelmed by the substantial vocabulary requirements of Biology.
Physics First began as an organized movement among educators around 1990, and has been slowly catching on throughout the United States. The most prominent movement championing Physics First is Leon Lederman's ARISE (American Renaissance in Science Education).
Many proponents of Physics First argue that turning this order around lays the foundations for better understanding of chemistry, which in turn will lead to more comprehension of biology. Due to the tangible nature of most introductory physics experiments, Physics First also lends itself well to an introduction to inquiry-based science education, where students are encouraged to probe the workings of the world in which they live.
The majority of high schools which have implemented "physics first" do so by way of offering two separate classes, at two separate levels: simple physics concepts in 9th grade, followed by more advanced physics courses in 11th or 12th grade. In schools with this curriculum, nearly all 9th grade students take a "Physical Science", or "Introduction to Physics Concepts" course. These courses focus on concepts that can be studied with skills from pre-algebra and algebra I. With these ideas in place, students then can be exposed to ideas with more physics related content in chemistry, and other science electives. After th
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
In some ways, a collection of gas molecules represents the simplest form of what?
A. matter
B. energy
C. life
D. protein
Answer:
|
|
sciq-9429
|
multiple_choice
|
What is the most common cause of dwarfism?
|
[
"achondroplasia",
"hypothyroidism",
"retinopathy",
"deformity"
] |
A
|
Relavent Documents:
Document 0:::
Achondroplasia in children is the most common form of dwarfism; it accounts for about 70% of all cases of dwarfism. Achondroplasia falls into the category of “disproportionate dwarfism”. It is linked to a mutation in the fibroblast growth factor receptor-3. More than 250,000 people in the world are diagnosed with achondroplasia. Achondroplasia diagnosis occurs somewhere between one in every 10,000 and one in every 30,000 live births.
Some symptoms of achondroplasia are short stature, a long and narrow trunk, shortening of the proximal segments of limbs, large head, mid-face hypoplasia, and joint hyperextension, among others. Achondroplasia is defined by central nervous system defects as well as the prior physical symptoms. Average height for an adult man or woman diagnosed with achondroplasia is about 120 centimeters (47.2 inches), although technically a maximum of 148 centimeters (58.2 inches) is also considered achondroplastic. Achondroplastic people typically have a long trunk and smaller upper legs and upper arms. Achondroplastic people of normal intelligence are able to lead independent and productive lives.
Presentation
Because achondroplastic children have different genes, their growth cycle should be expected to differ from that of a non-achondroplastic child. It is very typical for an achondroplastic child to snore because of their smaller than average size airways. There is no data that unfailingly states respiratory problems. Even though achondroplastic children have reduced lung volumes, this does not seem to result in respiratory problems.
Children diagnosed with achondroplasia usually have delayed motor milestones, otitis media, and bowing of the lower legs. Achondroplastic infants also need to be watched carefully the first few years of infancy for support problems. An infant or child’s hearing and sight needs to be monitored through the years. Along with their development of hearing and sight, their posture will not be guaranteed to develop perfe
Document 1:::
Achondroplasia is a genetic disorder with an autosomal dominant pattern of inheritance whose primary feature is dwarfism. In those with the condition, the arms and legs are short, while the torso is typically of normal length. Those affected have an average adult height of for males and for females. Other features can include an enlarged head and prominent forehead. Complications can include sleep apnea or recurrent ear infections. Achondroplasia includes the extremely rare short-limb skeletal dysplasia with severe combined immunodeficiency.
Achondroplasia is caused by a mutation in the fibroblast growth factor receptor 3 (FGFR3) gene that results in its protein being overactive. Achondroplasia results in impaired endochondral bone growth (bone growth within cartilage). The disorder has an autosomal dominant mode of inheritance, meaning only one mutated copy of the gene is required for the condition to occur. About 80% of cases occur in children of parents without the disease, and result from a new (de novo, or sporadic) mutation, which most commonly originates as a spontaneous change during spermatogenesis. The rest are inherited from a parent with the condition. The risk of a new mutation increases with the age of the father. In families with two affected parents, children who inherit both affected genes typically die before birth or in early infancy from breathing difficulties. The condition is generally diagnosed based on the clinical features but may be confirmed by genetic testing.
Treatments may include support groups and growth hormone therapy. Efforts to treat or prevent complications such as obesity, hydrocephalus, obstructive sleep apnea, middle ear infections or spinal stenosis may be required. Achondroplasia is the most common cause of dwarfism and affects about 1 in 27,500 people.
Signs and symptoms
Disproportionate dwarfism
Shortening of the proximal limbs (called rhizomelic shortening)
Short fingers and toes, with "trident hands" (short han
Document 2:::
Short stature refers to a height of a human which is below typical. Whether a person is considered short depends on the context. Because of the lack of preciseness, there is often disagreement about the degree of shortness that should be called short. Dwarfism is the condition of being very short, often caused by a medical condition.
In a medical context, short stature is typically defined as an adult height that is more than two standard deviations below a population’s mean for age and gender, which corresponds to the shortest 2.3% of individuals in that population. The median or typical adult height in developed countries is about for men and for women.
Causes
Shortness in children and young adults nearly always results from below-average growth in childhood, while shortness in older adults usually results from loss of height due to kyphosis of the spine or collapsed vertebrae from osteoporosis. The most common causes of short stature in childhood are constitutional growth delay or familial short stature.
From a medical perspective, severe shortness can be a variation of normal, resulting from the interplay of multiple familial genes. It can also be due to one or more of many abnormal conditions, such as chronic (prolonged) growth hormone or thyroid hormone deficiency, malnutrition, disease of a major organ system, mistreatment, treatment with certain drugs, chromosomal deletions. Human growth hormone (HGH) deficiency may occur at any time during infancy or childhood, with the most obvious sign being a noticeable slowing of growth. The deficiency may be genetic. Among children without growth hormone deficiency, short stature may be caused by Turner syndrome or Noonan syndrome, chronic kidney disease, being small for gestational age at birth, Prader–Willi syndrome, Wiedemann-Steiner syndrome, or other conditions. Genetic skeletal dysplasias also known as osteochondrodysplasia usually manifest in short-limbed disproportionate short stature.
When the cause is
Document 3:::
The STEM (Science, Technology, Engineering, and Mathematics) pipeline is a critical infrastructure for fostering the development of future scientists, engineers, and problem solvers. It's the educational and career pathway that guides individuals from early childhood through to advanced research and innovation in STEM-related fields.
Description
The "pipeline" metaphor is based on the idea that having sufficient graduates requires both having sufficient input of students at the beginning of their studies, and retaining these students through completion of their academic program. The STEM pipeline is a key component of workplace diversity and of workforce development that ensures sufficient qualified candidates are available to fill scientific and technical positions.
The STEM pipeline was promoted in the United States from the 1970s onwards, as “the push for STEM (science, technology, engineering, and mathematics) education appears to have grown from a concern for the low number of future professionals to fill STEM jobs and careers and economic and educational competitiveness.”
Today, this metaphor is commonly used to describe retention problems in STEM fields, called “leaks” in the pipeline. For example, the White House reported in 2012 that 80% of minority groups and women who enroll in a STEM field switch to a non-STEM field or drop out during their undergraduate education. These leaks often vary by field, gender, ethnic and racial identity, socioeconomic background, and other factors, drawing attention to structural inequities involved in STEM education and careers.
Current efforts
The STEM pipeline concept is a useful tool for programs aiming at increasing the total number of graduates, and is especially important in efforts to increase the number of underrepresented minorities and women in STEM fields. Using STEM methodology, educational policymakers can examine the quantity and retention of students at all stages of the K–12 educational process and beyo
Document 4:::
The Science, Technology, Engineering and Mathematics Network or STEMNET is an educational charity in the United Kingdom that seeks to encourage participation at school and college in science and engineering-related subjects (science, technology, engineering, and mathematics) and (eventually) work.
History
It is based at Woolgate Exchange near Moorgate tube station in London and was established in 1996. The chief executive is Kirsten Bodley. The STEMNET offices are housed within the Engineering Council.
Function
Its chief aim is to interest children in science, technology, engineering and mathematics. Primary school children can start to have an interest in these subjects, leading secondary school pupils to choose science A levels, which will lead to a science career. It supports the After School Science and Engineering Clubs at schools. There are also nine regional Science Learning Centres.
STEM ambassadors
To promote STEM subjects and encourage young people to take up jobs in these areas, STEMNET have around 30,000 ambassadors across the UK. these come from a wide selection of the STEM industries and include TV personalities like Rob Bell.
Funding
STEMNET used to receive funding from the Department for Education and Skills. Since June 2007, it receives funding from the Department for Children, Schools and Families and Department for Innovation, Universities and Skills, since STEMNET sits on the chronological dividing point (age 16) of both of the new departments.
See also
The WISE Campaign
Engineering and Physical Sciences Research Council
National Centre for Excellence in Teaching Mathematics
Association for Science Education
Glossary of areas of mathematics
Glossary of astronomy
Glossary of biology
Glossary of chemistry
Glossary of engineering
Glossary of physics
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is the most common cause of dwarfism?
A. achondroplasia
B. hypothyroidism
C. retinopathy
D. deformity
Answer:
|
|
sciq-6700
|
multiple_choice
|
What are the two types of vascular plants?
|
[
"dahlias and seed",
"seedless and seed",
"Water and spore",
"ants and seed"
] |
B
|
Relavent Documents:
Document 0:::
Vascular plants (), also called tracheophytes () or collectively Tracheophyta (), form a large group of land plants ( accepted known species) that have lignified tissues (the xylem) for conducting water and minerals throughout the plant. They also have a specialized non-lignified tissue (the phloem) to conduct products of photosynthesis. Vascular plants include the clubmosses, horsetails, ferns, gymnosperms (including conifers), and angiosperms (flowering plants). Scientific names for the group include Tracheophyta, Tracheobionta and Equisetopsida sensu lato. Some early land plants (the rhyniophytes) had less developed vascular tissue; the term eutracheophyte has been used for all other vascular plants, including all living ones.
Historically, vascular plants were known as "higher plants", as it was believed that they were further evolved than other plants due to being more complex organisms. However, this is an antiquated remnant of the obsolete scala naturae, and the term is generally considered to be unscientific.
Characteristics
Botanists define vascular plants by three primary characteristics:
Vascular plants have vascular tissues which distribute resources through the plant. Two kinds of vascular tissue occur in plants: xylem and phloem. Phloem and xylem are closely associated with one another and are typically located immediately adjacent to each other in the plant. The combination of one xylem and one phloem strand adjacent to each other is known as a vascular bundle. The evolution of vascular tissue in plants allowed them to evolve to larger sizes than non-vascular plants, which lack these specialized conducting tissues and are thereby restricted to relatively small sizes.
In vascular plants, the principal generation or phase is the sporophyte, which produces spores and is diploid (having two sets of chromosomes per cell). (By contrast, the principal generation phase in non-vascular plants is the gametophyte, which produces gametes and is haploid - with
Document 1:::
The following is a list of vascular plants, bryophytes and lichens which are constant species in one or more community of the British National Vegetation Classification system.
Vascular plants
Grasses
Sedges and rushes
Trees
Other dicotyledons
Other monocotyledons
Ferns
Clubmosses
Bryophytes
Mosses
Liverworts
Lichens
British National Vegetation Classification
Lists of biota of the United Kingdom
British National Vegetation Classification, constant
Document 2:::
Plant life-form schemes constitute a way of classifying plants alternatively to the ordinary species-genus-family scientific classification. In colloquial speech, plants may be classified as trees, shrubs, herbs (forbs and graminoids), etc. The scientific use of life-form schemes emphasizes plant function in the ecosystem and that the same function or "adaptedness" to the environment may be achieved in a number of ways, i.e. plant species that are closely related phylogenetically may have widely different life-form, for example Adoxa moschatellina and Sambucus nigra are from the same family, but the former is a small herbaceous plant and the latter is a shrub or tree. Conversely, unrelated species may share a life-form through convergent evolution.
While taxonomic classification is concerned with the production of natural classifications (being natural understood either in philosophical basis for pre-evolutionary thinking, or phylogenetically as non-polyphyletic), plant life form classifications uses other criteria than naturalness, like morphology, physiology and ecology.
Life-form and growth-form are essentially synonymous concepts, despite attempts to restrict the meaning of growth-form to types differing in shoot architecture. Most life form schemes are concerned with vascular plants only. Plant construction types may be used in a broader sense to encompass planktophytes, benthophytes (mainly algae) and terrestrial plants.
A popular life-form scheme is the Raunkiær system.
History
One of the earliest attempts to classify the life-forms of plants and animals was made by Aristotle, whose writings are lost. His pupil, Theophrastus, in Historia Plantarum (c. 350 BC), was the first who formally recognized plant habits: trees, shrubs and herbs.
Some earlier authors (e.g., Humboldt, 1806) did classify species according to physiognomy, but were explicit about the entities being merely practical classes without any relation to plant function. A marked exception was
Document 3:::
Microspores are land plant spores that develop into male gametophytes, whereas megaspores develop into female gametophytes. The male gametophyte gives rise to sperm cells, which are used for fertilization of an egg cell to form a zygote. Megaspores are structures that are part of the alternation of generations in many seedless vascular cryptogams, all gymnosperms and all angiosperms. Plants with heterosporous life cycles using microspores and megaspores arose independently in several plant groups during the Devonian period. Microspores are haploid, and are produced from diploid microsporocytes by meiosis.
Morphology
The microspore has three different types of wall layers. The outer layer is called the perispore, the next is the exospore, and the inner layer is the endospore. The perispore is the thickest of the three layers while the exospore and endospore are relatively equal in width.
Seedless vascular plants
In heterosporous seedless vascular plants, modified leaves called microsporophylls bear microsporangia containing many microsporocytes that undergo meiosis, each producing four microspores. Each microspore may develop into a male gametophyte consisting of a somewhat spherical antheridium within the microspore wall. Either 128 or 256 sperm cells with flagella are produced in each antheridium. The only heterosporous ferns are aquatic or semi-aquatic, including the genera Marsilea, Regnellidium, Pilularia, Salvinia, and Azolla. Heterospory also occurs in the lycopods in the spikemoss genus Selaginella and in the quillwort genus Isoëtes.
Types of seedless vascular plants:
Water ferns
Spikemosses
Quillworts
Gymnosperms
In seed plants the microspores develop into pollen grains each containing a reduced, multicellular male gametophyte. The megaspores, in turn, develop into reduced female gametophytes that produce egg cells that, once fertilized, develop into seeds. Pollen cones or microstrobili usually develop toward the tips of the lower branches in cluste
Document 4:::
In botany, a virtual herbarium is a herbarium in a digitized form. That is, it concerns a collection of digital images of preserved plants or plant parts. Virtual herbaria often are established to improve availability of specimens to a wider audience. However, there are digital herbaria that are not suitable for internet access because of the high resolution of scans and resulting large file sizes (several hundred megabytes per file). Additional information about each specimen, such as the location, the collector, and the botanical name are attached to every specimen. Frequently, further details such as related species and growth requirements are mentioned.
Specimen imaging
The standard hardware used for herbarium specimen imaging is the "HerbScan" scanner. It is an inverted flat-bed scanner which raises the specimen up to the scanning surface. This technology was developed because it is standard practice to never turn a herbarium specimen upside-down. Alternatively, some herbaria employ a flat-bed book scanner or a copy stand to achieve the same effect.
A small color chart and a ruler must be included on a herbarium sheet when it is imaged. The JSTOR Plant Science requires that the ruler bears the herbarium name and logo, and that a ColorChecker chart is used for any specimens to be contributed to the Global Plants Initiative (GPI).
Uses
Virtual herbaria are established in part to increase the longevity of specimens. Major herbaria participate in international loan programs, where a researcher can request specimens to be shipped in for study. This shipping contributes to the wear and tear of specimens. If, however, digital images are available, images of the specimens can be sent electronically. These images may be a sufficient substitute for the specimens themselves, or alternatively, the researcher can use the images to "preview" the specimens, to which ones should be sent out for further study. This process cuts down on the shipping, and thus the wear and
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What are the two types of vascular plants?
A. dahlias and seed
B. seedless and seed
C. Water and spore
D. ants and seed
Answer:
|
|
sciq-397
|
multiple_choice
|
What process refers to a separation of charge within an atom or molecule?
|
[
"polarization",
"rotation",
"diffusion",
"convection"
] |
A
|
Relavent Documents:
Document 0:::
The electric dipole moment is a measure of the separation of positive and negative electrical charges within a system, that is, a measure of the system's overall polarity. The SI unit for electric dipole moment is the coulomb-meter (C⋅m). The debye (D) is another unit of measurement used in atomic physics and chemistry.
Theoretically, an electric dipole is defined by the first-order term of the multipole expansion; it consists of two equal and opposite charges that are infinitesimally close together, although real dipoles have separated charge.
Elementary definition
Often in physics the dimensions of a massive object can be ignored and can be treated as a pointlike object, i.e. a point particle. Point particles with electric charge are referred to as point charges. Two point charges, one with charge and the other one with charge separated by a distance , constitute an electric dipole (a simple case of an electric multipole). For this case, the electric dipole moment has a magnitude and is directed from the negative charge to the positive one. Some authors may split in half and use since this quantity is the distance between either charge and the center of the dipole, leading to a factor of two in the definition.
A stronger mathematical definition is to use vector algebra, since a quantity with magnitude and direction, like the dipole moment of two point charges, can be expressed in vector form where is the displacement vector pointing from the negative charge to the positive charge. The electric dipole moment vector also points from the negative charge to the positive charge. With this definition the dipole direction tends to align itself with an external electric field (and note that the electric flux lines produced by the charges of the dipole itself, which point from positive charge to negative charge then tend to oppose the flux lines of the external field). Note that this sign convention is used in physics, while the opposite sign convention for th
Document 1:::
Photoinduced charge separation is the process of an electron in an atom or molecule, being excited to a higher energy level by the absorption of a photon and then leaving the atom or molecule to free space, or to a nearby electron acceptor.
Rutherford model
An atom consists of a positively-charged nucleus surrounded by bound electrons. The nucleus consists of uncharged neutrons and positively charged protons. Electrons are negatively charged. In the early part of the twentieth century Ernest Rutherford suggested that the electrons orbited the dense central nucleus in a manner analogous to planets orbiting the Sun. The centripetal force required to keep the electrons in orbit was provided by the Coulomb force of the protons in the nucleus acting upon the electrons; just like the gravitational force of the Sun acting on a planet provides the centripetal force necessary to keep the planet in orbit.
This model, although appealing, doesn't hold true in the real world. Synchrotron radiation would cause the orbiting electron to lose orbital energy and spiral inward since the vector quantity of acceleration of the particle multiplied by its mass (the value of the force required to keep the electron in circular motion) would be less than the electrical force the proton applied to the electron.
Once the electron spiralled into the nucleus the electron would combine with a proton to form a neutron, and the atom would cease to exist. This model is clearly wrong.
Bohr model
In 1913, Niels Bohr refined the Rutherford model by stating that the electrons existed in discrete quantized states called energy levels. This meant that the electrons could only occupy orbits at certain energies. The laws of quantum physics apply here, and they don't comply with the laws of classical newtonian mechanics.
An electron which is stationary and completely free from the atom has an energy of 0 joules (or 0 electronvolts). An electron which is described as being at the "ground state" has a (ne
Document 2:::
Atomic physics is the field of physics that studies atoms as an isolated system of electrons and an atomic nucleus. Atomic physics typically refers to the study of atomic structure and the interaction between atoms. It is primarily concerned with the way in which electrons are arranged around the nucleus and
the processes by which these arrangements change. This comprises ions, neutral atoms and, unless otherwise stated, it can be assumed that the term atom includes ions.
The term atomic physics can be associated with nuclear power and nuclear weapons, due to the synonymous use of atomic and nuclear in standard English. Physicists distinguish between atomic physics—which deals with the atom as a system consisting of a nucleus and electrons—and nuclear physics, which studies nuclear reactions and special properties of atomic nuclei.
As with many scientific fields, strict delineation can be highly contrived and atomic physics is often considered in the wider context of atomic, molecular, and optical physics. Physics research groups are usually so classified.
Isolated atoms
Atomic physics primarily considers atoms in isolation. Atomic models will consist of a single nucleus that may be surrounded by one or more bound electrons. It is not concerned with the formation of molecules (although much of the physics is identical), nor does it examine atoms in a solid state as condensed matter. It is concerned with processes such as ionization and excitation by photons or collisions with atomic particles.
While modelling atoms in isolation may not seem realistic, if one considers atoms in a gas or plasma then the time-scales for atom-atom interactions are huge in comparison to the atomic processes that are generally considered. This means that the individual atoms can be treated as if each were in isolation, as the vast majority of the time they are. By this consideration, atomic physics provides the underlying theory in plasma physics and atmospheric physics, even though
Document 3:::
In physics, a charge carrier is a particle or quasiparticle that is free to move, carrying an electric charge, especially the particles that carry electric charges in electrical conductors. Examples are electrons, ions and holes. The term is used most commonly in solid state physics. In a conducting medium, an electric field can exert force on these free particles, causing a net motion of the particles through the medium; this is what constitutes an electric current.
The electron and the proton are the elementary charge carriers, each carrying one elementary charge (e), of the same magnitude and opposite sign.
In conductors
In conducting media, particles serve to carry charge:
In many metals, the charge carriers are electrons. One or two of the valence electrons from each atom are able to move about freely within the crystal structure of the metal. The free electrons are referred to as conduction electrons, and the cloud of free electrons is called a Fermi gas. Many metals have electron and hole bands. In some, the majority carriers are holes.
In electrolytes, such as salt water, the charge carriers are ions, which are atoms or molecules that have gained or lost electrons so they are electrically charged. Atoms that have gained electrons so they are negatively charged are called anions, atoms that have lost electrons so they are positively charged are called cations. Cations and anions of the dissociated liquid also serve as charge carriers in melted ionic solids (see e.g. the Hall–Héroult process for an example of electrolysis of a melted ionic solid). Proton conductors are electrolytic conductors employing positive hydrogen ions as carriers.
In a plasma, an electrically charged gas which is found in electric arcs through air, neon signs, and the sun and stars, the electrons and cations of ionized gas act as charge carriers.
In a vacuum, free electrons can act as charge carriers. In the electronic component known as the vacuum tube (also called valve), the mobil
Document 4:::
Gas phase ion chemistry is a field of science encompassed within both chemistry and physics. It is the science that studies ions and molecules in the gas phase, most often enabled by some form of mass spectrometry. By far the most important applications for this science is in studying the thermodynamics and kinetics of reactions. For example, one application is in studying the thermodynamics of the solvation of ions. Ions with small solvation spheres of 1, 2, 3... solvent molecules can be studied in the gas phase and then extrapolated to bulk solution.
Theory
Transition state theory
Transition state theory is the theory of the rates of elementary reactions which assumes a special type of chemical equilibrium (quasi-equilibrium) between reactants and activated complexes.
RRKM theory
RRKM theory is used to compute simple estimates of the unimolecular ion decomposition reaction rates from a few characteristics of the potential energy surface.
Gas phase ion formation
The process of converting an atom or molecule into an ion by adding or removing charged particles such as electrons or other ions can occur in the gas phase. These processes are an important component of gas phase ion chemistry.
Associative ionization
Associative ionization is a gas phase reaction in which two atoms or molecules interact to form a single product ion.
where species A with excess internal energy (indicated by the asterisk) interacts with B to form the ion AB+.
One or both of the interacting species may have excess internal energy.
Charge-exchange ionization
Charge-exchange ionization (also called charge-transfer ionization) is a gas phase reaction between an ion and a neutral species
in which the charge of the ion is transferred to the neutral.
Chemical ionization
In chemical ionization, ions are produced through the reaction of ions of a reagent gas with other species. Some common reagent gases include: methane, ammonia, and isobutane.
Chemi-ionization
Chemi-ionization can
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What process refers to a separation of charge within an atom or molecule?
A. polarization
B. rotation
C. diffusion
D. convection
Answer:
|
|
sciq-5501
|
multiple_choice
|
What body part does encephalitis attack?
|
[
"liver",
"brain",
"lungs",
"heart"
] |
B
|
Relavent Documents:
Document 0:::
Several universities have designed interdisciplinary courses with a focus on human biology at the undergraduate level. There is a wide variation in emphasis ranging from business, social studies, public policy, healthcare and pharmaceutical research.
Americas
Human Biology major at Stanford University, Palo Alto (since 1970)
Stanford's Human Biology Program is an undergraduate major; it integrates the natural and social sciences in the study of human beings. It is interdisciplinary and policy-oriented and was founded in 1970 by a group of Stanford faculty (Professors Dornbusch, Ehrlich, Hamburg, Hastorf, Kennedy, Kretchmer, Lederberg, and Pittendrigh). It is a very popular major and alumni have gone to post-graduate education, medical school, law, business and government.
Human and Social Biology (Caribbean)
Human and Social Biology is a Level 4 & 5 subject in the secondary and post-secondary schools in the Caribbean and is optional for the Caribbean Secondary Education Certification (CSEC) which is equivalent to Ordinary Level (O-Level) under the British school system. The syllabus centers on structure and functioning (anatomy, physiology, biochemistry) of human body and the relevance to human health with Caribbean-specific experience. The syllabus is organized under five main sections: Living organisms and the environment, life processes, heredity and variation, disease and its impact on humans, the impact of human activities on the environment.
Human Biology Program at University of Toronto
The University of Toronto offers an undergraduate program in Human Biology that is jointly offered by the Faculty of Arts & Science and the Faculty of Medicine. The program offers several major and specialist options in: human biology, neuroscience, health & disease, global health, and fundamental genetics and its applications.
Asia
BSc (Honours) Human Biology at All India Institute of Medical Sciences, New Delhi (1980–2002)
BSc (honours) Human Biology at AIIMS (New
Document 1:::
Neurotrauma, brain damage or brain injury (BI) is the destruction or degeneration of brain cells. Brain injuries occur due to a wide range of internal and external factors. In general, brain damage refers to significant, undiscriminating trauma-induced damage.
A common category with the greatest number of injuries is traumatic brain injury (TBI) following physical trauma or head injury from an outside source, and the term acquired brain injury (ABI) is used in appropriate circles to differentiate brain injuries occurring after birth from injury, from a genetic disorder (GBI), or from a congenital disorder (CBI). Primary and secondary brain injuries identify the processes involved, while focal and diffuse brain injury describe the severity and localization.
Recent research has demonstrated that neuroplasticity, which allows the brain to reorganize itself by forming new neural connections throughout life, provides for rearrangement of its workings. This allows the brain to compensate for injury and disease.
Signs and symptoms
Symptoms of brain injuries vary based on the severity of the injury or how much of the brain is affected. The three categories used for classifying the severity of brain injuries are mild, moderate or severe.
Severity of injuries
Mild brain injuries
Symptoms of a mild brain injury include headaches, confusions, tinnitus, fatigue, changes in sleep patterns, mood or behavior. Other symptoms include trouble with memory, concentration, attention or thinking. Mental fatigue is a common debilitating experience and may not be linked by the patient to the original (minor) incident.
Moderate/severe brain injuries
Cognitive symptoms include confusion, aggressiveness, abnormal behavior, slurred speech, and coma or other disorders of consciousness. Physical symptoms include headaches that worsen or do not go away, vomiting or nausea, convulsions, brain pulsation, abnormal dilation of the eyes, inability to awaken from sleep, weakness in extremities a
Document 2:::
There is much to be discovered about the evolution of the brain and the principles that govern it. While much has been discovered, not everything currently known is well understood. The evolution of the brain has appeared to exhibit diverging adaptations within taxonomic classes such as Mammalia and more vastly diverse adaptations across other taxonomic classes.
Brain to body size scales allometrically. This means as body size changes, so do other physiological, anatomical, and biochemical constructs connecting the brain to the body. Small bodied mammals have relatively large brains compared to their bodies whereas large mammals (such as whales) have a smaller brain to body ratios. If brain weight is plotted against body weight for primates, the regression line of the sample points can indicate the brain power of a primate species. Lemurs for example fall below this line which means that for a primate of equivalent size, we would expect a larger brain size. Humans lie well above the line indicating that humans are more encephalized than lemurs. In fact, humans are more encephalized compared to all other primates. This means that human brains have exhibited a larger evolutionary increase in its complexity relative to its size. Some of these evolutionary changes have been found to be linked to multiple genetic factors, such as proteins and other organelles.
Early history of brain development
One approach to understanding overall brain evolution is to use a paleoarchaeological timeline to trace the necessity for ever increasing complexity in structures that allow for chemical and electrical signaling. Because brains and other soft tissues do not fossilize as readily as mineralized tissues, scientists often look to other structures as evidence in the fossil record to get an understanding of brain evolution. This, however, leads to a dilemma as the emergence of organisms with more complex nervous systems with protective bone or other protective tissues that can then
Document 3:::
The following diagram is provided as an overview of and topical guide to the human nervous system:
Human nervous system – the part of the human body that coordinates a person's voluntary and involuntary actions and transmits signals between different parts of the body. The human nervous system consists of two main parts: the central nervous system (CNS) and the peripheral nervous system (PNS). The CNS contains the brain and spinal cord. The PNS consists mainly of nerves, which are long fibers that connect the CNS to every other part of the body. The PNS includes motor neurons, mediating voluntary movement; the autonomic nervous system, comprising the sympathetic nervous system and the parasympathetic nervous system and regulating involuntary functions; and the enteric nervous system, a semi-independent part of the nervous system whose function is to control the gastrointestinal system.
Evolution of the human nervous system
Evolution of nervous systems
Evolution of human intelligence
Evolution of the human brain
Paleoneurology
Some branches of science that study the human nervous system
Neuroscience
Neurology
Paleoneurology
Central nervous system
The central nervous system (CNS) is the largest part of the nervous system and includes the brain and spinal cord.
Spinal cord
Brain
Brain – center of the nervous system.
Outline of the human brain
List of regions of the human brain
Principal regions of the vertebrate brain:
Peripheral nervous system
Peripheral nervous system (PNS) – nervous system structures that do not lie within the CNS.
Sensory system
A sensory system is a part of the nervous system responsible for processing sensory information. A sensory system consists of sensory receptors, neural pathways, and parts of the brain involved in sensory perception.
List of sensory systems
Sensory neuron
Perception
Visual system
Auditory system
Somatosensory system
Vestibular system
Olfactory system
Taste
Pain
Components of the nervous system
Neuron
I
Document 4:::
Acute disseminated encephalomyelitis (ADEM), or acute demyelinating encephalomyelitis, is a rare autoimmune disease marked by a sudden, widespread attack of inflammation in the brain and spinal cord. As well as causing the brain and spinal cord to become inflamed, ADEM also attacks the nerves of the central nervous system and damages their myelin insulation, which, as a result, destroys the white matter. The cause is often a trigger such as from viral infection or vaccinations.
ADEM's symptoms resemble the symptoms of multiple sclerosis (MS), so the disease itself is sorted into the classification of the multiple sclerosis borderline diseases. However, ADEM has several features that distinguish it from MS. Unlike MS, ADEM occurs usually in children and is marked with rapid fever, although adolescents and adults can get the disease too. ADEM consists of a single flare-up whereas MS is marked with several flare-ups (or relapses), over a long period of time. Relapses following ADEM are reported in up to a quarter of patients, but the majority of these 'multiphasic' presentations following ADEM likely represent MS. ADEM is also distinguished by a loss of consciousness, coma and death, which is very rare in MS, except in severe cases.
It affects about 8 per 1,000,000 people per year. Although it occurs in all ages, most reported cases are in children and adolescents, with the average age around 5 to 8 years old. The disease affects males and females almost equally. ADEM shows seasonal variation with higher incidence in winter and spring months which may coincide with higher viral infections during these months. The mortality rate may be as high as 5%; however, full recovery is seen in 50 to 75% of cases with increase in survival rates up to 70 to 90% with figures including minor residual disability as well. The average time to recover from ADEM flare-ups is one to six months.
ADEM produces multiple inflammatory lesions in the brain and spinal cord, particularly in the
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What body part does encephalitis attack?
A. liver
B. brain
C. lungs
D. heart
Answer:
|
|
sciq-11041
|
multiple_choice
|
Makemake and pluto are classified as what type of celestial object?
|
[
"dwarf planets",
"gas giants",
"black holes",
"nebula"
] |
A
|
Relavent Documents:
Document 0:::
Astrophysics is a science that employs the methods and principles of physics and chemistry in the study of astronomical objects and phenomena. As one of the founders of the discipline, James Keeler, said, Astrophysics "seeks to ascertain the nature of the heavenly bodies, rather than their positions or motions in space–what they are, rather than where they are." Among the subjects studied are the Sun (solar physics), other stars, galaxies, extrasolar planets, the interstellar medium and the cosmic microwave background. Emissions from these objects are examined across all parts of the electromagnetic spectrum, and the properties examined include luminosity, density, temperature, and chemical composition. Because astrophysics is a very broad subject, astrophysicists apply concepts and methods from many disciplines of physics, including classical mechanics, electromagnetism, statistical mechanics, thermodynamics, quantum mechanics, relativity, nuclear and particle physics, and atomic and molecular physics.
In practice, modern astronomical research often involves a substantial amount of work in the realms of theoretical and observational physics. Some areas of study for astrophysicists include their attempts to determine the properties of dark matter, dark energy, black holes, and other celestial bodies; and the origin and ultimate fate of the universe. Topics also studied by theoretical astrophysicists include Solar System formation and evolution; stellar dynamics and evolution; galaxy formation and evolution; magnetohydrodynamics; large-scale structure of matter in the universe; origin of cosmic rays; general relativity, special relativity, quantum and physical cosmology, including string cosmology and astroparticle physics.
History
Astronomy is an ancient science, long separated from the study of terrestrial physics. In the Aristotelian worldview, bodies in the sky appeared to be unchanging spheres whose only motion was uniform motion in a circle, while the earthl
Document 1:::
Astronomy education or astronomy education research (AER) refers both to the methods currently used to teach the science of astronomy and to an area of pedagogical research that seeks to improve those methods. Specifically, AER includes systematic techniques honed in science and physics education to understand what and how students learn about astronomy and determine how teachers can create more effective learning environments.
Education is important to astronomy as it impacts both the recruitment of future astronomers and the appreciation of astronomy by citizens and politicians who support astronomical research. Astronomy has been taught throughout much of recorded human history, and has practical application in timekeeping and navigation. Teaching astronomy contributes to an understanding of physics and the origin of the world around us, a shared cultural background, and a sense of wonder and exploration. It includes education of the general public through planetariums, books, and instructive presentations, plus programs and tools for amateur astronomy, and University-level degree programs for professional astronomers. Astronomy organizations provide educational functions and societies in about 100 nation states around the world.
In schools, particularly at the collegiate level, astronomy is aligned with physics and the two are often combined to form a Department of Physics and Astronomy. Some parts of astronomy education overlap with physics education, however, astronomy education has its own arenas, practitioners, journals, and research. This can be demonstrated in the identified 20-year lag between the emergence of AER and physics education research. The body of research in this field are available through electronic sources such as the Searchable Annotated Bibliography of Education Research (SABER) and the American Astronomical Society's database of the contents of their journal "Astronomy Education Review" (see link below).
The National Aeronautics and
Document 2:::
A dwarf planet is a small planetary-mass object that is in direct orbit of the Sun, smaller than any of the eight classical planets. The prototypical dwarf planet is Pluto. The interest of dwarf planets to planetary geologists is that they may be geologically active bodies, an expectation that was borne out in 2015 by the Dawn mission to and the New Horizons mission to Pluto.
Astronomers are in general agreement that at least the nine largest candidates are dwarf planets – in rough order of size, , , , , , , , , and – although there is some doubt for Orcus. Of these nine plus the tenth-largest candidate , two have been visited by spacecraft (Pluto and Ceres) and seven others have at least one known moon (Eris, Haumea, Makemake, Gonggong, Quaoar, Orcus, and Salacia), which allows their masses and thus an estimate of their densities to be determined. Mass and density in turn can be fit into geophysical models in an attempt to determine the nature of these worlds. Only one, Sedna, has neither been visited nor has any known moons, making an accurate estimate of mass difficult. Some astronomers include many smaller bodies as well, but there is no consensus that these are likely to be dwarf planets.
The term dwarf planet was coined by planetary scientist Alan Stern as part of a three-way categorization of planetary-mass objects in the Solar System: classical planets, dwarf planets, and satellite planets. Dwarf planets were thus conceived of as a category of planet. In 2006, however, the concept was adopted by the International Astronomical Union (IAU) as a category of sub-planetary objects, part of a three-way recategorization of bodies orbiting the Sun: planets, dwarf planets and small Solar System bodies. Thus Stern and other planetary geologists consider dwarf planets and large satellites to be planets, but since 2006, the IAU and perhaps the majority of astronomers have excluded them from the roster of planets.
History of the concept
Starting in 1801, astronom
Document 3:::
An asteroid spectral type is assigned to asteroids based on their reflectance spectrum, color, and sometimes albedo. These types are thought to correspond to an asteroid's surface composition. For small bodies that are not internally differentiated, the surface and internal compositions are presumably similar, while large bodies such as Ceres and Vesta are known to have internal structure. Over the years, there has been a number of surveys that resulted in a set of different taxonomic systems such as the Tholen, SMASS and Bus–DeMeo classifications.
Taxonomic systems
In 1975, astronomers Clark R. Chapman, David Morrison, and Ben Zellner developed a simple taxonomic system for asteroids based on color, albedo, and spectral shape. The three categories were labelled "C" for dark carbonaceous objects, "S" for stony (silicaceous) objects, and "U" for those that did not fit into either C or S. This basic division of asteroid spectra has since been expanded and clarified. A number of classification schemes are currently in existence, and while they strive to retain some mutual consistency, quite a few asteroids are sorted into different classes depending on the particular scheme. This is due to the use of different criteria for each approach. The two most widely used classifications are described below:
Overview of Tholen and SMASS
S3OS2 classification
The Small Solar System Objects Spectroscopic Survey (S3OS2 or S3OS2, also known as the Lazzaro classification) observed 820 asteroids, using the former ESO 1.52-metre telescope at La Silla Observatory during 1996–2001. This survey applied both the Tholen and Bus–Binzel (SMASS) taxonomy to the observed objects, many of which had previously not been classified. For the Tholen-like classification, the survey introduced a new "Caa-type", which shows a broad absorption band associated indicating an aqueous alteration of the body's surface. The Caa class corresponds to Tholen's C-type and to the SMASS hydrated Ch-type (inclu
Document 4:::
BL Herculis variables are a subclass of type II Cepheids with low luminosity and mass, that have a period of less than eight days. They are pulsating stars with light curves that frequently show a bump on the descending side for stars of the shortest periods and on the ascending side for longer period stars. Like other type II Cepheids, they are very old population II stars found in the galaxy’s halo and globular clusters. Also, compared to other type II Cepheids, BL Herculis variables have shorter periods and are fainter than W Virginis variables. Pulsating stars vary in spectral class as they vary in brightness and BL Herculis variables are normally class A at their brightest and class F when most dim. When plotted on the Hertzsprung–Russell diagram they fall in-between W Virginis and RR Lyrae variables.
The prototype star, BL Herculis, varies between magnitude 9.7 and 10.6 in a period of 1.3 days. The brightest BL Herculis variables, with their maximum magnitudes, are:
VY Pyxidis, 7.7
V553 Centauri, 8.2
SW Tauri, 9.3
RT Trianguli Australis, 9.4
V351 Cephei, 9.5
BL Herculis. 9.7
BD Cassiopeiae, 10.8
UY Eridani, 10.9
The BL Herculis stars show a wide variety of light curves, temperatures, and luminosity, and three subdivisions of the class have been defined, with the acronym AHB referring to above horizontal branch:
XX Virginis stars (AHB1), with very fast rises to maximum and low metallicity
CW stars (AHB2), W Virginis variables, longer periods, the bump on the ascending leg
BL Herculis stars (AHB3), shorter periods, the bump on the descending leg
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Makemake and pluto are classified as what type of celestial object?
A. dwarf planets
B. gas giants
C. black holes
D. nebula
Answer:
|
|
sciq-6395
|
multiple_choice
|
For each initial glucose molecules, two pyruvate molecules will enter where?
|
[
"the plasma",
"Golgi apparatus",
"the nucleus",
"the mitochondria"
] |
D
|
Relavent Documents:
Document 0:::
Pyruvate cycling commonly refers to an intracellular loop of spatial movements and chemical transformations involving pyruvate. Spatial movements occur between mitochondria and cytosol and chemical transformations create various Krebs cycle intermediates. In all variants, pyruvate is imported into the mitochondrion for processing through part of the Krebs cycle. In addition to pyruvate, alpha-ketoglutarate may also be imported. At various points, the intermediate product is exported to the cytosol for additional transformations and then re-imported. Three specific pyruvate cycles are generally considered, each named for the principal molecule exported from the mitochondrion: malate, citrate, and isocitrate. Other variants may exist, such as dissipative or "futile" pyruvate cycles.
This cycle is usually studied in relation to Glucose Stimulated Insulin Secretion ( or GSIS ) and there is thought to be a relationship between the insulin response and NADPH produced from this cycle but the specifics are not clear and particular confusion exists about the role of malic enzymes. It has been observed in various cell types including islet cells.
The pyruvate-malate cycle was described in liver and kidney preparations as early as 1971.
Document 1:::
Biochemistry
Pyruvate is an important chemical compound in biochemistry. It is the output of the metabolism of glucose known as glycolysis. One molecule of glucose breaks down into two molecules of pyruvate, which are then used to provide further energy, in one of two ways. Pyruvate is converted into acetyl-coenzyme A, which is the main input for a series of reactions known as
Document 2:::
Biochemistry or biological chemistry is the study of chemical processes within and relating to living organisms. A sub-discipline of both chemistry and biology, biochemistry may be divided into three fields: structural biology, enzymology, and metabolism. Over the last decades of the 20th century, biochemistry has become successful at explaining living processes through these three disciplines. Almost all areas of the life sciences are being uncovered and developed through biochemical methodology and research. Biochemistry focuses on understanding the chemical basis which allows biological molecules to give rise to the processes that occur within living cells and between cells, in turn relating greatly to the understanding of tissues and organs, as well as organism structure and function. Biochemistry is closely related to molecular biology, which is the study of the molecular mechanisms of biological phenomena.
Much of biochemistry deals with the structures, bonding, functions, and interactions of biological macromolecules, such as proteins, nucleic acids, carbohydrates, and lipids. They provide the structure of cells and perform many of the functions associated with life. The chemistry of the cell also depends upon the reactions of small molecules and ions. These can be inorganic (for example, water and metal ions) or organic (for example, the amino acids, which are used to synthesize proteins). The mechanisms used by cells to harness energy from their environment via chemical reactions are known as metabolism. The findings of biochemistry are applied primarily in medicine, nutrition and agriculture. In medicine, biochemists investigate the causes and cures of diseases. Nutrition studies how to maintain health and wellness and also the effects of nutritional deficiencies. In agriculture, biochemists investigate soil and fertilizers, with the goal of improving crop cultivation, crop storage, and pest control. In recent decades, biochemical principles a
Document 3:::
When pyruvate or lactate is used as the precursor for glycerol 3-phosphate, glyceroneogenesis follows the same pathway as gluconeogenesis until it gene
Document 4:::
Ethanol fermentation, also called alcoholic fermentation, is a biological process which converts sugars such as glucose, fructose, and sucrose into cellular energy, producing ethanol and carbon dioxide as by-products. Because yeasts perform this conversion in the absence of oxygen, alcoholic fermentation is considered an anaerobic process. It also takes place in some species of fish (including goldfish and carp) where (along with lactic acid fermentation) it provides energy when oxygen is scarce.
Ethanol fermentation is the basis for alcoholic beverages, ethanol fuel and bread dough rising.
Biochemical process of fermentation of sucrose
The chemical equations below summarize the fermentation of sucrose (C12H22O11) into ethanol (C2H5OH). Alcoholic fermentation converts one mole of glucose into two moles of ethanol and two moles of carbon dioxide, producing two moles of ATP in the process.
C6H12O6 → 2 C2H5OH + 2 CO2
Sucrose is a sugar composed of a glucose linked to a fructose. In the first step of alcoholic fermentation, the enzyme invertase cleaves the glycosidic linkage between the glucose and fructose molecules.
C12H22O11 + H2O + invertase → 2 C6H12O6
Next, each glucose molecule is broken down into two pyruvate molecules in a process known as glycolysis. Glycolysis is summarized by the equation:
C6H12O6 + 2 ADP + 2 Pi + 2 NAD+ → 2 CH3COCOO− + 2 ATP + 2 NADH + 2 H2O + 2 H+
CH3COCOO− is pyruvate, and Pi is inorganic phosphate. Finally, pyruvate is converted to ethanol and CO2 in two steps, regenerating oxidized NAD+ needed for glycolysis:
1. CH3COCOO− + H+ → CH3CHO + CO2
catalyzed by pyruvate decarboxylase
2. CH3CHO + NADH + H+ → C2H5OH + NAD+
This reaction is catalyzed by alcohol dehydrogenase (ADH1 in baker's yeast).
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
For each initial glucose molecules, two pyruvate molecules will enter where?
A. the plasma
B. Golgi apparatus
C. the nucleus
D. the mitochondria
Answer:
|
|
sciq-3676
|
multiple_choice
|
The largest population size that can be supported in an area without harming the environment is referred to as what?
|
[
"population density",
"carrying capacity",
"containing capacity",
"population shift"
] |
B
|
Relavent Documents:
Document 0:::
Outline of demography contains human demography and population related important concepts and high-level aggregated lists compiled in the useful categories.
The subheadings have been grouped by the following 4 categories:
Meta (lit. "highest" level) units, such as the universal important concepts related to demographics and places.
Macro (lit. "high" level) units where the "whole world" is the smallest unit of measurement, such as the aggregated summary demographics at global level. For example, United Nations.
Meso (lit. "middle" or "intermediate" level) units where the smallest unit of measurement cover more than one nation and more than one continent but not all the nations or continents. For example, summary list at continental level, e.g. Eurasia and Latin America or Middle East which cover two or more continents. Other examples include the intercontinental organisations e.g. the Commonwealth of Nations or the organisation of Arab states.
Micro (lit. "lower" or "smaller") level units where country is the smallest unit of measurement, such as the "globally aggregated lists" by the "individual countries" .
Please do not add sections on the items that are the nano (lit. "minor" or "tiny") level units as per the context described above, e.g. list of things within a city must be kept out.
Meta or important concepts
Global human population
World population
Demographics of the world
Fertility and intelligence
Human geography
Geographic mobility
Globalization
Human migration
List of lists on linguistics
Impact of human population
Human impact on the environment
Biological dispersal
Carrying capacity
Doomsday argument
Environmental migrant
Human overpopulation
Malthusian catastrophe
List of countries by carbon dioxide emissions
List of countries by carbon dioxide emissions per capita
List of countries by greenhouse gas emissions
List of countries by greenhouse gas emissions per capita
Overconsumption
Overexploitation
Population eco
Document 1:::
Demography (), also known as Demographics, is the statistical study of populations, especially human beings.
Demographic analysis examines and measures the dimensions and dynamics of populations; it can cover whole societies or groups defined by criteria such as education, nationality, religion, and ethnicity. Educational institutions usually treat demography as a field of sociology, though there are a number of independent demography departments. These methods have primarily been developed to study human populations, but are extended to a variety of areas where researchers want to know how populations of social actors can change across time through processes of birth, death, and migration. In the context of human biological populations, demographic analysis uses administrative records to develop an independent estimate of the population. Demographic analysis estimates are often considered a reliable standard for judging the accuracy of the census information gathered at any time. In the labor force, demographic analysis is used to estimate sizes and flows of populations of workers; in population ecology the focus is on the birth, death, migration and immigration of individuals in a population of living organisms, alternatively, in social human sciences could involve movement of firms and institutional forms. Demographic analysis is used in a wide variety of contexts. For example, it is often used in business plans, to describe the population connected to the geographic location of the business. Demographic analysis is usually abbreviated as DA. For the 2010 U.S. Census, The U.S. Census Bureau has expanded its DA categories. Also as part of the 2010 U.S. Census, DA now also includes comparative analysis between independent housing estimates, and census address lists at different key time points.
Patient demographics form the core of the data for any medical institution, such as patient and emergency contact information and patient medical record data. They allo
Document 2:::
Population density (in agriculture: standing stock or plant density) is a measurement of population per unit land area. It is mostly applied to humans, but sometimes to other living organisms too. It is a key geographical term.
Biological population densities
Population density is population divided by total land area, sometimes including seas and oceans, as appropriate.
Low densities may cause an extinction vortex and further reduce fertility. This is called the Allee effect after the scientist who identified it. Examples of the causes of reduced fertility in low population densities are:
Increased problems with locating sexual mates
Increased inbreeding
===Human densities===
Population density is the number of people per unit of area, usually transcribed as "per square kilometer" or square mile, and which may include or exclude, for example, areas of water or glaciers. Commonly this is calculated for a county, city, country, another territory or the entire world.
The world's population is around 8,000,000,000 and the Earth's total area (including land and water) is . Therefore, from this very crude type of calculation, the worldwide human population density is approximately 8,000,000,000 ÷ 510,000,000 = . However, if only the Earth's land area of is taken into account, then human population density is . This includes all continental and island land area, including Antarctica. However, if Antarctica is excluded, then population density rises to over .
The European Commission's Joint Research Centre (JRC) has developed a suite of (open and free) data and tools named the Global Human Settlement Layer (GHSL) to improve the science for policy support to the European Commission Directorate Generals and Services and as support to the United Nations system.
Several of the most densely populated territories in the world are city-states, microstates and urban dependencies. In fact, 95% of the world's population is concentrated on just 10% of the world's land.
Document 3:::
The carrying capacity of an environment is the maximum population size of a biological species that can be sustained by that specific environment, given the food, habitat, water, and other resources available. The carrying capacity is defined as the environment's maximal load, which in population ecology corresponds to the population equilibrium, when the number of deaths in a population equals the number of births (as well as immigration and emigration). The effect of carrying capacity on population dynamics is modelled with a logistic function. Carrying capacity is applied to the maximum population an environment can support in ecology, agriculture and fisheries. The term carrying capacity has been applied to a few different processes in the past before finally being applied to population limits in the 1950s. The notion of carrying capacity for humans is covered by the notion of sustainable population.
At the global scale, scientific data indicates that humans are living beyond the carrying capacity of planet Earth and that this cannot continue indefinitely. This scientific evidence comes from many sources worldwide. It was presented in detail in the Millennium Ecosystem Assessment of 2005, a collaborative effort involving more than 1,360 experts worldwide. More recent, detailed accounts are provided by ecological footprint accounting, and interdisciplinary research on planetary boundaries to safe human use of the biosphere. The Sixth Assessment Report on Climate Change from the IPCC and the First Assessment Report on Biodiversity and Ecosystem Services by the IPBES, large international summaries of the state of scientific knowledge regarding climate disruption and biodiversity loss, also support this view.
An early detailed examination of global limits was published in the 1972 book Limits to Growth, which has prompted follow-up commentary and analysis. A 2012 review in Nature by 22 international researchers expressed concerns that the Earth may be "approaching
Document 4:::
The Limits to Growth (LTG) is a 1972 report that discussed the possibility of exponential economic and population growth with finite supply of resources, studied by computer simulation. The study used the World3 computer model to simulate the consequence of interactions between the Earth and human systems. The model was based on the work of Jay Forrester of MIT, as described in his book World Dynamics.
Commissioned by the Club of Rome, the findings of the study were first presented at international gatherings in Moscow and Rio de Janeiro in the summer of 1971. The report's authors are Donella H. Meadows, Dennis L. Meadows, Jørgen Randers, and William W. Behrens III, representing a team of 17 researchers.
The report's findings suggest that in the absence of significant alterations in resource utilization, it is highly likely that there would be an abrupt and unmanageable decrease in both population and industrial capacity. Despite facing severe criticism and scrutiny upon its initial release, subsequent research aimed at verifying its predictions consistently supports the notion that there have been inadequate modifications made since 1972 to substantially alter its essence.
Since its publication, some 30 million copies of the book in 30 languages have been purchased. It continues to generate debate and has been the subject of several subsequent publications.
Beyond the Limits and The Limits to Growth: The 30-Year Update were published in 1992 and 2004 respectively, in 2012, a 40-year forecast from Jørgen Randers, one of the book's original authors, was published as 2052: A Global Forecast for the Next Forty Years, and in 2022 two of the original Limits to Growth authors, Dennis Meadows and Jørgen Randers, joined 19 other contributors to produce Limits and Beyond.
Purpose
In commissioning the MIT team to undertake the project that resulted in LTG, the Club of Rome had three objectives:
Gain insights into the limits of our world system and the constraints it put
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
The largest population size that can be supported in an area without harming the environment is referred to as what?
A. population density
B. carrying capacity
C. containing capacity
D. population shift
Answer:
|
|
sciq-5842
|
multiple_choice
|
What type of cells capture light energy, and use carbon dioxide as their carbon source?
|
[
"sporozoans",
"blood cells",
"photoautotrophs",
"chloroplasts"
] |
C
|
Relavent Documents:
Document 0:::
The evolution of photosynthesis refers to the origin and subsequent evolution of photosynthesis, the process by which light energy is used to assemble sugars from carbon dioxide and a hydrogen and electron source such as water. The process of photosynthesis was discovered by Jan Ingenhousz, a Dutch-born British physician and scientist, first publishing about it in 1779.
The first photosynthetic organisms probably evolved early in the evolutionary history of life and most likely used reducing agents such as hydrogen rather than water. There are three major metabolic pathways by which photosynthesis is carried out: C3 photosynthesis, C4 photosynthesis, and CAM photosynthesis. C3 photosynthesis is the oldest and most common form. A C3 plant uses the Calvin cycle for the initial steps that incorporate into organic material. A C4 plant prefaces the Calvin cycle with reactions that incorporate into four-carbon compounds. A CAM plant uses crassulacean acid metabolism, an adaptation for photosynthesis in arid conditions. C4 and CAM plants have special adaptations that save water.
Origin
Available evidence from geobiological studies of Archean (>2500 Ma) sedimentary rocks indicates that life existed 3500 Ma. Fossils of what are thought to be filamentous photosynthetic organisms have been dated at 3.4 billion years old, consistent with recent studies of photosynthesis. Early photosynthetic systems, such as those from green and purple sulfur and green and purple nonsulfur bacteria, are thought to have been anoxygenic, using various molecules as electron donors. Green and purple sulfur bacteria are thought to have used hydrogen and hydrogen sulfide as electron and hydrogen donors. Green nonsulfur bacteria used various amino and other organic acids. Purple nonsulfur bacteria used a variety of nonspecific organic and inorganic molecules. It is suggested that photosynthesis likely originated at low-wavelength geothermal light from acidic hydrothermal vents, Zn-tetrapyrroles w
Document 1:::
{{DISPLAYTITLE: C3 carbon fixation}}
carbon fixation is the most common of three metabolic pathways for carbon fixation in photosynthesis, the other two being and CAM. This process converts carbon dioxide and ribulose bisphosphate (RuBP, a 5-carbon sugar) into two molecules of 3-phosphoglycerate through the following reaction:
CO2 + H2O + RuBP → (2) 3-phosphoglycerate
This reaction was first discovered by Melvin Calvin, Andrew Benson and James Bassham in 1950. C3 carbon fixation occurs in all plants as the first step of the Calvin–Benson cycle. (In and CAM plants, carbon dioxide is drawn out of malate and into this reaction rather than directly from the air.)
Plants that survive solely on fixation ( plants) tend to thrive in areas where sunlight intensity is moderate, temperatures are moderate, carbon dioxide concentrations are around 200 ppm or higher, and groundwater is plentiful. The plants, originating during Mesozoic and Paleozoic eras, predate the plants and still represent approximately 95% of Earth's plant biomass, including important food crops such as rice, wheat, soybeans and barley.
plants cannot grow in very hot areas at today's atmospheric CO2 level (significantly depleted during hundreds of millions of years from above 5000 ppm) because RuBisCO incorporates more oxygen into RuBP as temperatures increase. This leads to photorespiration (also known as the oxidative photosynthetic carbon cycle, or C2 photosynthesis), which leads to a net loss of carbon and nitrogen from the plant and can therefore limit growth.
plants lose up to 97% of the water taken up through their roots by transpiration. In dry areas, plants shut their stomata to reduce water loss, but this stops from entering the leaves and therefore reduces the concentration of in the leaves. This lowers the :O2 ratio and therefore also increases photorespiration. and CAM plants have adaptations that allow them to survive in hot and dry areas, and they can therefore out-compete
Document 2:::
Giant retinal ganglion cells are photosensitive ganglion cells with large dendritic trees discovered in the human and macaque retina by Dacey et al. (2005).
Giant retinal ganglion cells contain a photo-pigment, melanopsin, allowing them to respond directly to light. They also receive connections from rods and cones, allowing them to encode colour and spatial information. Dacey et al. found the giants' receptive field sizes to be about three times the diameter of those of parasol ganglion cells.
When a giant is responding directly to light, Dacey et al. found its spectral sensitivity function to be similar in shape to those of rods and cones, but with a peak at 482 nm, in between S cones and rods. Dacey et al. also found giants' dynamic range to be 3-4 log units, far larger than any other photoreceptor type's and covering nearly the entire range of illuminations of natural daylight. Under naturalistic lighting conditions, responses to the rods and cones are superimposed on the melanopsin response of giant retinal ganglion cells. Giants encode colour via an S-Off, (L + M)-On opponency. Their spatial modulation transfer function is low-pass, with an upper limit of about 0.6 cycles per degree.
Dacey et al. propose that the giants subserve the subconscious, 'non-image-forming' functions of circadian photoentrainment and pupillary diameter, and via the rod and cone inputs, may help mediate conscious perception of irradiance.
Human eye anatomy
Histology
Circadian rhythm
Document 3:::
Disc shedding is the process by which photoreceptor cells in the retina are renewed. The disc formations in the outer segment of photoreceptors, which contain the photosensitive opsins, are completely renewed every ten days.
Photoreceptors
The retina contains two types of photoreceptor – rod cells and cone cells. There are about 6-7 million cones that mediate photopic vision, and they are concentrated in the macula at the center of the retina. There are about 120 million rods that are more sensitive than the cones and therefore mediate scotopic vision.
A vertebrate's photoreceptors are divided into three parts:
an outer segment that contains the photosensitive opsins
an inner segment that contains the cell's metabolic machinery (endoplasmic reticulum, Golgi complex, ribosomes, mitochondria)
a synaptic terminal at which contacts with second-order neurons of the retina are made
Discs
The photosensitive outer segment consists of a series of discrete membranous discs .
While in the rod, these discs lack any direct connection to the surface membrane (with the exception of a few recently formed basal discs that remain in continuity with the surface), the cone's photosensitive membrane is continuous with the surface membrane. The outer segment (OS) discs are densely packed with rhodopsin for high-sensitivity light detection. These discs are completely replaced once every ten days and this continuous renewal continues throughout the lifetime of the sighted animal.
After the opsins are synthesized, they fuse to the plasma membrane, which then invaginates with discs budding off internally, forming the tightly packed stacks of outer segment discs. From translation of opsin to formation of the discs takes just a couple of hours.
Shedding
Disc shedding was first described by RW Young in 1967. Discs mature along with their distal migration; aged discs shed at the distal tip and are engulfed by the neighboring retinal pigment epithelial (RPE) cells for degradation.
One e
Document 4:::
The cell is the basic structural and functional unit of all forms of life. Every cell consists of cytoplasm enclosed within a membrane, and contains many macromolecules such as proteins, DNA and RNA, as well as many small molecules of nutrients and metabolites. The term comes from the Latin word meaning 'small room'.
Cells can acquire specified function and carry out various tasks within the cell such as replication, DNA repair, protein synthesis, and motility. Cells are capable of specialization and mobility within the cell.
Most plant and animal cells are only visible under a light microscope, with dimensions between 1 and 100 micrometres. Electron microscopy gives a much higher resolution showing greatly detailed cell structure. Organisms can be classified as unicellular (consisting of a single cell such as bacteria) or multicellular (including plants and animals). Most unicellular organisms are classed as microorganisms.
The study of cells and how they work has led to many other studies in related areas of biology, including: discovery of DNA, cancer systems biology, aging and developmental biology.
Cell biology is the study of cells, which were discovered by Robert Hooke in 1665, who named them for their resemblance to cells inhabited by Christian monks in a monastery. Cell theory, first developed in 1839 by Matthias Jakob Schleiden and Theodor Schwann, states that all organisms are composed of one or more cells, that cells are the fundamental unit of structure and function in all living organisms, and that all cells come from pre-existing cells. Cells emerged on Earth about 4 billion years ago.
Discovery
With continual improvements made to microscopes over time, magnification technology became advanced enough to discover cells. This discovery is largely attributed to Robert Hooke, and began the scientific study of cells, known as cell biology. When observing a piece of cork under the scope, he was able to see pores. This was shocking at the time as i
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What type of cells capture light energy, and use carbon dioxide as their carbon source?
A. sporozoans
B. blood cells
C. photoautotrophs
D. chloroplasts
Answer:
|
|
sciq-10405
|
multiple_choice
|
Echinoderms lack what type of system that in humans consists of the brain and spinal cord?
|
[
"peripheral nervous system",
"endocrine system",
"autonomic system",
"centralized nervous system"
] |
D
|
Relavent Documents:
Document 0:::
The following diagram is provided as an overview of and topical guide to the human nervous system:
Human nervous system – the part of the human body that coordinates a person's voluntary and involuntary actions and transmits signals between different parts of the body. The human nervous system consists of two main parts: the central nervous system (CNS) and the peripheral nervous system (PNS). The CNS contains the brain and spinal cord. The PNS consists mainly of nerves, which are long fibers that connect the CNS to every other part of the body. The PNS includes motor neurons, mediating voluntary movement; the autonomic nervous system, comprising the sympathetic nervous system and the parasympathetic nervous system and regulating involuntary functions; and the enteric nervous system, a semi-independent part of the nervous system whose function is to control the gastrointestinal system.
Evolution of the human nervous system
Evolution of nervous systems
Evolution of human intelligence
Evolution of the human brain
Paleoneurology
Some branches of science that study the human nervous system
Neuroscience
Neurology
Paleoneurology
Central nervous system
The central nervous system (CNS) is the largest part of the nervous system and includes the brain and spinal cord.
Spinal cord
Brain
Brain – center of the nervous system.
Outline of the human brain
List of regions of the human brain
Principal regions of the vertebrate brain:
Peripheral nervous system
Peripheral nervous system (PNS) – nervous system structures that do not lie within the CNS.
Sensory system
A sensory system is a part of the nervous system responsible for processing sensory information. A sensory system consists of sensory receptors, neural pathways, and parts of the brain involved in sensory perception.
List of sensory systems
Sensory neuron
Perception
Visual system
Auditory system
Somatosensory system
Vestibular system
Olfactory system
Taste
Pain
Components of the nervous system
Neuron
I
Document 1:::
The following are two lists of animals ordered by the size of their nervous system. The first list shows number of neurons in their entire nervous system, indicating their overall neural complexity. The second list shows the number of neurons in the structure that has been found to be representative of animal intelligence. The human brain contains 86 billion neurons, with 16 billion neurons in the cerebral cortex.
Scientists are engaged in counting, quantification, in order to find answers to the question in the strategy of neuroscience and intelligence of "self-knowledge": how the evolution of a set of components and parameters (~1011 neurons, ~1014 synapses) of a complex system could lead to the phenomenon of the appearance of intelligence in the biological species "sapiens".
Overview
Neurons are the cells that transmit information in an animal's nervous system so that it can sense stimuli from its environment and behave accordingly. Not all animals have neurons; Trichoplax and sponges lack nerve cells altogether.
Neurons may be packed to form structures such as the brain of vertebrates or the neural ganglions of insects.
The number of neurons and their relative abundance in different parts of the brain is a determinant of neural function and, consequently, of behavior.
Whole nervous system
All numbers for neurons (except Caenorhabditis and Ciona), and all numbers for synapses (except Ciona) are estimations.
List of animal species by forebrain (cerebrum or pallium) neuron number
The question of what physical characteristic of an animal makes an animal intelligent has varied over the centuries. One early speculation was brain size (or weight, which provides the same ordering.) A second proposal was brain-to-body-mass ratio, and a third was encephalization quotient, sometimes referred to as EQ. The current best predictor is number of neurons in the forebrain, based on Herculano-Houzel's improved neuron counts. It accounts most accurately for variations
Document 2:::
The following outline is provided as an overview of and topical guide to neuroscience:
Neuroscience is the scientific study of the structure and function of the nervous system. It encompasses the branch of biology that deals with the anatomy, biochemistry, molecular biology, and physiology of neurons and neural circuits. It also encompasses cognition, and human behavior. Neuroscience has multiple concepts that each relate to learning abilities and memory functions. Additionally, the brain is able to transmit signals that cause conscious/unconscious behaviors that are responses verbal or non-verbal. This allows people to communicate with one another.
Branches of neuroscience
Neurophysiology
Neurophysiology is the study of the function (as opposed to structure) of the nervous system.
Brain mapping
Electrophysiology
Extracellular recording
Intracellular recording
Brain stimulation
Electroencephalography
Intermittent rhythmic delta activity
:Category: Neurophysiology
:Category: Neuroendocrinology
:Neuroendocrinology
Neuroanatomy
Neuroanatomy is the study of the anatomy of nervous tissue and neural structures of the nervous system.
Immunostaining
:Category: Neuroanatomy
Neuropharmacology
Neuropharmacology is the study of how drugs affect cellular function in the nervous system.
Drug
Psychoactive drug
Anaesthetic
Narcotic
Behavioral neuroscience
Behavioral neuroscience, also known as biological psychology, biopsychology, or psychobiology, is the application of the principles of biology to the study of mental processes and behavior in human and non-human animals.
Neuroethology
Developmental neuroscience
Developmental neuroscience aims to describe the cellular basis of brain development and to address the underlying mechanisms. The field draws on both neuroscience and developmental biology to provide insight into the cellular and molecular mechanisms by which complex nervous systems develop.
Aging and memory
Cognitive neuroscience
Cognitive ne
Document 3:::
Neuroecology studies ways in which the structure and function of the brain results from adaptations to a specific habitat and niche.
It integrates the multiple disciplines of neuroscience, which examines the biological basis of cognitive and emotional processes, such as perception, memory, and decision-making, with the field of ecology, which studies the relationship between living organisms and their physical environment.
In biology, the term 'adaptation' signifies the way evolutionary processes enhance an organism's fitness to survive within a specific ecological context. This fitness includes the development of physical, cognitive, and emotional adaptations specifically suited to the environmental conditions in which the organism or phenotype lives, and in which its species or genotype evolves.
Neuroecology concentrates specifically on neurological adaptations, particularly those of the brain. The purview of this study encompasses two areas. Firstly, neuroecology studies how the physical structure and functional activity of neural networks in a phenotype is influenced by characteristics of the environmental context. This includes the way social stressors, interpersonal relationships, and physical conditions precipitate persistent alterations in the individual brain, providing the neural correlates of cognitive and emotional responses. Secondly, neuroecology studies how neural structure and activity common to a genotype is determined by natural selection of traits that benefit survival and reproduction in a specific environment.
See also
Evolutionary ecology
Evolutionary psychology
Document 4:::
Several universities have designed interdisciplinary courses with a focus on human biology at the undergraduate level. There is a wide variation in emphasis ranging from business, social studies, public policy, healthcare and pharmaceutical research.
Americas
Human Biology major at Stanford University, Palo Alto (since 1970)
Stanford's Human Biology Program is an undergraduate major; it integrates the natural and social sciences in the study of human beings. It is interdisciplinary and policy-oriented and was founded in 1970 by a group of Stanford faculty (Professors Dornbusch, Ehrlich, Hamburg, Hastorf, Kennedy, Kretchmer, Lederberg, and Pittendrigh). It is a very popular major and alumni have gone to post-graduate education, medical school, law, business and government.
Human and Social Biology (Caribbean)
Human and Social Biology is a Level 4 & 5 subject in the secondary and post-secondary schools in the Caribbean and is optional for the Caribbean Secondary Education Certification (CSEC) which is equivalent to Ordinary Level (O-Level) under the British school system. The syllabus centers on structure and functioning (anatomy, physiology, biochemistry) of human body and the relevance to human health with Caribbean-specific experience. The syllabus is organized under five main sections: Living organisms and the environment, life processes, heredity and variation, disease and its impact on humans, the impact of human activities on the environment.
Human Biology Program at University of Toronto
The University of Toronto offers an undergraduate program in Human Biology that is jointly offered by the Faculty of Arts & Science and the Faculty of Medicine. The program offers several major and specialist options in: human biology, neuroscience, health & disease, global health, and fundamental genetics and its applications.
Asia
BSc (Honours) Human Biology at All India Institute of Medical Sciences, New Delhi (1980–2002)
BSc (honours) Human Biology at AIIMS (New
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Echinoderms lack what type of system that in humans consists of the brain and spinal cord?
A. peripheral nervous system
B. endocrine system
C. autonomic system
D. centralized nervous system
Answer:
|
|
sciq-10057
|
multiple_choice
|
What is the water that falls from the sky called?
|
[
"condensation",
"moisture",
"precipitation",
"evaporation"
] |
C
|
Relavent Documents:
Document 0:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 1:::
Evapotranspiration (ET) is the combined processes which move water from the Earth's surface into the atmosphere. It covers both water evaporation (movement of water to the air directly from soil, canopies, and water bodies) and transpiration (evaporation that occurs through the stomata, or openings, in plant leaves). Evapotranspiration is an important part of the local water cycle and climate, and measurement of it plays a key role in agricultural irrigation and water resource management.
Definition of evapotranspiration
Evapotranspiration is a combination of evaporation and transpiration, measured in order to better understand crop water requirements, irrigation scheduling, and watershed management. The two key components of evapotranspiration are:
Evaporation: the movement of water directly to the air from sources such as the soil and water bodies. It can be affected by factors including heat, humidity, solar radiation and wind speed.
Transpiration: the movement of water from root systems, through a plant, and exit into the air as water vapor. This exit occurs through stomata in the plant. Rate of transpiration can be influenced by factors including plant type, soil type, weather conditions and water content, and also cultivation practices.
Evapotranspiration is typically measured in millimeters of water (i.e. volume of water moved per unit area of the Earth's surface) in a set unit of time. Globally, it is estimated that on average between three-fifths and three-quarters of land precipitation is returned to the atmosphere via evapotranspiration.
Evapotranspiration does not, in general, account for other mechanisms which are involved in returning water to the atmosphere, though some of these, such as snow and ice sublimation in regions of high elevation or high latitude, can make a large contribution to atmospheric moisture even under standard conditions.
Factors that impact evapotranspiration levels
Primary factors
Because evaporation and transpiration
Document 2:::
In mathematical psychology and education theory, a knowledge space is a combinatorial structure used to formulate mathematical models describing the progression of a human learner. Knowledge spaces were introduced in 1985 by Jean-Paul Doignon and Jean-Claude Falmagne, and remain in extensive use in the education theory. Modern applications include two computerized tutoring systems, ALEKS and the defunct RATH.
Formally, a knowledge space assumes that a domain of knowledge is a collection of concepts or skills, each of which must be eventually mastered. Not all concepts are interchangeable; some require other concepts as prerequisites. Conversely, competency at one skill may ease the acquisition of another through similarity. A knowledge space marks out which collections of skills are feasible: they can be learned without mastering any other skills. Under reasonable assumptions, the collection of feasible competencies forms the mathematical structure known as an antimatroid.
Researchers and educators usually explore the structure of a discipline's knowledge space as a latent class model.
Motivation
Knowledge Space Theory attempts to address shortcomings of standardized testing when used in educational psychometry. Common tests, such as the SAT and ACT, compress a student's knowledge into a very small range of ordinal ranks, in the process effacing the conceptual dependencies between questions. Consequently, the tests cannot distinguish between true understanding and guesses, nor can they identify a student's particular weaknesses, only the general proportion of skills mastered. The goal of knowledge space theory is to provide a language by which exams can communicate
What the student can do and
What the student is ready to learn.
Model structure
Knowledge Space Theory-based models presume that an educational subject can be modeled as a finite set of concepts, skills, or topics. Each feasible state of knowledge about is then a subset of ; the set of
Document 3:::
Surface runoff (also known as overland flow or terrestrial runoff) is the unconfined flow of water over the ground surface, in contrast to channel runoff (or stream flow). It occurs when excess rainwater, stormwater, meltwater, or other sources, can no longer sufficiently rapidly infiltrate in the soil. This can occur when the soil is saturated by water to its full capacity, and the rain arrives more quickly than the soil can absorb it. Surface runoff often occurs because impervious areas (such as roofs and pavement) do not allow water to soak into the ground. Furthermore, runoff can occur either through natural or human-made processes.
Surface runoff is a major component of the water cycle. It is the primary agent of soil erosion by water. The land area producing runoff that drains to a common point is called a drainage basin.
Runoff that occurs on the ground surface before reaching a channel can be a nonpoint source of pollution, as it can carry human-made contaminants or natural forms of pollution (such as rotting leaves). Human-made contaminants in runoff include petroleum, pesticides, fertilizers and others. Much agricultural pollution is exacerbated by surface runoff, leading to a number of down stream impacts, including nutrient pollution that causes eutrophication.
In addition to causing water erosion and pollution, surface runoff in urban areas is a primary cause of urban flooding, which can result in property damage, damp and mold in basements, and street flooding.
Generation
Surface runoff is defined as precipitation (rain, snow, sleet, or hail) that reaches a surface stream without ever passing below the soil surface. It is distinct from direct runoff, which is runoff that reaches surface streams immediately after rainfall or melting snowfall and excludes runoff generated by the melting of snowpack or glaciers.
Snow and glacier melt occur only in areas cold enough for these to form permanently. Typically snowmelt will peak in the spring and glacie
Document 4:::
A pre-STEM program is a course of study at any two-year college that prepares a student to transfer to a four-year school to earn a bachelor's degree in a STEM field.
Overview
The concept of a pre-STEM program is being developed to address America's need for more college-trained professionals in science, technology, engineering, and mathematics (STEM). It is an innovation meant to fill a gap at community colleges that do not have 'major' degree paths that students identify with on their way to earning an Associates degree. Students must complete a considerable amount of STEM coursework before transferring from a two-year school to a four-year school and earn a baccalaureate degree in a STEM field. Schools with a pre-STEM program are able to identify those students and support them with STEM-specific academic and career advising, increasing the student's chances of going on to earn a STEM baccalaureate degree in a timely fashion.
With over 50% of America's college-bound students starting their college career at public or private two-year school, and with a very small proportion of students who start college at a two-year school matriculating to and earning STEM degrees from four-year schools, pre-STEM programs have great potential for broadening participation in baccalaureate STEM studies.
Example programs
The effectiveness of pre-STEM programs is being investigated by a consortium of schools in Missouri: Moberly Area Community College, St. Charles Community College, Metropolitan Community College, and Truman State University.
A larger group of schools met at the Belknap Springs Meetings in October 2009 to discuss the challenges and opportunities presented by STEM-focused partnerships between 2-year and 4-year schools. Each program represented a two-year school and a four-year school that were trying to increase the number of people who earn a baccalaureate degree in a STEM area through various means, some of which were pre-STEM programs. Other methods includes
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is the water that falls from the sky called?
A. condensation
B. moisture
C. precipitation
D. evaporation
Answer:
|
|
sciq-1385
|
multiple_choice
|
What consists of one or two closed loops through which current can flow?
|
[
"magnetic poles",
"cooling circuit",
"thermal reaction",
"electric circuit"
] |
D
|
Relavent Documents:
Document 0:::
Mathematical methods are integral to the study of electronics.
Mathematics in electronics
Electronics engineering careers usually include courses in calculus (single and multivariable), complex analysis, differential equations (both ordinary and partial), linear algebra and probability. Fourier analysis and Z-transforms are also subjects which are usually included in electrical engineering programs. Laplace transform can simplify computing RLC circuit behaviour.
Basic applications
A number of electrical laws apply to all electrical networks. These include
Faraday's law of induction: Any change in the magnetic environment of a coil of wire will cause a voltage (emf) to be "induced" in the coil.
Gauss's Law: The total of the electric flux out of a closed surface is equal to the charge enclosed divided by the permittivity.
Kirchhoff's current law: the sum of all currents entering a node is equal to the sum of all currents leaving the node or the sum of total current at a junction is zero
Kirchhoff's voltage law: the directed sum of the electrical potential differences around a circuit must be zero.
Ohm's law: the voltage across a resistor is the product of its resistance and the current flowing through it.at constant temperature.
Norton's theorem: any two-terminal collection of voltage sources and resistors is electrically equivalent to an ideal current source in parallel with a single resistor.
Thévenin's theorem: any two-terminal combination of voltage sources and resistors is electrically equivalent to a single voltage source in series with a single resistor.
Millman's theorem: the voltage on the ends of branches in parallel is equal to the sum of the currents flowing in every branch divided by the total equivalent conductance.
See also Analysis of resistive circuits.
Circuit analysis is the study of methods to solve linear systems for an unknown variable.
Circuit analysis
Components
There are many electronic components currently used and they all have thei
Document 1:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 2:::
Advanced Placement (AP) Physics C: Electricity and Magnetism (also known as AP Physics C: E&M or AP E&M) is an introductory physics course administered by the College Board as part of its Advanced Placement program. It is intended to proxy a second-semester calculus-based university course in electricity and magnetism. The content of Physics C: E&M overlaps with that of AP Physics 2, but Physics 2 is algebra-based and covers other topics outside of electromagnetism, while Physics C is calculus-based and only covers electromagnetism. Physics C: E&M may be combined with its mechanics counterpart to form a year-long course that prepares for both exams.
Course content
E&M is equivalent to an introductory college course in electricity and magnetism for physics or engineering majors. The course modules are:
Electrostatics
Conductors, capacitors, and dielectrics
Electric circuits
Magnetic fields
Electromagnetism.
Methods of calculus are used wherever appropriate in formulating physical principles and in applying them to physical problems. Therefore, students should have completed or be concurrently enrolled in a calculus class.
AP test
The course culminates in an optional exam for which high-performing students may receive some credit towards their college coursework, depending on the institution.
Registration
The AP examination for AP Physics C: Electricity and Magnetism is separate from the AP examination for AP Physics C: Mechanics. Before 2006, test-takers paid only once and were given the choice of taking either one or two parts of the Physics C test.
Format
The exam is typically administered on a Monday afternoon in May. The exam is configured in two categories: a 35-question multiple choice section and a 3-question free response section. Test takers are allowed to use an approved calculator during the entire exam. The test is weighted such that each section is worth half of the final score. This and AP Physics C: Mechanics are the shortest AP exams, with
Document 3:::
Computer science and engineering (CSE) is an academic program at many universities which comprises computer science classes (e.g. data structures and algorithms) and computer engineering classes (e.g computer architecture). There is no clear division in computing between science and engineering, just like in the field of materials science and engineering. CSE is also a term often used in Europe to translate the name of engineering informatics academic programs. It is offered in both undergraduate as well postgraduate with specializations.
Academic courses
Academic programs vary between colleges, but typically include a combination of topics in computer science, computer engineering, and electrical engineering. Undergraduate courses usually include programming, algorithms and data structures, computer architecture, operating systems, computer networks, parallel computing, embedded systems, algorithms design, circuit analysis and electronics, digital logic and processor design, computer graphics, scientific computing, software engineering, database systems, digital signal processing, virtualization, computer simulations and games programming. CSE programs also include core subjects of theoretical computer science such as theory of computation, numerical methods, machine learning, programming theory and paradigms. Modern academic programs also cover emerging computing fields like image processing, data science, robotics, bio-inspired computing, computational biology, autonomic computing and artificial intelligence. Most CSE programs require introductory mathematical knowledge, hence the first year of study is dominated by mathematical courses, primarily discrete mathematics, mathematical analysis, linear algebra, probability, and statistics, as well as the basics of electrical and electronic engineering, physics, and electromagnetism.
Example universities with CSE majors and departments
APJ Abdul Kalam Technological University
American International University-B
Document 4:::
A pre-STEM program is a course of study at any two-year college that prepares a student to transfer to a four-year school to earn a bachelor's degree in a STEM field.
Overview
The concept of a pre-STEM program is being developed to address America's need for more college-trained professionals in science, technology, engineering, and mathematics (STEM). It is an innovation meant to fill a gap at community colleges that do not have 'major' degree paths that students identify with on their way to earning an Associates degree. Students must complete a considerable amount of STEM coursework before transferring from a two-year school to a four-year school and earn a baccalaureate degree in a STEM field. Schools with a pre-STEM program are able to identify those students and support them with STEM-specific academic and career advising, increasing the student's chances of going on to earn a STEM baccalaureate degree in a timely fashion.
With over 50% of America's college-bound students starting their college career at public or private two-year school, and with a very small proportion of students who start college at a two-year school matriculating to and earning STEM degrees from four-year schools, pre-STEM programs have great potential for broadening participation in baccalaureate STEM studies.
Example programs
The effectiveness of pre-STEM programs is being investigated by a consortium of schools in Missouri: Moberly Area Community College, St. Charles Community College, Metropolitan Community College, and Truman State University.
A larger group of schools met at the Belknap Springs Meetings in October 2009 to discuss the challenges and opportunities presented by STEM-focused partnerships between 2-year and 4-year schools. Each program represented a two-year school and a four-year school that were trying to increase the number of people who earn a baccalaureate degree in a STEM area through various means, some of which were pre-STEM programs. Other methods includes
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What consists of one or two closed loops through which current can flow?
A. magnetic poles
B. cooling circuit
C. thermal reaction
D. electric circuit
Answer:
|
|
sciq-996
|
multiple_choice
|
Protists, fungi, animals, and plants all consist of what kind of cells?
|
[
"eukaryotic",
"parenchyma",
"xylem",
"enzymes"
] |
A
|
Relavent Documents:
Document 0:::
Cellular components are the complex biomolecules and structures of which cells, and thus living organisms, are composed. Cells are the structural and functional units of life. The smallest organisms are single cells, while the largest organisms are assemblages of trillions of cells. DNA, double stranded macromolecule that carries the hereditary information of the cell and found in all living cells; each cell carries chromosome(s) having a distinctive DNA sequence.
Examples include macromolecules such as proteins and nucleic acids, biomolecular complexes such as a ribosome, and structures such as membranes, and organelles. While the majority of cellular components are located within the cell itself, some may exist in extracellular areas of an organism.
Cellular components may also be called biological matter or biological material. Most biological matter has the characteristics of soft matter, being governed by relatively small energies. All known life is made of biological matter. To be differentiated from other theoretical or fictional life forms, such life may be called carbon-based, cellular, organic, biological, or even simply living – as some definitions of life exclude hypothetical types of biochemistry.
See also
Cell (biology)
Cell biology
Biomolecule
Organelle
Tissue (biology)
External links
https://web.archive.org/web/20130918033010/http://bioserv.fiu.edu/~walterm/FallSpring/review1_fall05_chap_cell3.htm
Document 1:::
The cell is the basic structural and functional unit of all forms of life. Every cell consists of cytoplasm enclosed within a membrane, and contains many macromolecules such as proteins, DNA and RNA, as well as many small molecules of nutrients and metabolites. The term comes from the Latin word meaning 'small room'.
Cells can acquire specified function and carry out various tasks within the cell such as replication, DNA repair, protein synthesis, and motility. Cells are capable of specialization and mobility within the cell.
Most plant and animal cells are only visible under a light microscope, with dimensions between 1 and 100 micrometres. Electron microscopy gives a much higher resolution showing greatly detailed cell structure. Organisms can be classified as unicellular (consisting of a single cell such as bacteria) or multicellular (including plants and animals). Most unicellular organisms are classed as microorganisms.
The study of cells and how they work has led to many other studies in related areas of biology, including: discovery of DNA, cancer systems biology, aging and developmental biology.
Cell biology is the study of cells, which were discovered by Robert Hooke in 1665, who named them for their resemblance to cells inhabited by Christian monks in a monastery. Cell theory, first developed in 1839 by Matthias Jakob Schleiden and Theodor Schwann, states that all organisms are composed of one or more cells, that cells are the fundamental unit of structure and function in all living organisms, and that all cells come from pre-existing cells. Cells emerged on Earth about 4 billion years ago.
Discovery
With continual improvements made to microscopes over time, magnification technology became advanced enough to discover cells. This discovery is largely attributed to Robert Hooke, and began the scientific study of cells, known as cell biology. When observing a piece of cork under the scope, he was able to see pores. This was shocking at the time as i
Document 2:::
A cell type is a classification used to identify cells that share morphological or phenotypical features. A multicellular organism may contain cells of a number of widely differing and specialized cell types, such as muscle cells and skin cells, that differ both in appearance and function yet have identical genomic sequences. Cells may have the same genotype, but belong to different cell types due to the differential regulation of the genes they contain. Classification of a specific cell type is often done through the use of microscopy (such as those from the cluster of differentiation family that are commonly used for this purpose in immunology). Recent developments in single cell RNA sequencing facilitated classification of cell types based on shared gene expression patterns. This has led to the discovery of many new cell types in e.g. mouse cortex, hippocampus, dorsal root ganglion and spinal cord.
Animals have evolved a greater diversity of cell types in a multicellular body (100–150 different cell types), compared
with 10–20 in plants, fungi, and protists. The exact number of cell types is, however, undefined, and the Cell Ontology, as of 2021, lists over 2,300 different cell types.
Multicellular organisms
All higher multicellular organisms contain cells specialised for different functions. Most distinct cell types arise from a single totipotent cell that differentiates into hundreds of different cell types during the course of development. Differentiation of cells is driven by different environmental cues (such as cell–cell interaction) and intrinsic differences (such as those caused by the uneven distribution of molecules during division). Multicellular organisms are composed of cells that fall into two fundamental types: germ cells and somatic cells. During development, somatic cells will become more specialized and form the three primary germ layers: ectoderm, mesoderm, and endoderm. After formation of the three germ layers, cells will continue to special
Document 3:::
This list of life sciences comprises the branches of science that involve the scientific study of life – such as microorganisms, plants, and animals including human beings. This science is one of the two major branches of natural science, the other being physical science, which is concerned with non-living matter. Biology is the overall natural science that studies life, with the other life sciences as its sub-disciplines.
Some life sciences focus on a specific type of organism. For example, zoology is the study of animals, while botany is the study of plants. Other life sciences focus on aspects common to all or many life forms, such as anatomy and genetics. Some focus on the micro-scale (e.g. molecular biology, biochemistry) other on larger scales (e.g. cytology, immunology, ethology, pharmacy, ecology). Another major branch of life sciences involves understanding the mindneuroscience. Life sciences discoveries are helpful in improving the quality and standard of life and have applications in health, agriculture, medicine, and the pharmaceutical and food science industries. For example, it has provided information on certain diseases which has overall aided in the understanding of human health.
Basic life science branches
Biology – scientific study of life
Anatomy – study of form and function, in plants, animals, and other organisms, or specifically in humans
Astrobiology – the study of the formation and presence of life in the universe
Bacteriology – study of bacteria
Biotechnology – study of combination of both the living organism and technology
Biochemistry – study of the chemical reactions required for life to exist and function, usually a focus on the cellular level
Bioinformatics – developing of methods or software tools for storing, retrieving, organizing and analyzing biological data to generate useful biological knowledge
Biolinguistics – the study of the biology and evolution of language.
Biological anthropology – the study of humans, non-hum
Document 4:::
Cell physiology is the biological study of the activities that take place in a cell to keep it alive. The term physiology refers to normal functions in a living organism. Animal cells, plant cells and microorganism cells show similarities in their functions even though they vary in structure.
General characteristics
There are two types of cells: prokaryotes and eukaryotes.
Prokaryotes were the first of the two to develop and do not have a self-contained nucleus. Their mechanisms are simpler than later-evolved eukaryotes, which contain a nucleus that envelops the cell's DNA and some organelles.
Prokaryotes
Prokaryotes have DNA located in an area called the nucleoid, which is not separated from other parts of the cell by a membrane. There are two domains of prokaryotes: bacteria and archaea. Prokaryotes have fewer organelles than eukaryotes. Both have plasma membranes and ribosomes (structures that synthesize proteins and float free in cytoplasm). Two unique characteristics of prokaryotes are fimbriae (finger-like projections on the surface of a cell) and flagella (threadlike structures that aid movement).
Eukaryotes
Eukaryotes have a nucleus where DNA is contained. They are usually larger than prokaryotes and contain many more organelles. The nucleus, the feature of a eukaryote that distinguishes it from a prokaryote, contains a nuclear envelope, nucleolus and chromatin. In cytoplasm, endoplasmic reticulum (ER) synthesizes membranes and performs other metabolic activities. There are two types, rough ER (containing ribosomes) and smooth ER (lacking ribosomes). The Golgi apparatus consists of multiple membranous sacs, responsible for manufacturing and shipping out materials such as proteins. Lysosomes are structures that use enzymes to break down substances through phagocytosis, a process that comprises endocytosis and exocytosis. In the mitochondria, metabolic processes such as cellular respiration occur. The cytoskeleton is made of fibers that support the str
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Protists, fungi, animals, and plants all consist of what kind of cells?
A. eukaryotic
B. parenchyma
C. xylem
D. enzymes
Answer:
|
|
sciq-9478
|
multiple_choice
|
What do you call the folded layers of rock where geologists look for oil?
|
[
"crests",
"synclines",
"anticlines",
"oscillations"
] |
C
|
Relavent Documents:
Document 0:::
Rock mechanics is a theoretical and applied science of the mechanical behavior of rocks and rock masses.
Compared to geology, it is the branch of mechanics concerned with the response of rock and rock masses to the force fields of their physical environment.
Background
Rock mechanics is part of a much broader subject of geomechanics, which is concerned with the mechanical responses of all geological materials, including soils.
Rock mechanics is concerned with the application of the principles of engineering mechanics to the design of structures built in or on rock. The structure could include many objects such as a drilling well, a mine shaft, a tunnel, a reservoir dam, a repository component, or a building. Rock mechanics is used in many engineering disciplines, but is primarily used in Mining, Civil, Geotechnical, Transportation, and Petroleum Engineering.
Rock mechanics answers questions such as, "is reinforcement necessary for a rock, or will it be able to handle whatever load it is faced with?" It also includes the design of reinforcement systems, such as rock bolting patterns.
Assessing the Project Site
Before any work begins, the construction site must be investigated properly to inform of the geological conditions of the site. Field observations, deep drilling, and geophysical surveys, can all give necessary information to develop a safe construction plan and create a site geological model. The level of investigation conducted at this site depends on factors such as budget, time frame, and expected geological conditions.
The first step of the investigation is the collection of maps and aerial photos to analyze. This can provide information about potential sinkholes, landslides, erosion, etc. Maps can provide information on the rock type of the site, geological structure, and boundaries between bedrock units.
Boreholes
Creating a borehole is a technique that consists of drilling through the ground in various areas at various depths, to get a bett
Document 1:::
The geologic record in stratigraphy, paleontology and other natural sciences refers to the entirety of the layers of rock strata. That is, deposits laid down by volcanism or by deposition of sediment derived from weathering detritus (clays, sands etc.). This includes all its fossil content and the information it yields about the history of the Earth: its past climate, geography, geology and the evolution of life on its surface. According to the law of superposition, sedimentary and volcanic rock layers are deposited on top of each other. They harden over time to become a solidified (competent) rock column, that may be intruded by igneous rocks and disrupted by tectonic events.
Correlating the rock record
At a certain locality on the Earth's surface, the rock column provides a cross section of the natural history in the area during the time covered by the age of the rocks. This is sometimes called the rock history and gives a window into the natural history of the location that spans many geological time units such as ages, epochs, or in some cases even multiple major geologic periods—for the particular geographic region or regions. The geologic record is in no one place entirely complete for where geologic forces one age provide a low-lying region accumulating deposits much like a layer cake, in the next may have uplifted the region, and the same area is instead one that is weathering and being torn down by chemistry, wind, temperature, and water. This is to say that in a given location, the geologic record can be and is quite often interrupted as the ancient local environment was converted by geological forces into new landforms and features. Sediment core data at the mouths of large riverine drainage basins, some of which go deep thoroughly support the law of superposition.
However using broadly occurring deposited layers trapped within differently located rock columns, geologists have pieced together a system of units covering most of the geologic time scale
Document 2:::
This is a list of free and open-source software for geological data handling and interpretation. The list is split into broad categories, depending on the intended use of the software and its scope of functionality.
Notice that 'free and open-source' requires that the source code is available and users are given a free software license. Simple being 'free of charge' is not sufficient—see gratis versus libre.
Well logging & Borehole visualisation
Geosciences software platforms
Geostatistics
Forward modeling
Geomodeling
Visualization, interpretation & analysis packages
Geographic information systems (GIS)
This important class of tools is already listed in the article List of GIS software.
Not true free and open-source projects
The following projects have unknown licensing, licenses or other conditions which place some restriction on use or redistribution, or which depend on non-open-source software like MATLAB or XVT (and therefore do not meet the Open Source Definition from the Open Source Initiative).
Document 3:::
Shale Gouge Ratio (typically abbreviated to SGR) is a mathematical algorithm that aims to predict the fault rock types for simple fault zones developed in sedimentary sequences dominated by sandstone and shale.
The parameter is widely used in the oil and gas exploration and production industries to enable quantitative predictions to be made regarding the hydrodynamic behavior of faults.
Definition
At any point on a fault surface, the shale gouge ratio is equal to the net shale/clay content of the rocks that have slipped past that point.
The SGR algorithm assumes complete mixing of the wall-rock components in any particular 'throw interval'. The parameter is a measure of the 'upscaled' composition of the fault zone.
Application to hydrocarbon exploration
Hydrocarbon exploration involves identifying and defining accumulations of hydrocarbons that are trapped in subsurface structures. These structures are often segmented by faults. For a thorough trap evaluation, it is necessary to predict whether the fault is sealing or leaking to hydrocarbons and also to provide an estimate of how 'strong' the fault seal might be. The 'strength' of a fault seal can be quantified in terms of subsurface pressure, arising from the buoyancy forces within the hydrocarbon column, that the fault can support before it starts to leak. When acting on a fault zone this subsurface pressure is termed capillary threshold pressure.
For faults developed in sandstone and shale sequences, the first order control on capillary threshold pressure is likely to be the composition, in particular the shale or clay content, of the fault-zone material. SGR is used to estimate the shale content of the fault zone.
In general, fault zones with higher clay content, equivalent to higher SGR values, can support higher capillary threshold pressures. On a broader scale, other factors also exert a control on the threshold pressure, such as depth of the rock sequence at the time of faulting, and the maxim
Document 4:::
Dr. Nahid Khazenie is a mechanical engineer who served as president of the IEEE Geoscience and Remote Sensing Society from 1998 to 1999.
Khazenie completed her undergraduate education at the Michigan Technological University before going on to receive several graduate degrees from the University of Texas at Austin, including a Ph.D. in 1987 for Mechanical Engineering and Operations Research. She joined the faculty and was a research scientist, specializing in remote sensing applications in agriculture and ocean studies. Because of her work there, she became a Senior Scientist appointment to the Naval Research Laboratory and then NASA as Earth Science Enterprise Education Programs Manager.
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What do you call the folded layers of rock where geologists look for oil?
A. crests
B. synclines
C. anticlines
D. oscillations
Answer:
|
|
sciq-1726
|
multiple_choice
|
Who first proposed that earth is a magnet?
|
[
"Marconi",
"william gilbert",
"Albert Einstein",
"Isaac Newton"
] |
B
|
Relavent Documents:
Document 0:::
The following is a chronology of discoveries concerning the magnetosphere.
1600 - William Gilbert in London suggests the Earth is a giant magnet.
1741 - Hiorter and Anders Celsius note that the polar aurora is accompanied by a disturbance of the magnetic needle.
1820 - Hans Christian Ørsted discovers electric currents create magnetic effects. André-Marie Ampère deduces that magnetism is basically the force between electric currents.
1833 - Carl Friedrich Gauss and Wilhelm Weber worked out the mathematical theory for separating the inner and outer Magnetosphere sources of Earth's magnetic field.
1843 - Samuel Schwabe, a German amateur astronomer, shows the existence of an 11-year sunspot cycle.
1859 - Richard Carrington in England observes a solar flare; 17 hours later a large magnetic storm begins.
1892 - George Ellery Hale introduces the spectroheliograph, observing the Sun in hydrogen light from the chromosphere, a sensitive way of detecting flares. He confirms the connection between flares and magnetic storms.
1900-3 - Kristian Birkeland experiments with beams of electrons aimed at a magnetized sphere ("terrella") in a vacuum chamber. The electrons hit near the magnetic poles, leading him to propose that the polar aurora is created by electron beams from the Sun. Birkeland also observes magnetic disturbances associated with the aurora, suggesting to him that localized "polar magnetic storms" exist in the auroral zone.
1902 - Marconi successfully sends radio signals across the Atlantic Ocean. Oliver Heaviside suggests that the radio waves found their way around the curving Earth because they were reflected from electrically conducting layer at the top of the atmosphere.
1926 - Gregory Breit and Merle Tuve measure the distance to the conducting layer—which R. Watson-Watt proposes naming "ionosphere"—by measuring the time needed for a radio signal to bounce back.
1930-1 - After Birkeland's "electron beam" theory is disproved, Sydney Chapman and Vincent Ferrar
Document 1:::
In physics, the dynamo theory proposes a mechanism by which a celestial body such as Earth or a star generates a magnetic field. The dynamo theory describes the process through which a rotating, convecting, and electrically conducting fluid can maintain a magnetic field over astronomical time scales. A dynamo is thought to be the source of the Earth's magnetic field and the magnetic fields of Mercury and the Jovian planets.
History of theory
When William Gilbert published de Magnete in 1600, he concluded that the Earth is magnetic and proposed the first hypothesis for the origin of this magnetism: permanent magnetism such as that found in lodestone. In 1822, André-Marie Ampère proposed that internal currents are responsible of Earth Magnetism
. In 1919, Joseph Larmor proposed that a dynamo might be generating the field. However, even after he advanced his hypothesis, some prominent scientists advanced alternative explanations. The Nobel Prize winner Patrick Blackett did a series of experiments looking for a fundamental relation between angular momentum and magnetic moment, but found none.
Walter M. Elsasser, considered a "father" of the presently accepted dynamo theory as an explanation of the Earth's magnetism, proposed that this magnetic field resulted from electric currents induced in the fluid outer core of the Earth. He revealed the history of the Earth's magnetic field through pioneering the study of the magnetic orientation of minerals in rocks.
In order to maintain the magnetic field against ohmic decay (which would occur for the dipole field in 20,000 years), the outer core must be convecting. The convection is likely some combination of thermal and compositional convection. The mantle controls the rate at which heat is extracted from the core. Heat sources include gravitational energy released by the compression of the core, gravitational energy released by the rejection of light elements (probably sulfur, oxygen, or silicon) at the inner core bo
Document 2:::
George Edward Backus (born May 24, 1930) is an American geophysicist, best known for his work with J. Freeman Gilbert on inverse methods for geophysical data. He is also notably credited with advancing the dynamo theory on the origin of the Earth's magnetic field.
Born in Chicago, Backus did his undergraduate and graduate training at the University of Chicago, receiving a M.S. in mathematics in 1950 and a Ph.D. in physics in 1956. During 1957–1958 he worked as a physicist in Project Matterhorn in Princeton, then moving to MIT for two years as assistant professor of mathematics. In 1960 he moved to Scripps Institution of Oceanography and the associated University of California, San Diego where he has spent the rest of his career. Backus was professor of geophysics until 1994, and subsequently a research professor of geophysics.
Backus is the winner of many awards, including the Gold Medal of the Royal Astronomical Society (1986) and the John Adam Fleming Medal of the American Geophysical Union (1986), and he was elected a fellow of the American Geophysical Union (1967) and of the John Simon Guggenheim Memorial Foundation (1963 and 1970). He is an elected member of the National Academy of Sciences, the American Academy of Arts and Sciences, and of the Académie des Sciences of France.
See also
Backus–Gilbert method
List of geophysicists
Document 3:::
De Magnete, Magneticisque Corporibus, et de Magno Magnete Tellure (On the Magnet and Magnetic Bodies, and on That Great Magnet the Earth) is a scientific work published in 1600 by the English physician and scientist William Gilbert. A highly influential and successful book, it exerted an immediate influence on many contemporary writers, including Francis Godwin and Mark Ridley.
Contents
In De Magnete, Gilbert described many of his experiments with his model Earth called the terrella. Gilbert made the claim that gravity was due to the same force and he believed that this held the Moon in orbit around the Earth.
The work then considered static electricity produced by amber. Amber is called elektron in Greek, and electrum in Latin, so Gilbert decided to refer to the phenomenon by the adjective electricus.
Summary
De Magnete consists of six books.
Book 1
Historical survey of magnetism and theory of Earth's magnetism. The lodestone in antiquity from Plato onwards and the gradual identification of iron ores. The south pole of a lodestone points to the north pole of the Earth and vice versa as the terrestrial globe is magnetic.
Book 2
Distinction between electricity and magnetism. An amber stick when rubbed affects a rotating needle made of any type of metal (a versorium) and attracts paper, leaves and even water. But electricity is different from heat and to magnetism which only attracts iron-bearing materials (he calls it coition). He shows the effects of cutting a spherical lodestone (which he calls a terrella) through the poles and equator and the direction of attraction at different points. Magnets act at a distance but the force has no permanent presence and is not hindered like light. Materials including gold, silver and diamonds are not affected by magnets, nor can one produce perpetual motion.
Book 3
The Earth's normal magnetism. He proposes (incorrectly) that the angle of the ecliptic and precession of the equinoxes are caused by magnetism. A lo
Document 4:::
A terrella (Latin for "little earth") is a small magnetised model ball representing the Earth, that is thought to have been invented by the English physician William Gilbert while investigating magnetism, and further developed 300 years later by the Norwegian scientist and explorer Kristian Birkeland, while investigating the aurora.
Terrellas have been used until the late 20th century to attempt to simulate the Earth's magnetosphere, but have now been replaced by computer simulations.
William Gilbert's terrella
William Gilbert, the royal physician to Queen Elizabeth I, devoted much of his time, energy and resources to the study of the Earth's magnetism. It had been known for centuries that a freely suspended compass needle pointed north. Earlier investigators (including Christopher Columbus) found that direction deviated somewhat from true north, and Robert Norman showed the force on the needle was not horizontal but slanted into the Earth.
William Gilbert's explanation was that the Earth itself was a giant magnet, and he demonstrated this by creating a scale model of the magnetic Earth, a "terrella", a sphere formed out of a lodestone. Passing a small compass over the terrella, Gilbert demonstrated that a horizontal compass would point towards the magnetic pole, while a dip needle, balanced on a horizontal axis perpendicular to the magnetic one, indicated the proper "magnetic inclination" between the magnetic force and the horizontal direction. Gilbert later reported his findings in De Magnete, Magneticisque Corporibus, et de Magno Magnete Tellure, published in 1600.
Kristian Birkeland's terrella
Kristian Birkeland was a Norwegian physicist who, around 1895, tried to explain why the lights of the polar aurora appeared only in regions centered at the magnetic poles.
He simulated the effect by directing cathode rays (later identified as electrons) at a terrella in a vacuum tank, and found they indeed produced a glow in regions around the poles of the terrella
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Who first proposed that earth is a magnet?
A. Marconi
B. william gilbert
C. Albert Einstein
D. Isaac Newton
Answer:
|
|
sciq-4956
|
multiple_choice
|
What do you call it when adult fish of the same species come together in a group and release gametes into the water at the same time?
|
[
"migration",
"spawning",
"schoaling",
"poaching"
] |
B
|
Relavent Documents:
Document 0:::
Fish migration is mass relocation by fish from one area or body of water to another. Many types of fish migrate on a regular basis, on time scales ranging from daily to annually or longer, and over distances ranging from a few metres to thousands of kilometres. Such migrations are usually done for better feeding or to reproduce, but in other cases the reasons are unclear.
Fish migrations involve movements of schools of fish on a scale and duration larger than those arising during normal daily activities. Some particular types of migration are anadromous, in which adult fish live in the sea and migrate into fresh water to spawn; and catadromous, in which adult fish live in fresh water and migrate into salt water to spawn.
Marine forage fish often make large migrations between their spawning, feeding and nursery grounds. Movements are associated with ocean currents and with the availability of food in different areas at different times of year. The migratory movements may partly be linked to the fact that the fish cannot identify their own offspring and moving in this way prevents cannibalism. Some species have been described by the United Nations Convention on the Law of the Sea as highly migratory species. These are large pelagic fish that move in and out of the exclusive economic zones of different nations, and these are covered differently in the treaty from other fish.
Salmon and striped bass are well-known anadromous fish, and freshwater eels are catadromous fish that make large migrations. The bull shark is a euryhaline species that moves at will from fresh to salt water, and many marine fish make a diel vertical migration, rising to the surface to feed at night and sinking to lower layers of the ocean by day. Some fish such as tuna move to the north and south at different times of year following temperature gradients. The patterns of migration are of great interest to the fishing industry. Movements of fish in fresh water also occur; often the fish swim upr
Document 1:::
In biology, any group of fish that stay together for social reasons are shoaling, and if the group is swimming in the same direction in a coordinated manner, they are schooling. In common usage, the terms are sometimes used rather loosely. About one quarter of fish species shoal all their lives, and about one half shoal for part of their lives.
Fish derive many benefits from shoaling behaviour including defence against predators (through better predator detection and by diluting the chance of individual capture), enhanced foraging success, and higher success in finding a mate. It is also likely that fish benefit from shoal membership through increased hydrodynamic efficiency.
Fish use many traits to choose shoalmates. Generally they prefer larger shoals, shoalmates of their own species, shoalmates similar in size and appearance to themselves, healthy fish, and kin (when recognized).
The oddity effect posits that any shoal member that stands out in appearance will be preferentially targeted by predators. This may explain why fish prefer to shoal with individuals that resemble themselves. The oddity effect thus tends to homogenize shoals.
Overview
An aggregation of fish is the general term for any collection of fish that have gathered together in some locality. Fish aggregations can be structured or unstructured. An unstructured aggregation might be a group of mixed species and sizes that have gathered randomly near some local resource, such as food or nesting sites.
If, in addition, the aggregation comes together in an interactive, social way, they may be said to be shoaling. Although shoaling fish can relate to each other in a loose way, with each fish swimming and foraging somewhat independently, they are nonetheless aware of the other members of the group as shown by the way they adjust behaviour such as swimming, so as to remain close to the other fish in the group. Shoaling groups can include fish of disparate sizes and can include mixed-species subgroups.
Document 2:::
Cleaner fish are fish that show a specialist feeding strategy by providing a service to other species, referred to as clients, by removing dead skin, ectoparasites, and infected tissue from the surface or gill chambers. This example of cleaning symbiosis represents mutualism and cooperation behaviour, an ecological interaction that benefits both parties involved. However, the cleaner fish may consume mucus or tissue, thus creating a form of parasitism called cheating. The client animals are typically fish of a different species, but can also be aquatic reptiles (sea turtles and marine iguana), mammals (manatees and whales), or octopuses. A wide variety of fish including wrasse, cichlids, catfish, pipefish, lumpsuckers, and gobies display cleaning behaviors across the globe in fresh, brackish, and marine waters but specifically concentrated in the tropics due to high parasite density. Similar behaviour is found in other groups of animals, such as cleaner shrimps.
There are two types of cleaner fish, obligate full time cleaners and facultative part time cleaners where different strategies occur based on resources and local abundance of fish. Cleaning behaviour takes place in pelagic waters as well as designated locations called cleaner stations. Cleaner fish interaction durations and memories of reoccurring clients are influenced by the neuroendocrine system of the fish, involving hormones arginine vasotocin, Isotocin and serotonin.
Conspicuous coloration is a method used by some cleaner fish, where they often display a brilliant blue stripe that spans the length of the body. Other species of fish, called mimics, imitate the behavior and phenotype of cleaner fish to gain access to client fish tissue.
The specialized feeding behaviour of cleaner fish has become a valuable resource in salmon aquaculture in Atlantic Canada, Scotland, Iceland and Norway for prevention of sea lice outbreaks which benefits the economy and environment by minimizing the use of chemical del
Document 3:::
Invertebrate drift is the downstream transport of invertebrate organisms in lotic freshwater systems such as rivers and streams. The term lotic comes from the Latin word lotus, meaning "washing", and is used to describe moving freshwater systems. This is in contrast with lentic coming from the Latin word lentus, meaning slow or motionless that typically describe still or standing waters such as lakes, ponds, and swamps.
Drift can service freshwater invertebrates by giving them an escape route from predation, or the use of a current to disperse progeny downstream. On occasion, however, invertebrates will inadvertently lose their footing, and drift downstream. For that, invertebrates counter a stream's flow through physical and behavioral adaptations. And just as invertebrates adapted to stabilize themselves in the water column, or use the stream's energy to their advantage, so too have predators adapted to catch invertebrates as they drift. Species of fish, commonly salmonids, catch drifting insects during the peak times after dusk, and before dawn. Fishermen can exploit this relationship using fly fishing techniques and lures that mimic drifting insects to catch these fishes.
Researchers have developed sampling techniques in lotic systems. From it, research as far back as 1928 has collected data on the phenomenon of drift. The study of invertebrate drift has progressed the field of stream ecology. Drift has been documented to impact community structure, benthic production, and the energy flow through trophic levels.
Mechanisms of drift
Types of drift
Invertebrate drift can be categorized by the conditions that caused the drift to occur.
Catastrophic drift: Disturbances such as floods physically dislodge animals.
Behavioral drift: Behavior such as escaping, and inadvertently losing foothold in the water column, cause animals to drift downstream. Active drift describes animals choosing to enter drift.
Distributional drift: Used by animals to disperse progeny
Document 4:::
Spawning triggers are environmental cues that cause marine animals to breed. Most commonly they involve sudden changes in the environment, such as changes in temperature, salinity, and/or the abundance of food. Catfish of the genus Corydoras, for example, spawn immediately after heavy rain, the specific cues being an increase in water level and a decrease in temperature. When water levels rise, it allows many fish access to areas further upstream, that are better suited for reproduction, that was not previously accessible. This is a dangerous game that fish play, though. If they wait too long, they may get trapped in small pockets of water, and eventually, die when the levels recede.Discus will breed when the temperature goes up and there is an overabundance of much needed such as mosquito larvae. Many fish stock up on food to ensure they make it through this exhausting period of time that is very hard on their bodies, while others go without eating during the spawning process because they are so focused on their offspring.
Spawning triggers is what allows many fish to synchronize their breeding, making it more probable that individual fish will find a mate. In most cases, if these triggers were not present, male and female fish would not be on the same page and the offspring would show for it. However, many fish do not respond to specific spawning triggers and will breed either constantly (e.g., guppies); at specific times of the year (e.g., grunion); or only at a certain point in their life cycle (e.g., eels). Some fish, like salmon, spend their whole life getting mature in the ocean, just to swim up many miles into various rivers to lay their eggs. This is such a grueling thing for their bodies, that some species die after laying their eggs. Although most commonly associated with fish, spawning triggers are also present in bivalves and corals.
In certain cases, aquarists can trigger spawning by duplicating the natural conditions where fish would breed. This c
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What do you call it when adult fish of the same species come together in a group and release gametes into the water at the same time?
A. migration
B. spawning
C. schoaling
D. poaching
Answer:
|
|
sciq-4664
|
multiple_choice
|
The cytoplasm divides during what process?
|
[
"genesys",
"cytokinesis",
"metamorphosis",
"osmosis"
] |
B
|
Relavent Documents:
Document 0:::
In cell biology, the cleavage furrow is the indentation of the cell's surface that begins the progression of cleavage, by which animal and some algal cells undergo cytokinesis, the final splitting of the membrane, in the process of cell division. The same proteins responsible for muscle contraction, actin and myosin, begin the process of forming the cleavage furrow, creating an actomyosin ring. Other cytoskeletal proteins and actin binding proteins are involved in the procedure.
Mechanism
Plant cells do not perform cytokinesis through this exact method but the two procedures are not totally different. Animal cells form an actin-myosin contractile ring within the equatorial region of the cell membrane that constricts to form the cleavage furrow. In plant cells, Golgi vesicle secretions form a cell plate or septum on the equatorial plane of the cell wall by the action of microtubules of the phragmoplast. The cleavage furrow in animal cells and the phragmoplast in plant cells are complex structures made up of microtubules and microfilaments that aide in the final separation of the cells into two identical daughter cells.
Cell cycle
The cell cycle begins with interphase when the DNA replicates, the cell grows and prepares to enter mitosis. Mitosis includes four phases: prophase, metaphase, anaphase, and telophase. Prophase is the initial phase when spindle fibers appear that function to move the chromosomes toward opposite poles. This spindle apparatus consists of microtubules, microfilaments and a complex network of various proteins. During metaphase, the chromosomes line up using the spindle apparatus in the middle of the cell along the equatorial plate. The chromosomes move to opposite poles during anaphase and remain attached to the spindle fibers by their centromeres. Animal cell cleavage furrow formation is caused by a ring of actin microfilaments called the contractile ring, which forms during early anaphase. Myosin is present in the region of the contracti
Document 1:::
Organelle biogenesis is the biogenesis, or creation, of cellular organelles in cells. Organelle biogenesis includes the process by which cellular organelles are split between daughter cells during mitosis; this process is called organelle inheritance.
Discovery
Following the discovery of cellular organelles in the nineteenth century, little was known about their function and synthesis until the development of electron microscopy and subcellular fractionation in the twentieth century. This allowed experiments on the function, structure, and biogenesis of these organelles to commence.
Mechanisms of protein sorting and retrieval have been found to give organelles their characteristic composition. It is known that cellular organelles can come from preexisting organelles; however, it is a subject of controversy whether organelles can be created without a preexisting one.
Process
Several processes are known to have developed for organelle biogenesis. These can range from de novo synthesis to the copying of a template organelle; the formation of an organelle 'from scratch' and using a preexisting organelle as a template to manufacture an organelle, respectively. The distinct structures of each organelle are thought to be caused by the different mechanisms of the processes which create them and the proteins that they are made up of. Organelles may also be 'split' between two cells during the process of cellular division (known as organelle inheritance), where the organelle of the parent cell doubles in size and then splits with each half being delivered to their respective daughter cells.
The process of organelle biogenesis is known to be regulated by specialized transcription networks that modulate the expression of the genes that code for specific organellar proteins. In order for organelle biogenesis to be carried out properly, the specific genes coding for the organellar proteins must be transcribed properly and the translation of the resulting mRNA must be succes
Document 2:::
Cytochemistry is the branch of cell biology dealing with the detection of cell constituents by means of biochemical analysis and visualization techniques. This is the study of the localization of cellular components through the use of staining methods. The term is also used to describe a process of identification of the biochemical content of cells. Cytochemistry is a science of localizing chemical components of cells and cell organelles on thin histological sections by using several techniques like enzyme localization, micro-incineration, micro-spectrophotometry, radioautography, cryo-electron microscopy, X-ray microanalysis by energy-dispersive X-ray spectroscopy, immunohistochemistry and cytochemistry, etc.
Freeze Fracture Enzyme Cytochemistry
Freeze fracture enzyme cytochemistry was initially mentioned in the study of Pinto de silva in 1987. It is a technique that allows the introduction of cytochemistry into a freeze fracture cell membrane. immunocytochemistry is used in this technique to label and visualize the cell membrane's molecules. This technique could be useful in analyzing the ultrastructure of cell membranes. The combination of immunocytochemistry and freeze fracture enzyme technique, research can identify and have a better understanding of the structure and distribution of a cell membrane.
Origin
Jean Brachet's research in Brussel demonstrated the localization and relative abundance between RNA and DNA in the cells of both animals and plants opened up the door into the research of cytochemistry. The work by Moller and Holter in 1976 about endocytosis which discussed the relationship between a cell's structure and function had established the needs of cytochemical research.
Aims
Cytochemical research aims to study individual cells that may contain several cell types within a tissue. It takes a nondestructive approach to study the localization of the cell. By remaining the cell components intact, researcher are able to study the intact cell activ
Document 3:::
The cell is the basic structural and functional unit of all forms of life. Every cell consists of cytoplasm enclosed within a membrane, and contains many macromolecules such as proteins, DNA and RNA, as well as many small molecules of nutrients and metabolites. The term comes from the Latin word meaning 'small room'.
Cells can acquire specified function and carry out various tasks within the cell such as replication, DNA repair, protein synthesis, and motility. Cells are capable of specialization and mobility within the cell.
Most plant and animal cells are only visible under a light microscope, with dimensions between 1 and 100 micrometres. Electron microscopy gives a much higher resolution showing greatly detailed cell structure. Organisms can be classified as unicellular (consisting of a single cell such as bacteria) or multicellular (including plants and animals). Most unicellular organisms are classed as microorganisms.
The study of cells and how they work has led to many other studies in related areas of biology, including: discovery of DNA, cancer systems biology, aging and developmental biology.
Cell biology is the study of cells, which were discovered by Robert Hooke in 1665, who named them for their resemblance to cells inhabited by Christian monks in a monastery. Cell theory, first developed in 1839 by Matthias Jakob Schleiden and Theodor Schwann, states that all organisms are composed of one or more cells, that cells are the fundamental unit of structure and function in all living organisms, and that all cells come from pre-existing cells. Cells emerged on Earth about 4 billion years ago.
Discovery
With continual improvements made to microscopes over time, magnification technology became advanced enough to discover cells. This discovery is largely attributed to Robert Hooke, and began the scientific study of cells, known as cell biology. When observing a piece of cork under the scope, he was able to see pores. This was shocking at the time as i
Document 4:::
Cell proliferation is the process by which a cell grows and divides to produce two daughter cells. Cell proliferation leads to an exponential increase in cell number and is therefore a rapid mechanism of tissue growth. Cell proliferation requires both cell growth and cell division to occur at the same time, such that the average size of cells remains constant in the population. Cell division can occur without cell growth, producing many progressively smaller cells (as in cleavage of the zygote), while cell growth can occur without cell division to produce a single larger cell (as in growth of neurons). Thus, cell proliferation is not synonymous with either cell growth or cell division, despite these terms sometimes being used interchangeably.
Stem cells undergo cell proliferation to produce proliferating "transit amplifying" daughter cells that later differentiate to construct tissues during normal development and tissue growth, during tissue regeneration after damage, or in cancer.
The total number of cells in a population is determined by the rate of cell proliferation minus the rate of cell death.
Cell size depends on both cell growth and cell division, with a disproportionate increase in the rate of cell growth leading to production of larger cells and a disproportionate increase in the rate of cell division leading to production of many smaller cells. Cell proliferation typically involves balanced cell growth and cell division rates that maintain a roughly constant cell size in the exponentially proliferating population of cells. Cell proliferation occurs by combining cell growth with regular "G1-S-M-G2" cell cycles to produce many diploid cell progeny.
In single-celled organisms, cell proliferation is largely responsive to the availability of nutrients in the environment (or laboratory growth medium).
In multicellular organisms, the process of cell proliferation is tightly controlled by gene regulatory networks encoded in the genome and executed mainly
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
The cytoplasm divides during what process?
A. genesys
B. cytokinesis
C. metamorphosis
D. osmosis
Answer:
|
|
sciq-2073
|
multiple_choice
|
What does the coefficient in front of an element in a formula indicate?
|
[
"place on the Periodic Table",
"number of atoms",
"time of the reaction",
"how much is involved in the reaction"
] |
D
|
Relavent Documents:
Document 0:::
In physical chemistry, there are numerous quantities associated with chemical compounds and reactions; notably in terms of amounts of substance, activity or concentration of a substance, and the rate of reaction. This article uses SI units.
Introduction
Theoretical chemistry requires quantities from core physics, such as time, volume, temperature, and pressure. But the highly quantitative nature of physical chemistry, in a more specialized way than core physics, uses molar amounts of substance rather than simply counting numbers; this leads to the specialized definitions in this article. Core physics itself rarely uses the mole, except in areas overlapping thermodynamics and chemistry.
Notes on nomenclature
Entity refers to the type of particle/s in question, such as atoms, molecules, complexes, radicals, ions, electrons etc.
Conventionally for concentrations and activities, square brackets [ ] are used around the chemical molecular formula. For an arbitrary atom, generic letters in upright non-bold typeface such as A, B, R, X or Y etc. are often used.
No standard symbols are used for the following quantities, as specifically applied to a substance:
the mass of a substance m,
the number of moles of the substance n,
partial pressure of a gas in a gaseous mixture p (or P),
some form of energy of a substance (for chemistry enthalpy H is common),
entropy of a substance S
the electronegativity of an atom or chemical bond χ.
Usually the symbol for the quantity with a subscript of some reference to the quantity is used, or the quantity is written with the reference to the chemical in round brackets. For example, the mass of water might be written in subscripts as mH2O, mwater, maq, mw (if clear from context) etc., or simply as m(H2O). Another example could be the electronegativity of the fluorine-fluorine covalent bond, which might be written with subscripts χF-F, χFF or χF-F etc., or brackets χ(F-F), χ(FF) etc.
Neither is standard. For the purpose of this a
Document 1:::
A physical coefficient is an important number that characterizes some physical property of a technical or scientific object. A coefficient also has a scientific reference which is the reliance on force.
Stoichiometric coefficient of a chemical compound
To find the coefficient of a chemical compound, you must balance the elements involved in it. For example, water:
H2O.
It just so happens that hydrogen (H) and oxygen (O) are both diatomic molecules, thus we have H2 and O2. To form water, one of the O atoms breaks off from the O2 molecule and react with the H2 compound to form H2O. But, there is one oxygen atom left. It reacts with another H2 molecule. Since it took two of each atom to balance the compound, we put the coefficient 2 in front of H2O:
2 H2O.
The total reaction is thus 2 H2 + O2 → 2 H2O.
Examples of physical coefficients
Coefficient of thermal expansion (thermodynamics) (dimensionless) - Relates the change in temperature to the change in a material's dimensions.
Partition coefficient (KD) (chemistry) - The ratio of concentrations of a compound in two phases of a mixture of two immiscible solvents at equilibrium.
Hall coefficient (electrical physics) - Relates a magnetic field applied to an element to the voltage created, the amount of current and the element thickness. It is a characteristic of the material from which the conductor is made.
Lift coefficient (CL or CZ) (aerodynamics) (dimensionless) - Relates the lift generated by an airfoil with the dynamic pressure of the fluid flow around the airfoil, and the planform area of the airfoil.
Ballistic coefficient (BC) (aerodynamics) (units of kg/m2) - A measure of a body's ability to overcome air resistance in flight. BC is a function of mass, diameter, and drag coefficient.
Transmission coefficient (quantum mechanics) (dimensionless) - Represents the probability flux of a transmitted wave relative to that of an incident wave. It is often used to describe the probability of a particle
Document 2:::
A chemical equation is the symbolic representation of a chemical reaction in the form of symbols and chemical formulas. The reactant entities are given on the left-hand side and the product entities are on the right-hand side with a plus sign between the entities in both the reactants and the products, and an arrow that points towards the products to show the direction of the reaction. The chemical formulas may be symbolic, structural (pictorial diagrams), or intermixed. The coefficients next to the symbols and formulas of entities are the absolute values of the stoichiometric numbers. The first chemical equation was diagrammed by Jean Beguin in 1615.
Structure
A chemical equation (see an example below) consists of a list of reactants (the starting substances) on the left-hand side, an arrow symbol, and a list of products (substances formed in the chemical reaction) on the right-hand side. Each substance is specified by its chemical formula, optionally preceded by a number called stoichiometric coefficient. The coefficient specifies how many entities (e.g. molecules) of that substance are involved in the reaction on a molecular basis. If not written explicitly, the coefficient is equal to 1. Multiple substances on any side of the equation are separated from each other by a plus sign.
As an example, the equation for the reaction of hydrochloric acid with sodium can be denoted:
Given the formulas are fairly simple, this equation could be read as "two H-C-L plus two N-A yields two N-A-C-L and H two." Alternately, and in general for equations involving complex chemicals, the chemical formulas are read using IUPAC nomenclature, which could verbalise this equation as "two hydrochloric acid molecules and two sodium atoms react to form two formula units of sodium chloride and a hydrogen gas molecule."
Reaction types
Different variants of the arrow symbol are used to denote the type of a reaction:
{|
| style="text-align: center; padding-right: 0.5em;" | -> || net forwa
Document 3:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 4:::
In mathematical psychology and education theory, a knowledge space is a combinatorial structure used to formulate mathematical models describing the progression of a human learner. Knowledge spaces were introduced in 1985 by Jean-Paul Doignon and Jean-Claude Falmagne, and remain in extensive use in the education theory. Modern applications include two computerized tutoring systems, ALEKS and the defunct RATH.
Formally, a knowledge space assumes that a domain of knowledge is a collection of concepts or skills, each of which must be eventually mastered. Not all concepts are interchangeable; some require other concepts as prerequisites. Conversely, competency at one skill may ease the acquisition of another through similarity. A knowledge space marks out which collections of skills are feasible: they can be learned without mastering any other skills. Under reasonable assumptions, the collection of feasible competencies forms the mathematical structure known as an antimatroid.
Researchers and educators usually explore the structure of a discipline's knowledge space as a latent class model.
Motivation
Knowledge Space Theory attempts to address shortcomings of standardized testing when used in educational psychometry. Common tests, such as the SAT and ACT, compress a student's knowledge into a very small range of ordinal ranks, in the process effacing the conceptual dependencies between questions. Consequently, the tests cannot distinguish between true understanding and guesses, nor can they identify a student's particular weaknesses, only the general proportion of skills mastered. The goal of knowledge space theory is to provide a language by which exams can communicate
What the student can do and
What the student is ready to learn.
Model structure
Knowledge Space Theory-based models presume that an educational subject can be modeled as a finite set of concepts, skills, or topics. Each feasible state of knowledge about is then a subset of ; the set of
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What does the coefficient in front of an element in a formula indicate?
A. place on the Periodic Table
B. number of atoms
C. time of the reaction
D. how much is involved in the reaction
Answer:
|
|
sciq-5525
|
multiple_choice
|
In a wheel and axle, when force is applied either to the wheel or axle, what aspect of the force does not change?
|
[
"pressure",
"direction",
"speed",
"shift"
] |
B
|
Relavent Documents:
Document 0:::
The circle of forces, traction circle, friction circle, or friction ellipse is a useful way to think about the dynamic interaction between a vehicle's tire and the road surface. The diagram below shows the tire from above, so that the road surface lies in the xy-plane. The vehicle to which the tire is attached is moving in the positive y direction.
In this example, the vehicle would be cornering to the right (i.e. the positive x direction points to the center of the corner). Note that the plane of rotation of the tire is at an angle to the actual direction that the tire is moving (the positive y direction). Put differently, rather than being allowed to simply "roll" in the direction that it is "pointing" (in this case, rightwards from the positive y direction), the tire instead must "slip" in a different direction from that which it is pointing in order to maintain its "forward" motion in the positive y direction. This difference between the direction the tire "points" (its plane of rotation) and the tire's actual direction of travel is the slip angle.
A tire can generate horizontal force where it meets the road surface by the mechanism of slip. That force is represented in the diagram by the vector F. Note that in this example, F is perpendicular to the plane of the tire. That is because the tire is rolling freely, with no torque applied to it by the vehicle's brakes or drive train. However, that is not always the case.
The magnitude of F is limited by the dashed circle, but it can be any combination of the components Fx and Fy that does not extend beyond the dashed circle. (For a real-world tire, the circle is likely to be closer to an ellipse, with the y axis slightly longer than the x axis.)
In the example, the tire is generating a component of force in the x direction (Fx) which, when transferred to the vehicle's chassis via the suspension system in combination with similar forces from the other tires, will cause the vehicle to turn to the right. Note that
Document 1:::
As described by the third of Newton's laws of motion of classical mechanics, all forces occur in pairs such that if one object exerts a force on another object, then the second object exerts an equal and opposite reaction force on the first. The third law is also more generally stated as: "To every action there is always opposed an equal reaction: or the mutual actions of two bodies upon each other are always equal, and directed to contrary parts." The attribution of which of the two forces is the action and which is the reaction is arbitrary. Either of the two can be considered the action, while the other is its associated reaction.
Examples
Interaction with ground
When something is exerting force on the ground, the ground will push back with equal force in the opposite direction. In certain fields of applied physics, such as biomechanics, this force by the ground is called 'ground reaction force'; the force by the object on the ground is viewed as the 'action'.
When someone wants to jump, he or she exerts additional downward force on the ground ('action'). Simultaneously, the ground exerts upward force on the person ('reaction'). If this upward force is greater than the person's weight, this will result in upward acceleration. When these forces are perpendicular to the ground, they are also called a normal force.
Likewise, the spinning wheels of a vehicle attempt to slide backward across the ground. If the ground is not too slippery, this results in a pair of friction forces: the 'action' by the wheel on the ground in backward direction, and the 'reaction' by the ground on the wheel in forward direction. This forward force propels the vehicle.
Gravitational forces
The Earth, among other planets, orbits the Sun because the Sun exerts a gravitational pull that acts as a centripetal force, holding the Earth to it, which would otherwise go shooting off into space. If the Sun's pull is considered an action, then Earth simultaneously exerts a reaction as a gravi
Document 2:::
Mechanical load is the physical stress on a mechanical system or component. Loads can be static or dynamic. Some loads are specified as part of the design criteria of a mechanical system. Depending on the usage, some mechanical loads can be measured by an appropriate test method in a laboratory or in the field.
Vehicle
It can be the external mechanical resistance against which a machine (such as a motor or engine), acts. The load can often be expressed as a curve of force versus speed.
For instance, a given car traveling on a road of a given slope presents a load which the engine must act against. Because air resistance increases with speed, the motor must put out more torque at a higher speed in order to maintain the speed. By shifting to a higher gear, one may be able to meet the requirement with a higher torque and a lower engine speed, whereas shifting to a lower gear has the opposite effect. Accelerating increases the load, whereas decelerating decreases the load.
Pump
Similarly, the load on a pump depends on the head against which the pump is pumping, and on the size of the pump.
Fan
Similar considerations apply to a fan. See Affinity laws.
See also
Structural load
Physical test
Document 3:::
Vehicle dynamics is the study of vehicle motion, e.g., how a vehicle's forward movement changes in response to driver inputs, propulsion system outputs, ambient conditions, air/surface/water conditions, etc.
Vehicle dynamics is a part of engineering primarily based on classical mechanics.
It may be applied for motorized vehicles (such as automobiles), bicycles and motorcycles, aircraft, and watercraft.
Factors affecting vehicle dynamics
The aspects of a vehicle's design which affect the dynamics can be grouped into drivetrain and braking, suspension and steering, distribution of mass, aerodynamics and tires.
Drivetrain and braking
Automobile layout (i.e. location of engine and driven wheels)
Powertrain
Braking system
Suspension and steering
Some attributes relate to the geometry of the suspension, steering and chassis. These include:
Ackermann steering geometry
Axle track
Camber angle
Caster angle
Ride height
Roll center
Scrub radius
Steering ratio
Toe
Wheel alignment
Wheelbase
Distribution of mass
Some attributes or aspects of vehicle dynamics are purely due to mass and its distribution. These include:
Center of mass
Moment of inertia
Roll moment
Sprung mass
Unsprung mass
Weight distribution
Aerodynamics
Some attributes or aspects of vehicle dynamics are purely aerodynamic. These include:
Automobile drag coefficient
Automotive aerodynamics
Center of pressure
Downforce
Ground effect in cars
Tires
Some attributes or aspects of vehicle dynamics can be attributed directly to the tires. These include:
Camber thrust
Circle of forces
Contact patch
Cornering force
Ground pressure
Pacejka's Magic Formula
Pneumatic trail
Radial Force Variation
Relaxation length
Rolling resistance
Self aligning torque
Skid
Slip angle
Slip (vehicle dynamics)
Spinout
Steering ratio
Tire load sensitivity
Vehicle behaviours
Some attributes or aspects of vehicle dynamics are purely dynamic. These include:
Body flex
Body roll
Bump Steer
Bu
Document 4:::
Cornering force or side force is the lateral (i.e., parallel to wheel axis) force produced by a vehicle tire during cornering.
Cornering force is generated by tire slip and is proportional to slip angle at low slip angles. The rate at which cornering force builds up is described by relaxation length. Slip angle describes the deformation of the tire contact patch, and this deflection of the contact patch deforms the tire in a fashion akin to a spring.
As with deformation of a spring, deformation of the tire contact patch generates a reaction force in the tire; the cornering force. Integrating the force generated by every tread element along the contact patch length gives the total cornering force. Although the term, "tread element" is used, the compliance in the tire that leads to this effect is actually a combination of sidewall deflection and deflection of the rubber within the contact patch. The exact ratio of sidewall compliance to tread compliance is a factor in tire construction and inflation pressure.
Because the tire deformation tends to reach a maximum behind the center of the contact patch, by a distance known as pneumatic trail, it tends to generate a torque about a vertical axis known as self aligning torque.
The diagram is misleading because the reaction force would appear to be acting in the wrong direction. It is simply a matter of convention to quote positive cornering force as acting in the opposite direction to positive tire slip so that calculations are simplified, since a vehicle cornering under the influence of a cornering force to the left will generate a tire slip to the right.
The same principles can be applied to a tire being deformed longitudinally, or in a combination of both longitudinal and lateral directions. The behaviour of a tire under combined longitudinal and lateral deformation can be described by a traction circle.
See also
Camber thrust
Lateral force variation
Circle of forces
Skidpad
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
In a wheel and axle, when force is applied either to the wheel or axle, what aspect of the force does not change?
A. pressure
B. direction
C. speed
D. shift
Answer:
|
|
sciq-2053
|
multiple_choice
|
What does the human heart do?
|
[
"destroy blood",
"make blood",
"use blood",
"pump blood"
] |
D
|
Relavent Documents:
Document 0:::
Cardiophysics is an interdisciplinary science that stands at the junction of cardiology and medical physics, with researchers using the methods of, and theories from, physics to study cardiovascular system at different levels of its organisation, from the molecular scale to whole organisms. Being formed historically as part of systems biology, cardiophysics designed to reveal connections between the physical mechanisms, underlying the organization of the cardiovascular system, and biological features of its functioning.
Zbigniew R. Struzik seems to be a first author who used the term in a scientific publication in 2004.
One can use interchangeably also the terms cardiovascular physics.
See also
Medical physics
Important publications in medical physics
Biomedicine
Biomedical engineering
Physiome
Nanomedicine
Document 1:::
Heart nanotechnology is the "Engineering of functional systems at the molecular scale" ("Nanotechnology Research").
Nanotechnology
Nanotechnology deals with structures and materials that are approximately one to one-hundred nanometers in length. At this microscopic level, quantum mechanics take place and are in effect, resulting in behaviors that would seem quite strange compared to what humans see with the naked eye (regular matter). Nanotechnology is used for a wide variety of fields of technology, ranging from energy to electronics to medicine. In the category of medicine, nanotechnology is still relatively new and has not yet been widely adopted by the field. It is possible that nanotechnology could be the new breakthrough of medicine and may eventually be the solution and cure for many of the health problems that humans encounter. Nanotechnology may lead to the cure for illnesses such as the common cold, diseases, and cancer. It is already starting to be used as a treatment for some serious health issues; more specifically it is being used to treat the heart and cancer.
Nanomedicine
Nanotechnology in the field of medicine is more commonly referred to as nanomedicine. Nanomedicine that deals with helping the heart is really starting to take off and gain in popularity compared to most of the other fields that nanomedicine currently has to offer. There are several heart problems that nanotechnology has promising evidence of being effective in the treatment of heart disease in the near future.
Examples
It should hopefully be able to treat heart valves that are defective; and detect and treat arterial plaque in the heart ("Nanotechnology Made Clear"). Nanomedicine should be able to help heal the hearts of people that have already been victims of heart disease and heart attacks. On the other hand, it will also play a key role in finding people with a high risk of having heart disease, and will be able to help prevent heart attacks from happening in the first p
Document 2:::
The Cardiac Electrophysiology Society (CES) is an international society of basic and clinical scientists and physicians interested in cardiac electrophysiology and arrhythmias. The Cardiac Electrophysiology Society's founder was George Burch in 1949 and its current president is Jonathan C. Makielski, M.D.
Document 3:::
Lucien Campeau (June 20, 1927March 15, 2010) was a Canadian cardiologist. He was a full professor at the Université de Montréal. He is best known for performing the world's first transradial coronary angiogram. Campeau was one of the founding staff of the Montreal Heart Institute, joining in 1957. He is also well known for developing the Canadian Cardiovascular Society grading of angina pectoris.
Education
Campeau received his M.D. degree from the University of Laval in 1953 and completed a fellowship in Cardiology at Johns Hopkins Hospital from 1956 to 1957. He later became a professor at University of Montreal in 1961 and was one of the co-founders of the Montreal Heart Institute.
In his lifetime, Campeau was awarded the Research Achievement Award of the Canadian Cardiovascular Society. In 2004, he was named “Cardiologue émérite 2004” by the Association des cardiologues du Québec.
Document 4:::
Merry L. Lindsey is an American cardiac physiologist. She is the Stokes-Shackleford Professor and Chair of the University of Nebraska Medical Center Department of Cellular and Integrative Physiology and the director of the Center for Heart and Vascular Research. In 2021, Lindsey was appointed editor-in-chief of the American Journal of Physiology. Heart and Circulatory Physiology.
Early life and education
Lindsey was born Stuart, Florida in 1970 and raised in South Florida, where she attended South Fork High School. Following high school, Lindsey earned her undergraduate degree in biology from Boston University and her PhD in cardiovascular sciences from Baylor College of Medicine.
Career
Upon completing her PhD, Lindsey worked at the Medical University of South Carolina as an assistant professor before joining the faculty at the University of Texas Health Science Center. In 2019, she left the Mississippi Center for Heart Research to accept an appointment as the Stokes-Shackleford Professor and Chair of the Department of Cellular and Integrative Physiology at the University of Nebraska Medical Center. Upon joining the department, Lindsey also became the founding director of the Center for Heart and Vascular Research. She joined Meharry Medical College as the dean of the School of Graduate Studies and Research.
In 2021, Lindsey was appointed editor-in-chief of the American Journal of Physiology. Heart and Circulatory Physiology, a journal published by the American Physiological Society. She received the Vincenzo Panagia Distinguished Lecture Award from the Institute of Cardiovascular Sciences at St-Boniface Hospital Research in 2021, and the Distinguished Investigator Award from the British Society for Cardiovascular Research in 2022.
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What does the human heart do?
A. destroy blood
B. make blood
C. use blood
D. pump blood
Answer:
|
|
sciq-9032
|
multiple_choice
|
When performing anaerobic exercise, your muscles work against what?
|
[
"resistance",
"inertia",
"momentum",
"gravity"
] |
A
|
Relavent Documents:
Document 0:::
In an isotonic contraction, tension remains the same, whilst the muscle's length changes. Isotonic contractions differ from isokinetic contractions in that in isokinetic contractions the muscle speed remains constant. While superficially identical, as the muscle's force changes via the length-tension relationship during a contraction, an isotonic contraction will keep force constant while velocity changes, but an isokinetic contraction will keep velocity constant while force changes. A near isotonic contraction is known as Auxotonic contraction.
There are two types of isotonic contractions: (1) concentric and (2) eccentric. In a concentric contraction, the muscle tension rises to meet the resistance, then remains the same as the muscle shortens. In eccentric, the muscle lengthens due to the resistance being greater than the force the muscle is producing.
Concentric
This type is typical of most exercise. The external force on the muscle is less than the force the muscle is generating - a shortening contraction. The effect is not visible during the classic biceps curl, which is in fact auxotonic because the resistance (torque due to the weight being lifted) does not remain the same through the exercise. Tension is highest at a parallel to the floor level, and eases off above and below this point. Therefore, tension changes as well as muscle length.
Eccentric
There are two main features to note regarding eccentric contractions. First, the absolute tensions achieved can be very high relative to the muscle's maximum tetanic tension generating capacity (you can set down a much heavier object than you can lift). Second, the absolute tension is relatively independent of lengthening velocity.
Muscle injury and soreness are selectively associated with eccentric contraction. Muscle strengthening using exercises that involve eccentric contractions is lower than using concentric exercises. However because higher levels of tension are easier to attain during exercises th
Document 1:::
Exertion is the physical or perceived use of energy. Exertion traditionally connotes a strenuous or costly effort, resulting in generation of force, initiation of motion, or in the performance of work. It often relates to muscular activity and can be quantified, empirically and by measurable metabolic response.
Physical
In physics, exertion is the expenditure of energy against, or inductive of, inertia as described by Isaac Newton's third law of motion. In physics, force exerted equivocates work done. The ability to do work can be either positive or negative depending on the direction of exertion relative to gravity. For example, a force exerted upwards, like lifting an object, creates positive work done on that object.
Exertion often results in force generated, a contributing dynamic of general motion. In mechanics it describes the use of force against a body in the direction of its motion (see vector).
Physiological
Exertion, physiologically, can be described by the initiation of exercise, or, intensive and exhaustive physical activity that causes cardiovascular stress or a sympathetic nervous response. This can be continuous or intermittent exertion.
Exertion requires, of the body, modified oxygen uptake, increased heart rate, and autonomic monitoring of blood lactate concentrations. Mediators of physical exertion include cardio-respiratory and musculoskeletal strength, as well as metabolic capability. This often correlates to an output of force followed by a refractory period of recovery. Exertion is limited by cumulative load and repetitive motions.
Muscular energy reserves, or stores for biomechanical exertion, stem from metabolic, immediate production of ATP and increased oxygen consumption. Muscular exertion generated depends on the muscle length and the velocity at which it is able to shorten, or contract.
Perceived exertion can be explained as subjective, perceived experience that mediates response to somatic sensations and mechanisms. A rating of pe
Document 2:::
Myology is the study of the muscular system, including the study of the structure, function and diseases of muscle. The muscular system consists of skeletal muscle, which contracts to move or position parts of the body (e.g., the bones that articulate at joints), smooth and cardiac muscle that propels, expels or controls the flow of fluids and contained substance.
See also
Myotomy
Oral myology
Document 3:::
Muscle hypertrophy or muscle building involves a hypertrophy or increase in size of skeletal muscle through a growth in size of its component cells. Two factors contribute to hypertrophy: sarcoplasmic hypertrophy, which focuses more on increased muscle glycogen storage; and myofibrillar hypertrophy, which focuses more on increased myofibril size. It is the primary focus of bodybuilding-related activities.
Hypertrophy stimulation
A range of stimuli can increase the volume of muscle cells. These changes occur as an adaptive response that serves to increase the ability to generate force or resist fatigue in anaerobic conditions.
Strength training
Strength training (resistance training) causes neural and muscular adaptations which increase the capacity of an athlete to exert force through voluntary muscular contraction: After an initial period of neuro-muscular adaptation, the muscle tissue expands by creating sarcomeres (contractile elements) and increasing non-contractile elements like sarcoplasmic fluid.
Muscular hypertrophy can be induced by progressive overload (a strategy of progressively increasing resistance or repetitions over successive bouts of exercise to maintain a high level of effort). However, the precise mechanisms are not clearly understood; currently accepted hypotheses involve some combination of mechanical tension, metabolic fatigue, and muscular damage.
Muscular hypertrophy plays an important role in competitive bodybuilding and strength sports like powerlifting, American football, and Olympic weightlifting.
Anaerobic training
The best approach to specifically achieve muscle growth remains controversial (as opposed to focusing on gaining strength, power, or endurance); it was generally considered that consistent anaerobic strength training will produce hypertrophy over the long term, in addition to its effects on muscular strength and endurance. Muscular hypertrophy can be increased through strength training and other short-duration, high-i
Document 4:::
Normal aging movement control in humans is about the changes in the muscles, motor neurons, nerves, sensory functions, gait, fatigue, visual and manual responses, in men and women as they get older but who do not have neurological, muscular (atrophy, dystrophy...) or neuromuscular disorder. With aging, neuromuscular movements are impaired, though with training or practice, some aspects may be prevented.
Force production
For voluntary force production, action potentials occur in the cortex. They propagate in the spinal cord, the motor neurons and the set of muscle fibers they innervate. This results in a twitch which properties are driven by two mechanisms: motor unit recruitment and rate coding. Both mechanisms are affected with aging. For instance, the number of motor units may decrease, the size of the motor units, i.e. the number of muscle fibers they innervate may increase, the frequency at which the action potentials are triggered may be reduced. Consequently, force production is generally impaired in old adults.
Aging is associated with decreases in muscle mass and strength. These decreases may be partially due to losses of alpha motor neurons. By the age of 70, these losses occur in both proximal and distal muscles. In biceps brachii and brachialis, old adults show decreased strength (by 1/3) correlated with a reduction in the number of motor units (by 1/2). Old adults show evidence that remaining motor units may become larger as motor units innervate collateral muscle fibers.
In first dorsal interosseus, almost all motor units are recruited at moderate rate coding, leading to 30-40% of maximal voluntary contraction (MVC). Motor unit discharge rates measured at 50% MVC are not significantly different in the young subjects from those observed in the old adults. However, for the maximal effort contractions, there is an appreciable difference in discharge rates between the two age groups. Discharge rates obtained at 100% of MVC are 64% smaller in the old adul
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
When performing anaerobic exercise, your muscles work against what?
A. resistance
B. inertia
C. momentum
D. gravity
Answer:
|
|
sciq-2153
|
multiple_choice
|
What is the predominant stage in the life cycle of most scyphozoans?
|
[
"medusa",
"gamete",
"spore",
"hydra"
] |
A
|
Relavent Documents:
Document 0:::
A trichocyst is an organelle found in certain ciliates and dinoflagellates.
A trichocyst can be found in tetrahymena and along cila pathways of several metabolic systems.
It is also a structure in the cortex of certain ciliate and flagellate protozoans consisting of a cavity and long, thin threads that can be ejected in response to certain stimuli. Trichocysts may be widely distributed over an organism or restricted to certain areas (e.g., tentacles, papillae, around the mouth). There are several types. Mucoid trichocysts are elongated inclusions that may be ejected as visible bodies after artificial stimulation. Filamentous trichocysts in Paramecium and other ciliates are discharged as filaments composed of a cross-striated shaft and a tip. Toxicysts (in Dileptus and certain other carnivorous protozoans) tend to be localized around the mouth. When discharged, a toxicyst expels a long, nonstriated filament with a rodlike tip, which paralyzes or kills other microorganisms; this filament is used to capture food and, presumably, in defense.
The functional significance of other trichocysts is uncertain, although those of Paramecium apparently can be extruded for anchorage during feeding.
Document 1:::
A polyp in zoology is one of two forms found in the phylum Cnidaria, the other being the medusa. Polyps are roughly cylindrical in shape and elongated at the axis of the vase-shaped body. In solitary polyps, the aboral (opposite to oral) end is attached to the substrate by means of a disc-like holdfast called a pedal disc, while in colonies of polyps it is connected to other polyps, either directly or indirectly. The oral end contains the mouth, and is surrounded by a circlet of tentacles.
Classes
In the class Anthozoa, comprising the sea anemones and corals, the individual is always a polyp; in the class Hydrozoa, however, the individual may be either a polyp or a medusa, with most species undergoing a life cycle with both a polyp stage and a medusa stage. In class Scyphozoa, the medusa stage is dominant, and the polyp stage may or may not be present, depending on the family. In those scyphozoans that have the larval planula metamorphose into a polyp, the polyp, also called a "scyphistoma," grows until it develops a stack of plate-like medusae that pinch off and swim away in a process known as strobilation. Once strobilation is complete, the polyp may die, or regenerate itself to repeat the process again later. With Cubozoans, the planula settles onto a suitable surface, and develops into a polyp. The cubozoan polyp then eventually metamorphoses directly into a Medusa.
Anatomy
The body of the polyp may be roughly compared in a structure to a sac, the wall of which is composed of two layers of cells. The outer layer is known technically as the ectoderm, the inner layer as the endoderm (or gastroderm). Between ectoderm and endoderm is a supporting layer of structureless gelatinous substance termed mesoglea, secreted by the cell layers of the body wall. The mesoglea can be thinner than the endoderm or ectoderm or comprise the bulk of the body as in larger jellyfish. The mesoglea can contain skeletal elements derived from cells migrated from ectoderm.
Th
Document 2:::
Schizophyte was a botanical classification proposed by Ferdinand Cohn to describe the class of primitive "plants" that reproduce solely by fission. It has been considered synonymous with the Protophyta of Sachs and the Monera of Haeckel. In modern taxonomy, it is equivalent with the concept of prokaryotes, single-celled microorganisms with no nucleus or other membrane-bound organelles, now divided into the domains Bacteria and Archaea.
Document 3:::
In biology, cell theory is a scientific theory first formulated in the mid-nineteenth century, that organisms are made up of cells, that they are the basic structural/organizational unit of all organisms, and that all cells come from pre-existing cells. Cells are the basic unit of structure in all organisms and also the basic unit of reproduction.
The theory was once universally accepted, but now some biologists consider non-cellular entities such as viruses living organisms, and thus disagree with the first tenet. As of 2021: "expert opinion remains divided roughly a third each between yes, no and don’t know". As there is no universally accepted definition of life, discussion still continues.
History
With continual improvements made to microscopes over time, magnification technology became advanced enough to discover cells. This discovery is largely attributed to Robert Hooke, and began the scientific study of cells, known as cell biology. When observing a piece of cork under the scope, he was able to see pores. This was shocking at the time as it was believed no one else had seen these. To further support his theory, Matthias Schleiden and Theodor Schwann both also studied cells of both animal and plants. What they discovered were significant differences between the two types of cells. This put forth the idea that cells were not only fundamental to plants, but animals as well.
Microscopes
The discovery of the cell was made possible through the invention of the microscope. In the first century BC, Romans were able to make glass. They discovered that objects appeared to be larger under the glass. The expanded use of lenses in eyeglasses in the 13th century probably led to wider spread use of simple microscopes (magnifying glasses) with limited magnification. Compound microscopes, which combine an objective lens with an eyepiece to view a real image achieving much higher magnification, first appeared in Europe around 1620. In 1665, Robert Hooke used a microscope
Document 4:::
Actinosphaerium is a genus of heliozoa, amoeboid unicellular organisms with many axopodial filaments that radiate out of their cell. It is classified within the monotypic family Actinosphaeriidae and suborder Actinosphaerina. Species of Actinophrys are distinguished by their large number of nuclei in each cell. Their axopodia sometimes terminate on the surface of nuclei. Vacuoles are abundant in the periphery of the cytoplasm.
Morphology
Actinosphaerium species belong to an informal group known as heliozoa, which are unicellular eukaryotes (or protists) that are heterotrophic (also known as protozoa) and present slender, radiating, specialized pseudopodia known as axopodia. Its cell structure has been studied profusely through electron microscopy during the 20th century. Actinosphaerium cells are spherical and multinucleate (i.e. have more than one cell nucleus), as opposed to Actinophrys species which are uninucleate. The axonemes of their axopodia may or may not end on the surface of their nuclei. Their cells range from 200 to 400 μm in diameter.
The cytoplasm of Actinosphaerium species is divided into a highly vacuolated ectoplasm (i.e. with numerous non-contractile vacuoles) and a less vacuolated endoplasm. Multiple long, slender axopodia radiate out of the cell body. Each axopodium is composed of a relatively stiff axial rod, surrounded by a thin layer of ectoplasm. The rods penetrate deep into the endoplasm and can terminate freely or close to the cell nuclei.
Ecology
Actinosphaerium is a freshwater genus of protists. It has been observed consuming a diverse range of prey such as midge larvae, sessile colonial ciliates and several rotifer species.
Systematics
Taxonomy
Actinosphaerium was created by German zoologist Ritter von Stein in 1857 to accommodate the species Actinophrys eichhornii (now Actinosphaerium eichhornii), distinguished from current Actinophrys species by a large number of nuclei.
In 1965, Hovasse divided Actinosphaerium to create t
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is the predominant stage in the life cycle of most scyphozoans?
A. medusa
B. gamete
C. spore
D. hydra
Answer:
|
|
sciq-4711
|
multiple_choice
|
What is the main type of cell in the nervous system called?
|
[
"neuron",
"axon",
"sensory",
"conduction"
] |
A
|
Relavent Documents:
Document 0:::
Nervous tissue, also called neural tissue, is the main tissue component of the nervous system. The nervous system regulates and controls body functions and activity. It consists of two parts: the central nervous system (CNS) comprising the brain and spinal cord, and the peripheral nervous system (PNS) comprising the branching peripheral nerves. It is composed of neurons, also known as nerve cells, which receive and transmit impulses, and neuroglia, also known as glial cells or glia, which assist the propagation of the nerve impulse as well as provide nutrients to the neurons.
Nervous tissue is made up of different types of neurons, all of which have an axon. An axon is the long stem-like part of the cell that sends action potentials to the next cell. Bundles of axons make up the nerves in the PNS and tracts in the CNS.
Functions of the nervous system are sensory input, integration, control of muscles and glands, homeostasis, and mental activity.
Structure
Nervous tissue is composed of neurons, also called nerve cells, and neuroglial cells. Four types of neuroglia found in the CNS are astrocytes, microglial cells, ependymal cells, and oligodendrocytes. Two types of neuroglia found in the PNS are satellite glial cells and Schwann cells. In the central nervous system (CNS), the tissue types found are grey matter and white matter. The tissue is categorized by its neuronal and neuroglial components.
Components
Neurons are cells with specialized features that allow them to receive and facilitate nerve impulses, or action potentials, across their membrane to the next neuron. They possess a large cell body (soma), with cell projections called dendrites and an axon. Dendrites are thin, branching projections that receive electrochemical signaling (neurotransmitters) to create a change in voltage in the cell. Axons are long projections that carry the action potential away from the cell body toward the next neuron. The bulb-like end of the axon, called the axon terminal, i
Document 1:::
The development of the nervous system in humans, or neural development or neurodevelopment involves the studies of embryology, developmental biology, and neuroscience to describe the cellular and molecular mechanisms by which the complex nervous system forms in humans, develops during prenatal development, and continues to develop postnatally.
Some landmarks of neural development in the embryo include the formation and differentiation of neurons from stem cell precursors (neurogenesis); the migration of immature neurons from their birthplaces in the embryo to their final positions; the outgrowth of axons from neurons and guidance of the motile growth cone through the embryo towards postsynaptic partners, the generation of synapses between these axons and their postsynaptic partners, the synaptic pruning that occurs in adolescence, and finally the lifelong changes in synapses which are thought to underlie learning and memory.
Typically, these neurodevelopmental processes can be broadly divided into two classes: activity-independent mechanisms and activity-dependent mechanisms. Activity-independent mechanisms are generally believed to occur as hardwired processes determined by genetic programs played out within individual neurons. These include differentiation, migration and axon guidance to their initial target areas. These processes are thought of as being independent of neural activity and sensory experience. Once axons reach their target areas, activity-dependent mechanisms come into play. Neural activity and sensory experience will mediate formation of new synapses, as well as synaptic plasticity, which will be responsible for refinement of the nascent neural circuits.
Development of the human brain
Overview
The central nervous system (CNS) is derived from the ectoderm—the outermost tissue layer of the embryo. In the third week of human embryonic development the neuroectoderm appears and forms the neural plate along the dorsal side of the embryo. The neural
Document 2:::
The neuron doctrine is the concept that the nervous system is made up of discrete individual cells, a discovery due to decisive neuro-anatomical work of Santiago Ramón y Cajal and later presented by, among others, H. Waldeyer-Hartz. The term neuron (spelled neurone in British English) was itself coined by Waldeyer as a way of identifying the cells in question. The neuron doctrine, as it became known, served to position neurons as special cases under the broader cell theory evolved some decades earlier. He appropriated the concept not from his own research but from the disparate observation of the histological work of Albert von Kölliker, Camillo Golgi, Franz Nissl, Santiago Ramón y Cajal, Auguste Forel and others.
Historical context
Theodor Schwann proposed in 1839 that the tissues of all organisms are composed of cells. Schwann was expanding on the proposal of his good friend Matthias Jakob Schleiden the previous year that all plant tissues were composed of cells. The nervous system stood as an exception. Although nerve cells had been described in tissue by numerous investigators including Jan Purkinje, Gabriel Valentin, and Robert Remak, the relationship between the nerve cells and other features such as dendrites and axons was not clear. The connections between the large cell bodies and smaller features could not be observed, and it was possible that neurofibrils would stand as an exception to cell theory as non-cellular components of living tissue. Technical limitations of microscopy and tissue preparation were largely responsible. Chromatic aberration, spherical aberration and the dependence on natural light all played a role in limiting microscope performance in the early 19th century. Tissue was typically lightly mashed in water and pressed between a glass slide and cover slip. There was also a limited number of dyes and fixatives available prior to the middle of the 19th century.
A landmark development came from Camillo Golgi who invented a silver
Document 3:::
The following are two lists of animals ordered by the size of their nervous system. The first list shows number of neurons in their entire nervous system, indicating their overall neural complexity. The second list shows the number of neurons in the structure that has been found to be representative of animal intelligence. The human brain contains 86 billion neurons, with 16 billion neurons in the cerebral cortex.
Scientists are engaged in counting, quantification, in order to find answers to the question in the strategy of neuroscience and intelligence of "self-knowledge": how the evolution of a set of components and parameters (~1011 neurons, ~1014 synapses) of a complex system could lead to the phenomenon of the appearance of intelligence in the biological species "sapiens".
Overview
Neurons are the cells that transmit information in an animal's nervous system so that it can sense stimuli from its environment and behave accordingly. Not all animals have neurons; Trichoplax and sponges lack nerve cells altogether.
Neurons may be packed to form structures such as the brain of vertebrates or the neural ganglions of insects.
The number of neurons and their relative abundance in different parts of the brain is a determinant of neural function and, consequently, of behavior.
Whole nervous system
All numbers for neurons (except Caenorhabditis and Ciona), and all numbers for synapses (except Ciona) are estimations.
List of animal species by forebrain (cerebrum or pallium) neuron number
The question of what physical characteristic of an animal makes an animal intelligent has varied over the centuries. One early speculation was brain size (or weight, which provides the same ordering.) A second proposal was brain-to-body-mass ratio, and a third was encephalization quotient, sometimes referred to as EQ. The current best predictor is number of neurons in the forebrain, based on Herculano-Houzel's improved neuron counts. It accounts most accurately for variations
Document 4:::
These are timelines of brain development events in different animal species.
Mouse brain development timeline
Macaque brain development timeline
Human brain development timeline
See also
Encephalization quotient
Evolution of the brain
Neural development
External links
Translating Neurodevelopmental Time Across Mammalian Species
Vertebrate developmental biology
Embryology of nervous system
Developmental neuroscience
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is the main type of cell in the nervous system called?
A. neuron
B. axon
C. sensory
D. conduction
Answer:
|
|
sciq-9138
|
multiple_choice
|
What are the largest known proteins called?
|
[
"obscures",
"sporozoans",
"aminos",
"titins"
] |
D
|
Relavent Documents:
Document 0:::
This is a list of topics in molecular biology. See also index of biochemistry articles.
Document 1:::
The Institute of Biophysics, Chinese Academy of Sciences, based in Beijing, China, focuses on biophysically oriented basic research in the life sciences. It was established by Bei Shizhang in 1958, from the former Beijing Experimental Biology Institute founded in 1957. Xu Tao is the current Director.
The main research focus of the Institute is on the fields of protein science and brain & cognitive sciences. The Institute has two National Key Laboratories—"The National Laboratory of Biomacromolecules" and "The State Laboratory of Brain and Cognitive Sciences". The establishment of the National Laboratory of Protein Science was given approval by China's Ministry of Science and Technology (MOST) in December 2006. Research in the field of protein science emphasizes the following areas: 3D-structure and function of proteins, bio-membranes and membrane proteins, protein translation and folding, protein interaction networks, the molecular basis of infection and immunity, the molecular basis of sensation and cognition, protein and peptide drugs, and new techniques and methods in protein science research. Research areas in the brain and cognitive sciences include neural processes and mechanisms in complex cognition, expression of visual perception and attention, neural mechanisms of perceptional information processing, and dysfunction in brain cognition.
The Institute has received National Natural Science Foundation, '973', '863', 'Knowledge Innovation Program', and other major research grants, supporting outstanding research in a range of areas. The achievements of the Institute in terms of awards, publications, patents, and applied research maintain the Institute at the highest level nationally, and it has worldwide recognition for research in the life sciences. Among other connections, since 2008 it has hosted an intensive course in macromolecular crystallography as a resource closely modeled on the course at Cold Spring Harbor Laboratory on Long Island, USA, and invol
Document 2:::
MicrobeLibrary is a permanent collection of over 1400 original peer-reviewed resources for teaching undergraduate microbiology. It is provided by the American Society for Microbiology, Washington DC, United States.
Contents include curriculum activities; images and animations; reviews of books, websites and other resources; and articles from Focus on Microbiology Education, Microbiology Education and Microbe. Around 40% of the materials are free to educators and students, the remainder require a subscription. the service is suspended with the message to:
"Please check back with us in 2017".
External links
MicrobeLibrary
Microbiology
Document 3:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 4:::
Proteins are a class of biomolecules composed of amino acid chains.
Biochemistry
Antifreeze protein, class of polypeptides produced by certain fish, vertebrates, plants, fungi and bacteria
Conjugated protein, protein that functions in interaction with other chemical groups attached by covalent bonds
Denatured protein, protein which has lost its functional conformation
Matrix protein, structural protein linking the viral envelope with the virus core
Protein A, bacterial surface protein that binds antibodies
Protein A/G, recombinant protein that binds antibodies
Protein C, anticoagulant
Protein G, bacterial surface protein that binds antibodies
Protein L, bacterial surface protein that binds antibodies
Protein S, plasma glycoprotein
Protein Z, glycoprotein
Protein catabolism, the breakdown of proteins into amino acids and simple derivative compounds
Protein complex, group of two or more associated proteins
Protein electrophoresis, method of analysing a mixture of proteins by means of gel electrophoresis
Protein folding, process by which a protein assumes its characteristic functional shape or tertiary structure
Protein isoform, version of a protein with some small differences
Protein kinase, enzyme that modifies other proteins by chemically adding phosphate groups to them
Protein ligands, atoms, molecules, and ions which can bind to specific sites on proteins
Protein microarray, piece of glass on which different molecules of protein have been affixed at separate locations in an ordered manner
Protein phosphatase, enzyme that removes phosphate groups that have been attached to amino acid residues of proteins
Protein purification, series of processes intended to isolate a single type of protein from a complex mixture
Protein sequencing, protein method
Protein splicing, intramolecular reaction of a particular protein in which an internal protein segment is removed from a precursor protein
Protein structure, unique three-dimensional shape of amino
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What are the largest known proteins called?
A. obscures
B. sporozoans
C. aminos
D. titins
Answer:
|
|
sciq-10111
|
multiple_choice
|
What are composed of alternating positive and negative ions?
|
[
"magnetic crystals",
"radioactive substances",
"ionic crystals",
"polar bonds"
] |
C
|
Relavent Documents:
Document 0:::
An ion () is an atom or molecule with a net electrical charge. The charge of an electron is considered to be negative by convention and this charge is equal and opposite to the charge of a proton, which is considered to be positive by convention. The net charge of an ion is not zero because its total number of electrons is unequal to its total number of protons.
A cation is a positively charged ion with fewer electrons than protons while an anion is a negatively charged ion with more electrons than protons. Opposite electric charges are pulled towards one another by electrostatic force, so cations and anions attract each other and readily form ionic compounds.
Ions consisting of only a single atom are termed atomic or monatomic ions, while two or more atoms form molecular ions or polyatomic ions. In the case of physical ionization in a fluid (gas or liquid), "ion pairs" are created by spontaneous molecule collisions, where each generated pair consists of a free electron and a positive ion. Ions are also created by chemical interactions, such as the dissolution of a salt in liquids, or by other means, such as passing a direct current through a conducting solution, dissolving an anode via ionization.
History of discovery
The word ion was coined from Greek neuter present participle of ienai (), meaning "to go". A cation is something that moves down ( pronounced kato, meaning "down") and an anion is something that moves up (, meaning "up"). They are so called because ions move toward the electrode of opposite charge. This term was introduced (after a suggestion by the English polymath William Whewell) by English physicist and chemist Michael Faraday in 1834 for the then-unknown species that goes from one electrode to the other through an aqueous medium. Faraday did not know the nature of these species, but he knew that since metals dissolved into and entered a solution at one electrode and new metal came forth from a solution at the other electrode; that some kind of
Document 1:::
The ionic strength of a solution is a measure of the concentration of ions in that solution. Ionic compounds, when dissolved in water, dissociate into ions. The total electrolyte concentration in solution will affect important properties such as the dissociation constant or the solubility of different salts. One of the main characteristics of a solution with dissolved ions is the ionic strength. Ionic strength can be molar (mol/L solution) or molal (mol/kg solvent) and to avoid confusion the units should be stated explicitly. The concept of ionic strength was first introduced by Lewis and Randall in 1921 while describing the activity coefficients of strong electrolytes.
Quantifying ionic strength
The molar ionic strength, I, of a solution is a function of the concentration of all ions present in that solution.
where one half is because we are including both cations and anions, ci is the molar concentration of ion i (M, mol/L), zi is the charge number of that ion, and the sum is taken over all ions in the solution. For a 1:1 electrolyte such as sodium chloride, where each ion is singly-charged, the ionic strength is equal to the concentration. For the electrolyte MgSO4, however, each ion is doubly-charged, leading to an ionic strength that is four times higher than an equivalent concentration of sodium chloride:
Generally multivalent ions contribute strongly to the ionic strength.
Calculation example
As a more complex example, the ionic strength of a mixed solution 0.050 M in Na2SO4 and 0.020 M in KCl is:
Non-ideal solutions
Because in non-ideal solutions volumes are no longer strictly additive it is often preferable to work with molality b (mol/kg of H2O) rather than molarity c (mol/L). In that case, molal ionic strength is defined as:
in which
i = ion identification number
z = charge of ion
b = molality (mol solute per Kg solvent)
Importance
The ionic strength plays a central role in the Debye–Hückel theory that describes the strong deviations from id
Document 2:::
A solvated electron is a free electron in (solvated in) a solution, and is the smallest possible anion. Solvated electrons occur widely. Often, discussions of solvated electrons focus on their solutions in ammonia, which are stable for days, but solvated electrons also occur in water and other solvents in fact, in any solvent that mediates outer-sphere electron transfer. The solvated electron is responsible for a great deal of radiation chemistry.
Ammonia solutions
Liquid ammonia will dissolve all of the alkali metals and other electropositive metals such as Ca, Sr, Ba, Eu, and Yb (also Mg using an electrolytic process), giving characteristic blue solutions. For alkali metals in liquid ammonia, the solution is blue when dilute and copper-colored when more concentrated (> 3 molar). These solutions conduct electricity. The blue colour of the solution is due to ammoniated electrons, which absorb energy in the visible region of light. The diffusivity of the solvated electron in liquid ammonia can be determined using potential-step chronoamperometry.
Solvated electrons in ammonia are the anions of salts called electrides.
Na + 6 NH3 → [Na(NH3)6]+e−
The reaction is reversible: evaporation of the ammonia solution produces a film of metallic sodium.
Case study: Li in NH3
A lithium–ammonia solution at −60 °C is saturated at about 15 mol% metal (MPM). When the concentration is increased in this range electrical conductivity increases from 10−2 to 104 Ω−1cm−1 (larger than liquid mercury). At around 8 MPM, a "transition to the metallic state" (TMS) takes place (also called a "metal-to-nonmetal transition" (MNMT)). At 4 MPM a liquid-liquid phase separation takes place: the less dense gold-colored phase becomes immiscible from a denser blue phase. Above 8 MPM the solution is bronze/gold-colored. In the same concentration range the overall density decreases by 30%.
Other solvents
Alkali metals also dissolve in some small primary amines, such as methylamine and ethylami
Document 3:::
An ionic liquid (IL) is a salt in the liquid state. In some contexts, the term has been restricted to salts whose melting point is below a specific temperature, such as . While ordinary liquids such as water and gasoline are predominantly made of electrically neutral molecules, ionic liquids are largely made of ions. These substances are variously called liquid electrolytes, ionic melts, ionic fluids, fused salts, liquid salts, or ionic glasses.
Ionic liquids have many potential applications. They are powerful solvents and can be used as electrolytes. Salts that are liquid at near-ambient temperature are important for electric battery applications, and have been considered as sealants due to their very low vapor pressure.
Any salt that melts without decomposing or vaporizing usually yields an ionic liquid. Sodium chloride (NaCl), for example, melts at into a liquid that consists largely of sodium cations () and chloride anions (). Conversely, when an ionic liquid is cooled, it often forms an ionic solid—which may be either crystalline or glassy.
The ionic bond is usually stronger than the Van der Waals forces between the molecules of ordinary liquids. Because of these strong interactions, salts tend to have high lattice energies, manifested in high melting points. Some salts, especially those with organic cations, have low lattice energies and thus are liquid at or below room temperature. Examples include compounds based on the 1-ethyl-3-methylimidazolium (EMIM) cation and include: EMIM:Cl, EMIMAc (acetate anion), EMIM dicyanamide, ()()·, that melts at ; and 1-butyl-3,5-dimethylpyridinium bromide which becomes a glass below .
Low-temperature ionic liquids can be compared to ionic solutions, liquids that contain both ions and neutral molecules, and in particular to the so-called deep eutectic solvents, mixtures of ionic and non-ionic solid substances which have much lower melting points than the pure compounds. Certain mixtures of nitrate salts can have melt
Document 4:::
An electrolyte is a medium containing ions that is electrically conducting through the movement of those ions, but not conducting electrons. This includes most soluble salts, acids, and bases dissolved in a polar solvent, such as water. Upon dissolving, the substance separates into cations and anions, which disperse uniformly throughout the solvent. Solid-state electrolytes also exist. In medicine and sometimes in chemistry, the term electrolyte refers to the substance that is dissolved.
Electrically, such a solution is neutral. If an electric potential is applied to such a solution, the cations of the solution are drawn to the electrode that has an abundance of electrons, while the anions are drawn to the electrode that has a deficit of electrons. The movement of anions and cations in opposite directions within the solution amounts to a current. Some gases, such as hydrogen chloride (HCl), under conditions of high temperature or low pressure can also function as electrolytes. Electrolyte solutions can also result from the dissolution of some biological (e.g., DNA, polypeptides) or synthetic polymers (e.g., polystyrene sulfonate), termed "polyelectrolytes", which contain charged functional groups. A substance that dissociates into ions in solution or in the melt acquires the capacity to conduct electricity. Sodium, potassium, chloride, calcium, magnesium, and phosphate in a liquid phase are examples of electrolytes.
In medicine, electrolyte replacement is needed when a person has prolonged vomiting or diarrhea, and as a response to sweating due to strenuous athletic activity. Commercial electrolyte solutions are available, particularly for sick children (such as oral rehydration solution, Suero Oral, or Pedialyte) and athletes (sports drinks). Electrolyte monitoring is important in the treatment of anorexia and bulimia.
In science, electrolytes are one of the main components of electrochemical cells.
In clinical medicine, mentions of electrolytes usually refer m
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What are composed of alternating positive and negative ions?
A. magnetic crystals
B. radioactive substances
C. ionic crystals
D. polar bonds
Answer:
|
|
sciq-3761
|
multiple_choice
|
What do you call a simple machine consisting of a sloping surface that connects lower and higher elevations?
|
[
"jack",
"inclined plane",
"wedge",
"lever"
] |
B
|
Relavent Documents:
Document 0:::
Machine element or hardware refers to an elementary component of a machine. These elements consist of three basic types:
structural components such as frame members, bearings, axles, splines, fasteners, seals, and lubricants,
mechanisms that control movement in various ways such as gear trains, belt or chain drives, linkages, cam and follower systems, including brakes and clutches, and
control components such as buttons, switches, indicators, sensors, actuators and computer controllers.
While generally not considered to be a machine element, the shape, texture and color of covers are an important part of a machine that provide a styling and operational interface between the mechanical components of a machine and its users.
Machine elements are basic mechanical parts and features used as the building blocks of most machines. Most are standardized to common sizes, but customs are also common for specialized applications.
Machine elements may be features of a part (such as screw threads or integral plain bearings) or they may be discrete parts in and of themselves such as wheels, axles, pulleys, rolling-element bearings, or gears. All of the simple machines may be described as machine elements, and many machine elements incorporate concepts of one or more simple machines. For example, a leadscrew incorporates a screw thread, which is an inclined plane wrapped around a cylinder.
Many mechanical design, invention, and engineering tasks involve a knowledge of various machine elements and an intelligent and creative combining of these elements into a component or assembly that fills a need (serves an application).
Structural elements
Beams,
Struts,
Bearings,
Fasteners
Keys,
Splines,
Cotter pin,
Seals
Machine guardings
Mechanical elements
Engine,
Electric motor,
Actuator,
Shafts,
Couplings
Belt,
Chain,
Cable drives,
Gear train,
Clutch,
Brake,
Flywheel,
Cam,
follower systems,
Linkage,
Simple machine
Types
Shafts
Document 1:::
A hoist is a device used for lifting or lowering a load by means of a drum or lift-wheel around which rope or chain wraps. It may be manually operated, electrically or pneumatically driven and may use chain, fiber or wire rope as its lifting medium. The most familiar form is an elevator, the car of which is raised and lowered by a hoist mechanism. Most hoists couple to their loads using a lifting hook. Today, there are a few governing bodies for the North American overhead hoist industry which include the Hoist Manufactures Institute, ASME, and the Occupational Safety and Health Administration. HMI is a product counsel of the Material Handling Industry of America consisting of hoist manufacturers promoting safe use of their products.
Types
The word “hoist” is used to describe many different types of equipment that lift and lower loads. For example, many people use “hoist” to describe an elevator. The information contained here pertains specially to overhead, construction and mine hoist.
Overhead hoist
Overhead hoists are defined in the American Society of Mechanical Engineers (ASME) B30 standards as a machinery unit that is used for lifting or lowering a freely suspended (unguided) load. These units are typically used in an industrial setting and may be part of an overhead crane.
A specific overhead hoist configuration is usually defined by the lifting medium, operation and suspension. The lifting medium is the type of component used to transmit and cause the vertical motion and includes wire rope, chain or synthetic strap, or rope. The operation defines the type of power used to operate the hoisting motion and includes manual power, electric power, hydraulic power or air power. The suspension defines the type of mounting method used to suspend the hoist and includes hook, clevis, lug, trolley, deck, base, wall or ceiling.
The most commonly used overhead hoist is electrical powered with wire rope or chain as the lifting medium. Both wire rope and ch
Document 2:::
Mechanical engineering is a discipline centered around the concept of using force multipliers, moving components, and machines. It utilizes knowledge of mathematics, physics, materials sciences, and engineering technologies. It is one of the oldest and broadest of the engineering disciplines.
Dawn of civilization to early middle ages
Engineering arose in early civilization as a general discipline for the creation of large scale structures such as irrigation, architecture, and military projects. Advances in food production through irrigation allowed a portion of the population to become specialists in Ancient Babylon.
All six of the classic simple machines were known in the ancient Near East. The wedge and the inclined plane (ramp) were known since prehistoric times. The wheel, along with the wheel and axle mechanism, was invented in Mesopotamia (modern Iraq) during the 5th millennium BC. The lever mechanism first appeared around 5,000 years ago in the Near East, where it was used in a simple balance scale, and to move large objects in ancient Egyptian technology. The lever was also used in the shadoof water-lifting device, the first crane machine, which appeared in Mesopotamia circa 3000 BC, and then in ancient Egyptian technology circa 2000 BC. The earliest evidence of pulleys date back to Mesopotamia in the early 2nd millennium BC, and ancient Egypt during the Twelfth Dynasty (1991-1802 BC). The screw, the last of the simple machines to be invented, first appeared in Mesopotamia during the Neo-Assyrian period (911-609) BC. The Egyptian pyramids were built using three of the six simple machines, the inclined plane, the wedge, and the lever, to create structures like the Great Pyramid of Giza.
The Assyrians were notable in their use of metallurgy and incorporation of iron weapons. Many of their advancements were in military equipment. They were not the first to develop them, but did make advancements on the wheel and the chariot. They made use of pivot-able axl
Document 3:::
A gristmill (also: grist mill, corn mill, flour mill, feed mill or feedmill) grinds cereal grain into flour and middlings. The term can refer to either the grinding mechanism or the building that holds it. Grist is grain that has been separated from its chaff in preparation for grinding.
History
Early history
The Greek geographer Strabo reports in his Geography a water-powered grain-mill to have existed near the palace of king Mithradates VI Eupator at Cabira, Asia Minor, before 71 BC.
The early mills had horizontal paddle wheels, an arrangement which later became known as the "Norse wheel", as many were found in Scandinavia. The paddle wheel was attached to a shaft which was, in turn, attached to the centre of the millstone called the "runner stone". The turning force produced by the water on the paddles was transferred directly to the runner stone, causing it to grind against a stationary "bed", a stone of a similar size and shape. This simple arrangement required no gears, but had the disadvantage that the speed of rotation of the stone was dependent on the volume and flow of water available and was, therefore, only suitable for use in mountainous regions with fast-flowing streams. This dependence on the volume and speed of flow of the water also meant that the speed of rotation of the stone was highly variable and the optimum grinding speed could not always be maintained.
Vertical wheels were in use in the Roman Empire by the end of the first century BC, and these were described by Vitruvius. The rotating mill is considered "one of the greatest discoveries of the human race". It was a very physically demanding job for workers, where the slave workers were considered little different from animals, the miseries of which were depicted in iconography and Apuleius' The Golden Ass. The peak of Roman technology is probably the Barbegal aqueduct and mill where water with a 19-metre fall drove sixteen water wheels, giving a grinding capacity estimated at 28 tons per
Document 4:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What do you call a simple machine consisting of a sloping surface that connects lower and higher elevations?
A. jack
B. inclined plane
C. wedge
D. lever
Answer:
|
|
sciq-3662
|
multiple_choice
|
Water can be boiled to kill giardia and most other what?
|
[
"bacteria",
"parasites",
"pathogens",
"germs"
] |
C
|
Relavent Documents:
Document 0:::
Bacteriological water analysis is a method of analysing water to estimate the numbers of bacteria present and, if needed, to find out what sort of bacteria they are. It represents one aspect of water quality. It is a microbiological analytical procedure which uses samples of water and from these samples determines the concentration of bacteria. It is then possible to draw inferences about the suitability of the water for use from these concentrations. This process is used, for example, to routinely confirm that water is safe for human consumption or that bathing and recreational waters are safe to use.
The interpretation and the action trigger levels for different waters vary depending on the use made of the water. Whilst very stringent levels apply to drinking water, more relaxed levels apply to marine bathing waters, where much lower volumes of water are expected to be ingested by users.
Approach
The common feature of all these routine screening procedures is that the primary analysis is for indicator organisms rather than the pathogens that might cause concern. Indicator organisms are bacteria such as non-specific coliforms, Escherichia coli and Pseudomonas aeruginosa that are very commonly found in the human or animal gut and which, if detected, may suggest the presence of sewage. Indicator organisms are used because even when a person is infected with a more pathogenic bacteria, they will still be excreting many millions times more indicator organisms than pathogens. It is therefore reasonable to surmise that if indicator organism levels are low, then pathogen levels will be very much lower or absent. Judgements as to suitability of water for use are based on very extensive precedents and relate to the probability of any sample population of bacteria being able to be infective at a reasonable statistical level of confidence.
Analysis is usually performed using culture, biochemical and sometimes optical methods. When indicator organisms levels exceed pre-set
Document 1:::
A boil-water advisory (BWA), boil-water notice, boil-water warning, boil-water order, or boil order is a public-health advisory or directive issued by governmental or other health authorities to consumers when a community's drinking water is or could be contaminated by pathogens.
Under a BWA, the Centers for Disease Control and Prevention recommends that water be brought to a rolling boil for one minute before it is consumed in order to kill protozoa, bacteria, and viruses. At altitudes above , boiling should be extended to 3 minutes, as the lower boiling point at high altitudes requires more time to kill such organisms. A boil water advisory usually lasts up to 24-48 hours, but sometimes more.
BWA's are typically issued when monitoring of water being served to consumers detects E. coli or other microbiological indicators of sewage contamination. Another reason for a BWA is a failure of distribution system integrity evidenced by a loss of system pressure. While loss of pressure does not necessarily mean the water has been contaminated, it does mean that pathogens may be able to enter the piped-water system and thus be carried to consumers. In the United States, this has been defined as a drop below .
History
John Snow's 1849 recommendation that water be "filtered and boiled before it is used" is one of the first practical applications of the germ theory of disease in the area of public health and is the antecedent to the modern boil water advisory. Snow demonstrated a clear understanding of germ theory in his writings. He first published his theory in an 1849 essay On the Mode of Communication of Cholera, in which he correctly suggested that the fecal-oral route was the mode of communication, and that the disease replicated itself in the lower intestines. Snow later went so far as to accurately propose in his 1855 edition of the work that the structure of cholera was that of a cell. Snow's ideas were not fully accepted until years after his death in 1858.
The
Document 2:::
The branches of microbiology can be classified into pure and applied sciences. Microbiology can be also classified based on taxonomy, in the cases of bacteriology, mycology, protozoology, and phycology. There is considerable overlap between the specific branches of microbiology with each other and with other disciplines, and certain aspects of these branches can extend beyond the traditional scope of microbiology
In general the field of microbiology can be divided in the more fundamental branch (pure microbiology) and the applied microbiology (biotechnology). In the more fundamental field the organisms are studied as the subject itself on a deeper (theoretical) level.
Applied microbiology refers to the fields where the micro-organisms are applied in certain processes such as brewing or fermentation. The organisms itself are often not studied as such, but applied to sustain certain processes.
Pure microbiology
Bacteriology: the study of bacteria
Mycology: the study of fungi
Protozoology: the study of protozoa
Phycology/algology: the study of algae
Parasitology: the study of parasites
Immunology: the study of the immune system
Virology: the study of viruses
Nematology: the study of nematodes
Microbial cytology: the study of microscopic and submicroscopic details of microorganisms
Microbial physiology: the study of how the microbial cell functions biochemically. Includes the study of microbial growth, microbial metabolism and microbial cell structure
Microbial pathogenesis: the study of pathogens which happen to be microbes
Microbial ecology: the relationship between microorganisms and their environment
Microbial genetics: the study of how genes are organized and regulated in microbes in relation to their cellular functions Closely related to the field of molecular biology
Cellular microbiology: a discipline bridging microbiology and cell biology
Evolutionary microbiology: the study of the evolution of microbes. This field can be subdivided into:
Micr
Document 3:::
Microbial cytology is the study of microscopic and submicroscopic details of microorganisms. Origin of "Microbial" 1880-85; < Greek mīkro- micro- small + bíos life). "Cytology" 1857; < Cyto-is derived from the Greek "kytos" meaning "hollow, as a cell or container." + -logy meaning "the study of"). Microbial cytology is analyzed under a microscope for cells which were collected from a part of the body. The main purpose of microbial cytology is to see the structure of the cells, and how they form and operate.
Document 4:::
Industrial microbiology is a branch of biotechnology that applies microbial sciences to create industrial products in mass quantities, often using microbial cell factories. There are multiple ways to manipulate a microorganism in order to increase maximum product yields. Introduction of mutations into an organism may be accomplished by introducing them to mutagens. Another way to increase production is by gene amplification, this is done by the use of plasmids, and vectors. The plasmids and/ or vectors are used to incorporate multiple copies of a specific gene that would allow more enzymes to be produced that eventually cause more product yield. The manipulation of organisms in order to yield a specific product has many applications to the real world like the production of some antibiotics, vitamins, enzymes, amino acids, solvents, alcohol and daily products. Microorganisms play a big role in the industry, with multiple ways to be used. Medicinally, microbes can be used for creating antibiotics in order to treat infection. Microbes can also be used for the food industry as well. Microbes are very useful in creating some of the mass produced products that are consumed by people. The chemical industry also uses microorganisms in order to synthesize amino acids and organic solvents. Microbes can also be used in an agricultural application for use as a biopesticide instead of using dangerous chemicals and or inoculants to help plant proliferation.
Medical application
The medical application to industrial microbiology is the production of new drugs synthesized in a specific organism for medical purposes. Production of antibiotics is necessary for the treatment of many bacterial infections. Some natural occurring antibiotics and precursors, are produced through a process called fermentation. The microorganisms grow in a liquid media where the population size is controlled in order to yield the greatest amount of product. In this environment nutrient, pH, temperature, an
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Water can be boiled to kill giardia and most other what?
A. bacteria
B. parasites
C. pathogens
D. germs
Answer:
|
|
sciq-827
|
multiple_choice
|
What do farmers use greenhouses for?
|
[
"to predict the growing season",
"to decrease the growing season",
"to allow the growing season",
"to extend the growing season"
] |
D
|
Relavent Documents:
Document 0:::
The Crop Science Centre (informally known as 3CS during its planning) is an alliance between the University of Cambridge and National Institute of Agricultural Botany (NIAB).
History
The Crop Science Centre development plans began in 2015, between the University of Cambridge's Department of Plant Sciences, NIAB (National Institute of Agricultural Botany) and the Sainsbury Laboratory. The research institute received £16.9m funding in 2017 from the UK Research Partnership Investment Fund (UKRPIF) from Research England (United Kingdom Research and Innovation or UKRI) to build a new state-of-the-art building, designed exclusively for crop research, which opened on the 1st October 2020.
Structure
The Crop Science Centre is based at NIAB's Lawrence Weaver Road HQ site in Cambridge.
Document 1:::
A pre-STEM program is a course of study at any two-year college that prepares a student to transfer to a four-year school to earn a bachelor's degree in a STEM field.
Overview
The concept of a pre-STEM program is being developed to address America's need for more college-trained professionals in science, technology, engineering, and mathematics (STEM). It is an innovation meant to fill a gap at community colleges that do not have 'major' degree paths that students identify with on their way to earning an Associates degree. Students must complete a considerable amount of STEM coursework before transferring from a two-year school to a four-year school and earn a baccalaureate degree in a STEM field. Schools with a pre-STEM program are able to identify those students and support them with STEM-specific academic and career advising, increasing the student's chances of going on to earn a STEM baccalaureate degree in a timely fashion.
With over 50% of America's college-bound students starting their college career at public or private two-year school, and with a very small proportion of students who start college at a two-year school matriculating to and earning STEM degrees from four-year schools, pre-STEM programs have great potential for broadening participation in baccalaureate STEM studies.
Example programs
The effectiveness of pre-STEM programs is being investigated by a consortium of schools in Missouri: Moberly Area Community College, St. Charles Community College, Metropolitan Community College, and Truman State University.
A larger group of schools met at the Belknap Springs Meetings in October 2009 to discuss the challenges and opportunities presented by STEM-focused partnerships between 2-year and 4-year schools. Each program represented a two-year school and a four-year school that were trying to increase the number of people who earn a baccalaureate degree in a STEM area through various means, some of which were pre-STEM programs. Other methods includes
Document 2:::
Eric Toensmeier is an author of several books focused on climate change, biodiversity, and nutrition. He is also a lecturer at Yale University and a Senior Biosequestration Fellow at Project Drawdown.
Previously, he managed a farm program for Nuestras Raices, a nonprofit organization that ran community gardens in Holyoke, Massachusetts; and ran a mail-order seed company with Jonathan Bates for plants that would improve the soil or attract beneficial insects.
Career
Toensmeier's works include Paradise Lot, Perennial Vegetables: From Artichoke to ‘Zuiki’ Taro, a Gardener’s Guide to Over 100 Delicious, Easy-to-Grow Edibles, Edible Forests Gardens: Ecological Vision and Theory for Temperate Climate Permaculture (co-authored), and Perennial Vegetables: A neglected resource for biodiversity, carbon sequestration, and nutrition (co-authored).
In Perennial Vegetables: A neglected resource, Toensmeier provides information on crops that could help address malnutrition issues, including trees with edible, nutritious leaves. Perennial plants are those that grow all-year-round and do not require replanting or reseeding. They are used in regenerative and sustainable farming.
In February 2020, Toensmeier's 2016 book, Carbon Farming: A Global Toolkit for Stabilizing the Climate with Tree Crops and Regenerative Agriculture Practices, was released as an ebook. In the book, Toensmeier argues that carbon farming has the potential to return carbon dioxide in the atmosphere to 350 parts per million, while also providing food for people and regenerating soil. Carbon farming is defined as the use of agriculture to remove excess carbon from the air and soil by storing it in trees and plants. Toensmeier practices carbon farming at his home in Holyoke, Massachusetts.
Toensmeier is an adviser for Summersweet Gardens Nursery at Perennial Pleasures, whose owner was inspired by Toensmeier's book, Perennial Vegetables.
More recently, Toensmeier became a senior fellow at the climate think t
Document 3:::
The University of Florida Institute of Food and Agricultural Sciences (UF/IFAS) is a teaching, research and Extension scientific organization focused on agriculture and natural resources. It is a partnership of federal, state, and county governments that includes an Extension office in each of Florida's 67 counties, 12 off-campus research and education centers, five demonstration units, the University of Florida College of Agricultural and Life Sciences (including the School of Forest, Fisheries and Geomatics Sciences and the School of Natural Resources and Environment), three 4-H camps, portions of the UF College of Veterinary Medicine, the Florida Sea Grant program, the Emerging Pathogens Institute, the UF Water Institute and the UF Genetics Institute.
UF/IFAS research and development covers natural resource industries that have a $101 billion annual impact. The program is ranked #1 in the nation in federally financed higher education R&D expenditures in agricultural sciences and natural resources conservation by the National Science Foundation for FY 2019. Because of this mission and the diversity of Florida's climate and agricultural commodities, IFAS has facilities located throughout Florida. On July 13, 2020, Dr. J. Scott Angle became leader of UF/IFAS and UF's vice president for agriculture and natural resources.
History
Research
The mission of UF/IFAS is to develop knowledge in agricultural, human, and natural resources, and to make that knowledge accessible to sustain and enhance the quality of human life. Faculty members pursue fundamental and applied research that furthers understanding of natural and human systems. Research is supported by state and federally appropriated funds and supplemented by grants and contracts. UF/IFAS received $155.6 million in annual research expenditures in sponsored research for FY 2021.
The Florida Agricultural Experiment Station administers and supports research programs in UF/IFAS. The research program was created in
Document 4:::
Agrophysics is a branch of science bordering on agronomy and physics,
whose objects of study are the agroecosystem - the biological objects, biotope and biocoenosis affected by human activity, studied and described using the methods of physical sciences. Using the achievements of the exact sciences to solve major problems in agriculture, agrophysics involves the study of materials and processes occurring in the production and processing of agricultural crops, with particular emphasis on the condition of the environment and the quality of farming materials and food production.
Agrophysics is closely related to biophysics, but is restricted to the physics of the plants, animals, soil and an atmosphere involved in agricultural activities and biodiversity. It is different from biophysics in having the necessity of taking into account the specific features of biotope and biocoenosis, which involves the knowledge of nutritional science and agroecology, agricultural technology, biotechnology, genetics etc.
The needs of agriculture, concerning the past experience study of the local complex soil and next plant-atmosphere systems, lay at the root of the emergence of a new branch – agrophysics – dealing this with experimental physics.
The scope of the branch starting from soil science (physics) and originally limited to the study of relations within the soil environment, expanded over time onto influencing the properties of agricultural crops and produce as foods and raw postharvest materials, and onto the issues of quality, safety and labeling concerns, considered distinct from the field of nutrition for application in food science.
Research centres focused on the development of the agrophysical sciences include the Institute of Agrophysics, Polish Academy of Sciences in Lublin, and the Agrophysical Research Institute, Russian Academy of Sciences in St. Petersburg.
See also
Agriculture science
Agroecology
Genomics
Metagenomics
Metabolomics
Physics (Aristotle)
Proteomi
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What do farmers use greenhouses for?
A. to predict the growing season
B. to decrease the growing season
C. to allow the growing season
D. to extend the growing season
Answer:
|
|
sciq-8584
|
multiple_choice
|
A protein’s shape is determined, most fundamentally, by the sequence of what type of acids?
|
[
"amino",
"lipids",
"lactic",
"acetic"
] |
A
|
Relavent Documents:
Document 0:::
Amino acids are organic compounds that contain both amino and carboxylic acid functional groups. Although over 500 amino acids exist in nature, by far the most important are the 22 α-amino acids incorporated into proteins. Only these 22 appear in the genetic code of all life.
Amino acids can be classified according to the locations of the core structural functional groups, as alpha- (α-), beta- (β-), gamma- (γ-) or delta- (δ-) amino acids; other categories relate to polarity, ionization, and side chain group type (aliphatic, acyclic, aromatic, containing hydroxyl or sulfur, etc.). In the form of proteins, amino acid residues form the second-largest component (water being the largest) of human muscles and other tissues. Beyond their role as residues in proteins, amino acids participate in a number of processes such as neurotransmitter transport and biosynthesis. It is thought that they played a key role in enabling life on Earth and its emergence.
Amino acids are formally named by the IUPAC-IUBMB Joint Commission on Biochemical Nomenclature in terms of the fictitious "neutral" structure shown in the illustration. For example, the systematic name of alanine is 2-aminopropanoic acid, based on the formula . The Commission justified this approach as follows:
The systematic names and formulas given refer to hypothetical forms in which amino groups are unprotonated and carboxyl groups are undissociated. This convention is useful to avoid various nomenclatural problems but should not be taken to imply that these structures represent an appreciable fraction of the amino-acid molecules.
History
The first few amino acids were discovered in the early 1800s. In 1806, French chemists Louis-Nicolas Vauquelin and Pierre Jean Robiquet isolated a compound from asparagus that was subsequently named asparagine, the first amino acid to be discovered. Cystine was discovered in 1810, although its monomer, cysteine, remained undiscovered until 1884. Glycine and leucine were discovere
Document 1:::
Proteins are a class of macromolecular organic compounds that are essential to life. They consist of a long polypeptide chain that usually adopts a single stable three-dimensional structure. They fulfill a wide variety of functions including providing structural stability to cells, catalyze chemical reactions that produce or store energy or synthesize other biomolecules including nucleic acids and proteins, transport essential nutrients, or serve other roles such as signal transduction. They are selectively transported to various compartments of the cell or in some cases, secreted from the cell.
This list aims to organize information on how proteins are most often classified: by structure, by function, or by location.
Structure
Proteins may be classified as to their three-dimensional structure (also known a protein fold). The two most widely used classification schemes are:
CATH database
Structural Classification of Proteins database (SCOP)
Both classification schemes are based on a hierarchy of fold types. At the top level are all alpha proteins (domains consisting of alpha helices), all beta proteins (domains consisting of beta sheets), and mixed alpha helix/beta sheet proteins.
While most proteins adopt a single stable fold, a few proteins can rapidly interconvert between one or more folds. These are referred to as metamorphic proteins. Finally other proteins appear not to adopt any stable conformation and are referred to as intrinsically disordered.
Proteins frequently contain two or more domains, each have a different fold separated by intrinsically disordered regions. These are referred to as multi-domain proteins.
Function
Proteins may also be classified based on their celluar function. A widely used classification is PANTHER (protein analysis through evolutionary relationships) classification system.
Structural
Protein#Structural proteins
Catalytic
Enzymes classified according to their Enzyme Commission number (EC). Note that strictly speaki
Document 2:::
Biomolecular structure is the intricate folded, three-dimensional shape that is formed by a molecule of protein, DNA, or RNA, and that is important to its function. The structure of these molecules may be considered at any of several length scales ranging from the level of individual atoms to the relationships among entire protein subunits. This useful distinction among scales is often expressed as a decomposition of molecular structure into four levels: primary, secondary, tertiary, and quaternary. The scaffold for this multiscale organization of the molecule arises at the secondary level, where the fundamental structural elements are the molecule's various hydrogen bonds. This leads to several recognizable domains of protein structure and nucleic acid structure, including such secondary-structure features as alpha helixes and beta sheets for proteins, and hairpin loops, bulges, and internal loops for nucleic acids.
The terms primary, secondary, tertiary, and quaternary structure were introduced by Kaj Ulrik Linderstrøm-Lang in his 1951 Lane Medical Lectures at Stanford University.
Primary structure
The primary structure of a biopolymer is the exact specification of its atomic composition and the chemical bonds connecting those atoms (including stereochemistry). For a typical unbranched, un-crosslinked biopolymer (such as a molecule of a typical intracellular protein, or of DNA or RNA), the primary structure is equivalent to specifying the sequence of its monomeric subunits, such as amino acids or nucleotides.
The primary structure of a protein is reported starting from the amino N-terminus to the carboxyl C-terminus, while the primary structure of DNA or RNA molecule is known as the nucleic acid sequence reported from the 5' end to the 3' end.
The nucleic acid sequence refers to the exact sequence of nucleotides that comprise the whole molecule. Often, the primary structure encodes sequence motifs that are of functional importance. Some examples of such motif
Document 3:::
Amino acid replacement is a change from one amino acid to a different amino acid in a protein due to point mutation in the corresponding DNA sequence. It is caused by nonsynonymous missense mutation which changes the codon sequence to code other amino acid instead of the original.
Conservative and radical replacements
Not all amino acid replacements have the same effect on function or structure of protein. The magnitude of this process may vary depending on how similar or dissimilar the replaced amino acids are, as well as on their position in the sequence or the structure. Similarity between amino acids can be calculated based on substitution matrices, physico-chemical distance, or simple properties such as amino acid size or charge (see also amino acid chemical properties). Usually amino acids are thus classified into two types:
Conservative replacement - an amino acid is exchanged into another that has similar properties. This type of replacement is expected to rarely result in dysfunction in the corresponding protein .
Radical replacement - an amino acid is exchanged into another with different properties. This can lead to changes in protein structure or function, which can cause potentially lead to changes in phenotype, sometimes pathogenic. A well known example in humans is sickle cell anemia, due to a mutation in beta globin where at position 6 glutamic acid (negatively charged) is exchanged with valine (not charged).
Physicochemical distances
Physicochemical distance is a measure that assesses the difference between replaced amino acids. The value of distance is based on properties of amino acids. There are 134 physicochemical properties that can be used to estimate similarity between amino acids. Each physicochemical distance is based on different composition of properties.
Grantham's distance
Grantham's distance depends on three properties: composition, polarity and molecular volume.
Distance difference D for each pair of amino acid i and j is c
Document 4:::
PSMα
PSMα forms an amphipathic α-helix structure that composes the entire length of the peptide. These peptides are relatively short, being composed of only 20-25 amino acids. With regards to charge, α-typ
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
A protein’s shape is determined, most fundamentally, by the sequence of what type of acids?
A. amino
B. lipids
C. lactic
D. acetic
Answer:
|
|
scienceQA-12124
|
multiple_choice
|
Which of the following organisms is the tertiary consumer in this food web?
|
[
"pine vole",
"beaver",
"swallowtail caterpillar",
"bobcat"
] |
D
|
Tertiary consumers eat secondary consumers. So, in a food web, tertiary consumers have arrows pointing to them from secondary consumers. Secondary consumers have arrows pointing to them from primary consumers. And primary consumers have arrows pointing to them from producers.
The swallowtail caterpillar has an arrow pointing to it from the persimmon tree. The persimmon tree is not a secondary consumer, so the swallowtail caterpillar is not a tertiary consumer.
The pine vole has an arrow pointing to it from the persimmon tree and the swallowtail caterpillar. Neither the persimmon tree nor the swallowtail caterpillar is a secondary consumer, so the pine vole is not a tertiary consumer.
The bobcat has an arrow pointing to it from the gray fox. The gray fox is a secondary consumer, so the bobcat is a tertiary consumer.
The gray fox has an arrow pointing to it from the pine vole. The pine vole is a secondary consumer, so the gray fox is a tertiary consumer.
The beaver has an arrow pointing to it from the silver maple. The silver maple is not a secondary consumer, so the beaver is not a tertiary consumer.
|
Relavent Documents:
Document 0:::
Consumer–resource interactions are the core motif of ecological food chains or food webs, and are an umbrella term for a variety of more specialized types of biological species interactions including prey-predator (see predation), host-parasite (see parasitism), plant-herbivore and victim-exploiter systems. These kinds of interactions have been studied and modeled by population ecologists for nearly a century. Species at the bottom of the food chain, such as algae and other autotrophs, consume non-biological resources, such as minerals and nutrients of various kinds, and they derive their energy from light (photons) or chemical sources. Species higher up in the food chain survive by consuming other species and can be classified by what they eat and how they obtain or find their food.
Classification of consumer types
The standard categorization
Various terms have arisen to define consumers by what they eat, such as meat-eating carnivores, fish-eating piscivores, insect-eating insectivores, plant-eating herbivores, seed-eating granivores, and fruit-eating frugivores and omnivores are meat eaters and plant eaters. An extensive classification of consumer categories based on a list of feeding behaviors exists.
The Getz categorization
Another way of categorizing consumers, proposed by South African American ecologist Wayne Getz, is based on a biomass transformation web (BTW) formulation that organizes resources into five components: live and dead animal, live and dead plant, and particulate (i.e. broken down plant and animal) matter. It also distinguishes between consumers that gather their resources by moving across landscapes from those that mine their resources by becoming sessile once they have located a stock of resources large enough for them to feed on during completion of a full life history stage.
In Getz's scheme, words for miners are of Greek etymology and words for gatherers are of Latin etymology. Thus a bestivore, such as a cat, preys on live animal
Document 1:::
The trophic level of an organism is the position it occupies in a food web. A food chain is a succession of organisms that eat other organisms and may, in turn, be eaten themselves. The trophic level of an organism is the number of steps it is from the start of the chain. A food web starts at trophic level 1 with primary producers such as plants, can move to herbivores at level 2, carnivores at level 3 or higher, and typically finish with apex predators at level 4 or 5. The path along the chain can form either a one-way flow or a food "web". Ecological communities with higher biodiversity form more complex trophic paths.
The word trophic derives from the Greek τροφή (trophē) referring to food or nourishment.
History
The concept of trophic level was developed by Raymond Lindeman (1942), based on the terminology of August Thienemann (1926): "producers", "consumers", and "reducers" (modified to "decomposers" by Lindeman).
Overview
The three basic ways in which organisms get food are as producers, consumers, and decomposers.
Producers (autotrophs) are typically plants or algae. Plants and algae do not usually eat other organisms, but pull nutrients from the soil or the ocean and manufacture their own food using photosynthesis. For this reason, they are called primary producers. In this way, it is energy from the sun that usually powers the base of the food chain. An exception occurs in deep-sea hydrothermal ecosystems, where there is no sunlight. Here primary producers manufacture food through a process called chemosynthesis.
Consumers (heterotrophs) are species that cannot manufacture their own food and need to consume other organisms. Animals that eat primary producers (like plants) are called herbivores. Animals that eat other animals are called carnivores, and animals that eat both plants and other animals are called omnivores.
Decomposers (detritivores) break down dead plant and animal material and wastes and release it again as energy and nutrients into
Document 2:::
Grazing is a method of feeding in which a herbivore feeds on low-growing plants such as grasses or other multicellular organisms, such as algae. Many species of animals can be said to be grazers, from large animals such as hippopotamuses to small aquatic snails. Grazing behaviour is a type of feeding strategy within the ecology of a species. Specific grazing strategies include graminivory (eating grasses); coprophagy (producing part-digested pellets which are reingested); pseudoruminant (having a multi-chambered stomach but not chewing the cud); and grazing on plants other than grass, such as on marine algae.
Grazing's ecological effects can include redistributing nutrients, keeping grasslands open or favouring a particular species over another.
Ecology
Many small selective herbivores follow larger grazers which skim off the highest, tough growth of grasses, exposing tender shoots. For terrestrial animals, grazing is normally distinguished from browsing in that grazing is eating grass or forbs, whereas browsing is eating woody twigs and leaves from trees and shrubs. Grazing differs from predation because the organism being grazed upon may not be killed. It differs from parasitism because the two organisms live together in a constant state of physical externality (i.e. low intimacy). Water animals that feed by rasping algae and other micro-organisms from stones are called grazers–scrapers.
Graminivory
Graminivory is a form of grazing involving feeding primarily on grass (specifically "true" grasses in the Poaceae). Horses, cattle, capybara, hippopotamuses, grasshoppers, geese, and giant pandas are graminivores. Giant pandas (Ailuropoda melanoleuca) are obligate bamboo grazers, 99% of their diet consisting of sub-alpine bamboo species.
Coprophagy
Rabbits are herbivores that feed by grazing on grass, forbs, and leafy weeds. They graze heavily and rapidly for about the first half-hour of a grazing period (usually in the late afternoon), followed by about half an
Document 3:::
The soil food web is the community of organisms living all or part of their lives in the soil. It describes a complex living system in the soil and how it interacts with the environment, plants, and animals.
Food webs describe the transfer of energy between species in an ecosystem. While a food chain examines one, linear, energy pathway through an ecosystem, a food web is more complex and illustrates all of the potential pathways. Much of this transferred energy comes from the sun. Plants use the sun’s energy to convert inorganic compounds into energy-rich, organic compounds, turning carbon dioxide and minerals into plant material by photosynthesis. Plant flowers exude energy-rich nectar above ground and plant roots exude acids, sugars, and ectoenzymes into the rhizosphere, adjusting the pH and feeding the food web underground.
Plants are called autotrophs because they make their own energy; they are also called producers because they produce energy available for other organisms to eat. Heterotrophs are consumers that cannot make their own food. In order to obtain energy they eat plants or other heterotrophs.
Above ground food webs
In above ground food webs, energy moves from producers (plants) to primary consumers (herbivores) and then to secondary consumers (predators). The phrase, trophic level, refers to the different levels or steps in the energy pathway. In other words, the producers, consumers, and decomposers are the main trophic levels. This chain of energy transferring from one species to another can continue several more times, but eventually ends. At the end of the food chain, decomposers such as bacteria and fungi break down dead plant and animal material into simple nutrients.
Methodology
The nature of soil makes direct observation of food webs difficult. Since soil organisms range in size from less than 0.1 mm (nematodes) to greater than 2 mm (earthworms) there are many different ways to extract them. Soil samples are often taken using a metal
Document 4:::
A herbivore is an animal anatomically and physiologically adapted to eating plant material, for example foliage or marine algae, for the main component of its diet. As a result of their plant diet, herbivorous animals typically have mouthparts adapted to rasping or grinding. Horses and other herbivores have wide flat teeth that are adapted to grinding grass, tree bark, and other tough plant material.
A large percentage of herbivores have mutualistic gut flora that help them digest plant matter, which is more difficult to digest than animal prey. This flora is made up of cellulose-digesting protozoans or bacteria.
Etymology
Herbivore is the anglicized form of a modern Latin coinage, herbivora, cited in Charles Lyell's 1830 Principles of Geology. Richard Owen employed the anglicized term in an 1854 work on fossil teeth and skeletons. Herbivora is derived from Latin herba 'small plant, herb' and vora, from vorare 'to eat, devour'.
Definition and related terms
Herbivory is a form of consumption in which an organism principally eats autotrophs such as plants, algae and photosynthesizing bacteria. More generally, organisms that feed on autotrophs in general are known as primary consumers.
Herbivory is usually limited to animals that eat plants. Insect herbivory can cause a variety of physical and metabolic alterations in the way the host plant interacts with itself and other surrounding biotic factors. Fungi, bacteria, and protists that feed on living plants are usually termed plant pathogens (plant diseases), while fungi and microbes that feed on dead plants are described as saprotrophs. Flowering plants that obtain nutrition from other living plants are usually termed parasitic plants. There is, however, no single exclusive and definitive ecological classification of consumption patterns; each textbook has its own variations on the theme.
Evolution of herbivory
The understanding of herbivory in geological time comes from three sources: fossilized plants, which may
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Which of the following organisms is the tertiary consumer in this food web?
A. pine vole
B. beaver
C. swallowtail caterpillar
D. bobcat
Answer:
|
sciq-882
|
multiple_choice
|
The heart, blood vessels, and blood make up which system in the body?
|
[
"pulminary",
"cardiovascular",
"respiratory",
"digestive"
] |
B
|
Relavent Documents:
Document 0:::
The blood circulatory system is a system of organs that includes the heart, blood vessels, and blood which is circulated throughout the entire body of a human or other vertebrate. It includes the cardiovascular system, or vascular system, that consists of the heart and blood vessels (from Greek kardia meaning heart, and from Latin vascula meaning vessels). The circulatory system has two divisions, a systemic circulation or circuit, and a pulmonary circulation or circuit. Some sources use the terms cardiovascular system and vascular system interchangeably with the circulatory system.
The network of blood vessels are the great vessels of the heart including large elastic arteries, and large veins; other arteries, smaller arterioles, capillaries that join with venules (small veins), and other veins. The circulatory system is closed in vertebrates, which means that the blood never leaves the network of blood vessels. Some invertebrates such as arthropods have an open circulatory system. Diploblasts such as sponges, and comb jellies lack a circulatory system.
Blood is a fluid consisting of plasma, red blood cells, white blood cells, and platelets; it is circulated around the body carrying oxygen and nutrients to the tissues and collecting and disposing of waste materials. Circulated nutrients include proteins and minerals and other components include hemoglobin, hormones, and gases such as oxygen and carbon dioxide. These substances provide nourishment, help the immune system to fight diseases, and help maintain homeostasis by stabilizing temperature and natural pH.
In vertebrates, the lymphatic system is complementary to the circulatory system. The lymphatic system carries excess plasma (filtered from the circulatory system capillaries as interstitial fluid between cells) away from the body tissues via accessory routes that return excess fluid back to blood circulation as lymph. The lymphatic system is a subsystem that is essential for the functioning of the bloo
Document 1:::
Cardiovascular physiology is the study of the cardiovascular system, specifically addressing the physiology of the heart ("cardio") and blood vessels ("vascular").
These subjects are sometimes addressed separately, under the names cardiac physiology and circulatory physiology.
Although the different aspects of cardiovascular physiology are closely interrelated, the subject is still usually divided into several subtopics.
Heart
Cardiac output (= heart rate * stroke volume. Can also be calculated with Fick principle, palpating method.)
Stroke volume (= end-diastolic volume − end-systolic volume)
Ejection fraction (= stroke volume / end-diastolic volume)
Cardiac output is mathematically ` to systole
Inotropic, chronotropic, and dromotropic states
Cardiac input (= heart rate * suction volume Can be calculated by inverting terms in Fick principle)
Suction volume (= end-systolic volume + end-diastolic volume)
Injection fraction (=suction volume / end-systolic volume)
Cardiac input is mathematically ` to diastole
Electrical conduction system of the heart
Electrocardiogram
Cardiac marker
Cardiac action potential
Frank–Starling law of the heart
Wiggers diagram
Pressure volume diagram
Regulation of blood pressure
Baroreceptor
Baroreflex
Renin–angiotensin system
Renin
Angiotensin
Juxtaglomerular apparatus
Aortic body and carotid body
Autoregulation
Cerebral Autoregulation
Hemodynamics
Under most circumstances, the body attempts to maintain a steady mean arterial pressure.
When there is a major and immediate decrease (such as that due to hemorrhage or standing up), the body can increase the following:
Heart rate
Total peripheral resistance (primarily due to vasoconstriction of arteries)
Inotropic state
In turn, this can have a significant impact upon several other variables:
Stroke volume
Cardiac output
Pressure
Pulse pressure (systolic pressure - diastolic pressure)
Mean arterial pressure (usually approximated with diastolic pressure +
Document 2:::
Cardiophysics is an interdisciplinary science that stands at the junction of cardiology and medical physics, with researchers using the methods of, and theories from, physics to study cardiovascular system at different levels of its organisation, from the molecular scale to whole organisms. Being formed historically as part of systems biology, cardiophysics designed to reveal connections between the physical mechanisms, underlying the organization of the cardiovascular system, and biological features of its functioning.
Zbigniew R. Struzik seems to be a first author who used the term in a scientific publication in 2004.
One can use interchangeably also the terms cardiovascular physics.
See also
Medical physics
Important publications in medical physics
Biomedicine
Biomedical engineering
Physiome
Nanomedicine
Document 3:::
A biological system is a complex network which connects several biologically relevant entities. Biological organization spans several scales and are determined based different structures depending on what the system is. Examples of biological systems at the macro scale are populations of organisms. On the organ and tissue scale in mammals and other animals, examples include the circulatory system, the respiratory system, and the nervous system. On the micro to the nanoscopic scale, examples of biological systems are cells, organelles, macromolecular complexes and regulatory pathways. A biological system is not to be confused with a living system, such as a living organism.
Organ and tissue systems
These specific systems are widely studied in human anatomy and are also present in many other animals.
Respiratory system: the organs used for breathing, the pharynx, larynx, bronchi, lungs and diaphragm.
Digestive system: digestion and processing food with salivary glands, oesophagus, stomach, liver, gallbladder, pancreas, intestines, rectum and anus.
Cardiovascular system (heart and circulatory system): pumping and channeling blood to and from the body and lungs with heart, blood and blood vessels.
Urinary system: kidneys, ureters, bladder and urethra involved in fluid balance, electrolyte balance and excretion of urine.
Integumentary system: skin, hair, fat, and nails.
Skeletal system: structural support and protection with bones, cartilage, ligaments and tendons.
Endocrine system: communication within the body using hormones made by endocrine glands such as the hypothalamus, pituitary gland, pineal body or pineal gland, thyroid, parathyroid and adrenals, i.e., adrenal glands.
Lymphatic system: structures involved in the transfer of lymph between tissues and the blood stream; includes the lymph and the nodes and vessels. The lymphatic system includes functions including immune responses and development of antibodies.
Immune system: protects the organism from
Document 4:::
The human body is the structure of a human being. It is composed of many different types of cells that together create tissues and subsequently organs and then organ systems. They ensure homeostasis and the viability of the human body.
It comprises a head, hair, neck, torso (which includes the thorax and abdomen), arms and hands, legs and feet.
The study of the human body includes anatomy, physiology, histology and embryology. The body varies anatomically in known ways. Physiology focuses on the systems and organs of the human body and their functions. Many systems and mechanisms interact in order to maintain homeostasis, with safe levels of substances such as sugar and oxygen in the blood.
The body is studied by health professionals, physiologists, anatomists, and artists to assist them in their work.
Composition
The human body is composed of elements including hydrogen, oxygen, carbon, calcium and phosphorus. These elements reside in trillions of cells and non-cellular components of the body.
The adult male body is about 60% water for a total water content of some . This is made up of about of extracellular fluid including about of blood plasma and about of interstitial fluid, and about of fluid inside cells. The content, acidity and composition of the water inside and outside cells is carefully maintained. The main electrolytes in body water outside cells are sodium and chloride, whereas within cells it is potassium and other phosphates.
Cells
The body contains trillions of cells, the fundamental unit of life. At maturity, there are roughly 3037trillion cells in the body, an estimate arrived at by totaling the cell numbers of all the organs of the body and cell types. The body is also host to about the same number of non-human cells as well as multicellular organisms which reside in the gastrointestinal tract and on the skin. Not all parts of the body are made from cells. Cells sit in an extracellular matrix that consists of proteins such as collagen,
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
The heart, blood vessels, and blood make up which system in the body?
A. pulminary
B. cardiovascular
C. respiratory
D. digestive
Answer:
|
|
sciq-70
|
multiple_choice
|
What phenomenon is crowding out other species and making all the other causes of extinction worse?
|
[
"migration",
"human suburbanization",
"human overpopulation",
"climate change"
] |
C
|
Relavent Documents:
Document 0:::
Biodiversity loss includes the worldwide extinction of different species, as well as the local reduction or loss of species in a certain habitat, resulting in a loss of biological diversity. The latter phenomenon can be temporary or permanent, depending on whether the environmental degradation that leads to the loss is reversible through ecological restoration/ecological resilience or effectively permanent (e.g. through land loss). The current global extinction (frequently called the sixth mass extinction or Anthropocene extinction), has resulted in a biodiversity crisis being driven by human activities which push beyond the planetary boundaries and so far has proven irreversible.
The main direct threats to conservation (and thus causes for biodiversity loss) fall in eleven categories: Residential and commercial development; farming activities; energy production and mining; transportation and service corridors; biological resource usages; human intrusions and activities that alter, destroy, disturb habitats and species from exhibiting natural behaviors; natural system modification; invasive and problematic species, pathogens and genes; pollution; catastrophic geological events, climate change, and so on.
Numerous scientists and the IPBES Global Assessment Report on Biodiversity and Ecosystem Services assert that human population growth and overconsumption are the primary factors in this decline. However other scientists have criticized this, saying that loss of habitat is caused mainly by "the growth of commodities for export" and that population has very little to do with overall consumption, due to country wealth disparities.
Climate change is another threat to global biodiversity. For example, coral reefs – which are biodiversity hotspots – will be lost within the century if global warming continues at the current rate. However, habitat destruction e.g. for the expansion of agriculture, is currently the more significant driver of contemporary biodiversity lo
Document 1:::
In nature and human societies, many phenomena have causal relationships where one phenomenon A (a cause) impacts another phenomenon B (an effect). Establishing causal relationships is the aim of many scientific studies across fields ranging from biology and physics to social sciences and economics. It is also a subject of accident analysis, and can be considered a prerequisite for effective policy making.
To describe causal relationships between phenomena, non-quantitative visual notations are common, such as arrows, e.g. in the nitrogen cycle or many chemistry and mathematics textbooks. Mathematical conventions are also used, such as plotting an independent variable on a horizontal axis and a dependent variable on a vertical axis, or the notation to denote that a quantity "" is a dependent variable which is a function of an independent variable "". Causal relationships are also described using quantitative mathematical expressions.
The following examples illustrate various types of causal relationships. These are followed by different notations used to represent causal relationships.
Examples
What follows does not necessarily assume the convention whereby denotes an independent variable, and
denotes a function of the independent variable . Instead, and denote two quantities with an a priori unknown causal relationship, which can be related by a mathematical expression.
Ecosystem example: correlation without causation
Imagine the number of days of weather below zero degrees Celsius, , causes ice to form on a lake, , and it causes bears to go into hibernation . Even though does not cause and vice-versa, one can write an equation relating and . This equation may be used to successfully calculate the number of hibernating bears , given the surface area of the lake covered by ice. However, melting the ice in a region of the lake by pouring salt onto it, will not cause bears to come out of hibernation. Nor will waking the bears by physically disturbing the
Document 2:::
Overpopulation or overabundance is a phenomenon in which a species' population becomes larger than the carrying capacity of its environment. This may be caused by increased birth rates, lowered mortality rates, reduced predation or large scale migration, leading to an overabundant species and other animals in the ecosystem competing for food, space, and resources. The animals in an overpopulated area may then be forced to migrate to areas not typically inhabited, or die off without access to necessary resources.
Judgements regarding overpopulation always involve both facts and values. Animals often are judged overpopulated when their numbers cause impacts that people find dangerous, damaging, expensive, or otherwise harmful. Societies may be judged overpopulated when their human numbers cause impacts that degrade ecosystem services, decrease human health and well-being, or crowd other species out of existence.
Background
In ecology, overpopulation is a concept used primarily in wildlife management. Typically, an overpopulation causes the entire population of the species in question to become weaker, as no single individual is able to find enough food or shelter. As such, overpopulation is thus characterized by an increase in the diseases and parasite-load which live upon the species in question, as the entire population is weaker. Other characteristics of overpopulation are lower fecundity, adverse effects on the environment (soil, vegetation or fauna) and lower average body weights. Especially the worldwide increase of deer populations, which usually show irruptive growth, is proving to be of ecological concern. Ironically, where ecologists were preoccupied with conserving or augmenting deer populations only a century ago, the focus has now shifted in the direct opposite, and ecologists are now more concerned with limiting the populations of such animals.
Supplemental feeding of charismatic species or interesting game species is a major problem in causing overp
Document 3:::
Background extinction rate, also known as the normal extinction rate, refers to the standard rate of extinction in Earth's geological and biological history before humans became a primary contributor to extinctions. This is primarily the pre-human extinction rates during periods in between major extinction events. Currently there have been five mass extinctions that have happened since the beginning of time all resulting in a variety of reasons.
Overview
Extinctions are a normal part of the evolutionary process, and the background extinction rate is a measurement of "how often" they naturally occur. Normal extinction rates are often used as a comparison to present day extinction rates, to illustrate the higher frequency of extinction today than in all periods of non-extinction events before it.
Background extinction rates have not remained constant, although changes are measured over geological time, covering millions of years.
Measurement
Background extinction rates are typically measured in order to give a specific classification to a species and this is obtained over a certain period of time. There is three different ways to calculate background extinction rate.. The first is simply the number of species that normally go extinct over a given period of time. For example, at the background rate one species of bird will go extinct every estimated 400 years. Another way the extinction rate can be given is in million species years (MSY). For example, there is approximately one extinction estimated per million species years. From a purely mathematical standpoint this means that if there are a million species on the planet earth, one would go extinct every year, while if there was only one species it would go extinct in one million years, etc. The third way is in giving species survival rates over time. For example, given normal extinction rates species typically exist for 5–10 million years before going extinct.
Lifespan estimates
Some species lifespan es
Document 4:::
The Sixth Extinction: An Unnatural History is a 2014 non-fiction book written by Elizabeth Kolbert and published by Henry Holt and Company. The book argues that the Earth is in the midst of a modern, man-made, sixth extinction. In the book, Kolbert chronicles previous mass extinction events, and compares them to the accelerated, widespread extinctions during our present time. She also describes specific species extinguished by humans, as well as the ecologies surrounding prehistoric and near-present extinction events. The author received the Pulitzer Prize for General Non-Fiction for the book in 2015.
The target audience is the general reader, and scientific descriptions are rendered in understandable prose. The writing blends explanations of her treks to remote areas with interviews of scientists, researchers, and guides, without advocating a position, in pursuit of objectivity. Hence, the sixth mass extinction theme is applied to flora and fauna existing in diverse habitats, such as the Panamanian rainforest, the Great Barrier Reef, the Andes, Bikini Atoll, city zoos, and the author's own backyard. The book also applies this theme to a number of other habitats and organisms throughout the world. After researching the current mainstream view of the relevant peer-reviewed science, Kolbert estimates flora and fauna loss by the end of the 21st century to be between 20 and 50 percent "of all living species on earth".
Anthropocene
Kolbert equates current, general unawareness of this issue to previous widespread disbelief of it during the centuries preceding the late 1700s; at that time, it was believed that prehistoric mass extinctions had never occurred. It was also believed there were no natural forces powerful enough to extinguish species en masse. Likewise, in our own time, the possible finality presented by this issue results in denialism. But scientific studies have shown that human behavior disrupts Earth's balanced and interconnected systems, "putting our own
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What phenomenon is crowding out other species and making all the other causes of extinction worse?
A. migration
B. human suburbanization
C. human overpopulation
D. climate change
Answer:
|
|
sciq-10185
|
multiple_choice
|
Induced mutations are caused by environmental factors. what term is used to describe a mutation that has no environmental cause?
|
[
"evident mutation",
"numerous mutation",
"cancerous mutation",
"spontaneous mutation"
] |
D
|
Relavent Documents:
Document 0:::
In biology, and especially in genetics, a mutant is an organism or a new genetic character arising or resulting from an instance of mutation, which is generally an alteration of the DNA sequence of the genome or chromosome of an organism. It is a characteristic that would not be observed naturally in a specimen. The term mutant is also applied to a virus with an alteration in its nucleotide sequence whose genome is in the nuclear genome. The natural occurrence of genetic mutations is integral to the process of evolution. The study of mutants is an integral part of biology; by understanding the effect that a mutation in a gene has, it is possible to establish the normal function of that gene.
Mutants arise by mutation
Mutants arise by mutations occurring in pre-existing genomes as a result of errors of DNA replication or errors of DNA repair. Errors of replication often involve translesion synthesis by a DNA polymerase when it encounters and bypasses a damaged base in the template strand. A DNA damage is an abnormal chemical structure in DNA, such as a strand break or an oxidized base, whereas a mutation, by contrast, is a change in the sequence of standard base pairs. Errors of repair occur when repair processes inaccurately replace a damaged DNA sequence. The DNA repair process microhomology-mediated end joining is particularly error-prone.
Etymology
Although not all mutations have a noticeable phenotypic effect, the common usage of the word "mutant" is generally a pejorative term, only used for genetically or phenotypically noticeable mutations. Previously, people used the word "sport" (related to spurt) to refer to abnormal specimens. The scientific usage is broader, referring to any organism differing from the wild type. The word finds its origin in the Latin term mūtant- (stem of mūtāns), which means "to change".
Mutants should not be confused with organisms born with developmental abnormalities, which are caused by errors during morphogenesis. In a devel
Document 1:::
Neutral mutations are changes in DNA sequence that are neither beneficial nor detrimental to the ability of an organism to survive and reproduce. In population genetics, mutations in which natural selection does not affect the spread of the mutation in a species are termed neutral mutations. Neutral mutations that are inheritable and not linked to any genes under selection will be lost or will replace all other alleles of the gene. That loss or fixation of the gene proceeds based on random sampling known as genetic drift. A neutral mutation that is in linkage disequilibrium with other alleles that are under selection may proceed to loss or fixation via genetic hitchhiking and/or background selection.
While many mutations in a genome may decrease an organism’s ability to survive and reproduce, also known as fitness, those mutations are selected against and are not passed on to future generations. The most commonly-observed mutations that are detectable as variation in the genetic makeup of organisms and populations appear to have no visible effect on the fitness of individuals and are therefore neutral. The identification and study of neutral mutations has led to the development of the neutral theory of molecular evolution, which is an important and often-controversial theory that proposes that most molecular variation within and among species is essentially neutral and not acted on by selection. Neutral mutations are also the basis for using molecular clocks to identify such evolutionary events as speciation and adaptive or evolutionary radiations.
History
Charles Darwin commented on the idea of neutral mutation in his work, hypothesizing that mutations that do not give an advantage or disadvantage may fluctuate or become fixed apart from natural selection. "Variations neither useful nor injurious would not be affected by natural selection, and would be left either a fluctuating element, as perhaps we see in certain polymorphic species, or would ultimately become
Document 2:::
A mutant protein is the protein product encoded by a gene with mutation. Mutated protein can have single amino acid change (minor, but still in many cases significant change leading to disease) or wide-range amino acid changes by e.g. truncation of C-terminus after introducing premature stop codon.
See also
Site-directed mutagenesis
Phi value analysis
missense mutation
nonsense mutation
point mutation
frameshift mutation
silent mutation
single-nucleotide polymorphism
Document 3:::
Biological tests of necessity and sufficiency refer to experimental methods and techniques that seek to test or provide evidence for specific kinds of causal relationships in biological systems. A necessary cause is one without which it would be impossible for an effect to occur, while a sufficient cause is one whose presence guarantees the occurrence of an effect. These concepts are largely based on but distinct from ideas of necessity and sufficiency in logic.
Tests of necessity, among which are methods of lesioning or gene knockout, and tests of sufficiency, among which are methods of isolation or discrete stimulation of factors, have become important in current-day experimental designs, and application of these tests have led to a number of notable discoveries and findings in the biological sciences.
Definitions
In biological research, experiments or tests are often used to study predicted causal relationships between two phenomena. These causal relationships may be described in terms of the logical concepts of necessity and sufficiency.
Consider the statement that a phenomenon x causes a phenomenon y. X would be a necessary cause of y when the occurrence of y implies that x needed to have occurred. However, only the occurrence of the necessary condition x may not always result in y also occurring. In other words, when some factor is necessary to cause an effect, it is impossible to have the effect without the cause. X would instead be a sufficient cause of y when the occurrence of x implies that y must then occur. in other words, when some factor is sufficient to cause an effect, the presence of the cause guarantees the occurrence of the effect. However, a different cause z may also cause y, meaning that y may occur without x occurring. For a concrete example, consider the conditional statement "if an object is a square, then it has four sides". It is a necessary condition that an object has four sides if it is true that it is a square; conversely, the obj
Document 4:::
A mutation accumulation (MA) experiment is a genetic experiment in which isolated and inbred lines of organisms (so-called MA lines) are maintained such that the effect of natural selection is minimized, with the aim of quantitatively estimating the rates at which spontaneous mutations (mutations not caused by exogenous mutagens) occur in the studied organism. Spontaneous mutation rates may be directly estimated using molecular techniques such as DNA sequencing, or indirectly estimated using phenotypic assays (observing how an organism’s phenotype changes as mutations accumulate).
The earliest mutation accumulation experiments were performed by American geneticist Hermann Joseph Muller in the 1920s, using Drosophila melanogaster.
Principles and procedures
All MA lines used in a MA experiment are bred from a single common ancestor, and are often propagated by single-progeny descent, where a single offspring is randomly selected to sire the next generation of organisms. This serves to prevent the loss of mutant alleles through sexual reproduction. Notably, single-progeny descent is only possible if the organism being studied is capable of asexual reproduction or self-fertilization.
A control line is maintained parallel to the MA lines and under the same conditions, except organisms are allowed to reproduce sexually (they are not constrained to single-progeny descent). The assumption underlying this procedure is that the larger, sexually reproducing population of the control line will cause all spontaneous mutations to be ‘weeded out’ by sexual reproduction. Mutations that arise in MA lines are heterozygous at first, and can become fixed or lost at random in subsequent generations. Thus, the control line will be relatively free of mutations, and can be compared with the MA lines to assess the impact of the mutations that have accumulated therein. Both the MA lines and the control line are maintained under relaxed natural selection, to minimize the strain that n
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Induced mutations are caused by environmental factors. what term is used to describe a mutation that has no environmental cause?
A. evident mutation
B. numerous mutation
C. cancerous mutation
D. spontaneous mutation
Answer:
|
|
ai2_arc-1111
|
multiple_choice
|
Currently, Antarctica is the coldest and driest place on Earth. Paleontologists have found dinosaur fossils in rock layers below the ice. Based on this evidence, which is a reasonable conclusion about the climate of Antarctica in the past?
|
[
"Antarctica has always had the same climate as at present.",
"At some time in the past, Antarctica was a hot, dry desert.",
"At some time in the past, Antarctica was much warmer and wetter.",
"At some time in the past, Antarctica was colder and wetter than it is now."
] |
C
|
Relavent Documents:
Document 0:::
The history of life on Earth is closely associated with environmental change on multiple spatial and temporal scales. Climate change is a long-term change in the average weather patterns that have come to define Earth’s local, regional and global climates. These changes have a broad range of observed effects that are synonymous with the term. Climate change is any significant long term change in the expected pattern, whether due to natural variability or as a result of human activity. Predicting the effects that climate change will have on plant biodiversity can be achieved using various models, however bioclimatic models are most commonly used.
Environmental conditions play a key role in defining the function and geographic distributions of plants, in combination with other factors, thereby modifying patterns of biodiversity. Changes in long term environmental conditions that can be collectively coined climate change are known to have had enormous impacts on current plant diversity patterns; further impacts are expected in the future. It is predicted that climate change will remain one of the major drivers of biodiversity patterns in the future. Climate change is thought to be one of several factors causing the currently ongoing human-triggered mass extinction, which is changing the distribution and abundance of many plants.
Palaeo context
The Earth has experienced a constantly changing climate in the time since plants first evolved. In comparison to the present day, this history has seen Earth as cooler, warmer, drier and wetter, and (carbon dioxide) concentrations have been both higher and lower. These changes have been reflected by constantly shifting vegetation, for example forest communities dominating most areas in interglacial periods, and herbaceous communities dominating during glacial periods. It has been shown through fossil records that past climatic change has been a major driver of the processes of speciation and extinction. The best known example
Document 1:::
Blood Falls is an outflow of an iron oxide–tainted plume of saltwater, flowing from the tongue of Taylor Glacier onto the ice-covered surface of West Lake Bonney in the Taylor Valley of the McMurdo Dry Valleys in Victoria Land, East Antarctica.
Iron-rich hypersaline water sporadically emerges from small fissures in the ice cascades. The saltwater source is a subglacial pool of unknown size overlain by about of ice several kilometers from its tiny outlet at Blood Falls.
The reddish deposit was found in 1911 by the Australian geologist Thomas Griffith Taylor, who first explored the valley that bears his name. The Antarctica pioneers first attributed the red color to red algae, but later it was proven to be due to iron oxides.
Geochemistry
Poorly soluble hydrous ferric oxides are deposited at the surface of ice after the ferrous ions present in the unfrozen saltwater are oxidized in contact with atmospheric oxygen. The more soluble ferrous ions initially are dissolved in old seawater trapped in an ancient pocket remaining from the Antarctic Ocean when a fjord was isolated by the glacier in its progression during the Miocene period, some 5 million years ago, when the sea level was higher than today.
Unlike most Antarctic glaciers, the Taylor Glacier is not frozen to the bedrock, probably because of the presence of salts concentrated by the crystallization of the ancient seawater imprisoned below it. Salt cryo-concentration occurred in the deep relict seawater when pure ice crystallized and expelled its dissolved salts as it cooled down because of the heat exchange of the captive liquid seawater with the enormous ice mass of the glacier. As a consequence, the trapped seawater was concentrated in brines with a salinity two to three times that of the mean ocean water. A second mechanism sometimes also explaining the formation of hypersaline brines is the water evaporation of surface lakes directly exposed to the very dry polar atmosphere in the McMurdo Dry Valleys. Th
Document 2:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 3:::
The Older Dryas was a stadial (cold) period between the Bølling and Allerød interstadials (warmer phases), about 14,000 years Before Present, towards the end of the Pleistocene. Its date is not well defined, with estimates varying by 400 years, but its duration is agreed to have been around 200 years.
The gradual warming since the Last Glacial Maximum (27,000 to 24,000 years BP) has been interrupted by two cold spells: the Older Dryas and the Younger Dryas (c. 12,900–11,650 BP). In northern Scotland, the glaciers were thicker and deeper during the Older Dryas than the succeeding Younger Dryas, and there is no evidence of human occupation of Britain. In Northwestern Europe there was also an earlier Oldest Dryas (18.5–17 ka BP 15–14 ka BP). The Dryas are named after an indicator genus, the Arctic and Alpine plant Dryas, the remains of which are found in higher concentrations in deposits from colder periods.
The Older Dryas was a variable cold, dry Blytt–Sernander period, observed in climatological evidence in only some regions, dependent on latitude. In regions in which it is not observed, the Bølling–Allerød is considered a single interstadial period. Evidence of the Older Dryas is strongest in northern Eurasia, particularly part of Northern Europe, roughly equivalent to Pollen zone Ic.
Dates
In the Greenland oxygen isotope record, the Older Dryas appears as a downward peak establishing a small, low-intensity gap between the Bølling and the Allerød. That configuration presents a difficulty in estimating its time, as it is more of a point than a segment. The segment is small enough to escape the resolution of most carbon-14 series, as the points are not close enough together to find the segment.
One approach to the problem assigns a point and then picks an arbitrary segment. The Older Dryas is sometimes considered to be "centered" near 14,100 BP or to be 100 to 150 years long "at" 14,250 BP.
A second approach finds carbon-14 or other dates as close to the end of
Document 4:::
Glacio-geological databases compile data on glacially associated sedimentary deposits and erosional activity from former and current ice-sheets, usually from published peer-reviewed sources. Their purposes are generally directed towards two ends: (Mode 1) compiling information about glacial landforms, which often inform about former ice-flow directions; and (Mode 2) compiling information which dates the absence or presence of ice.
These databases are used for a variety of purposes: (i) as bibliographic tools for researchers; (ii) as the quantitative basis of mapping of landforms or dates of ice presence/absence; and (iii) as quantitative databases which are used to constrain physically based mathematical models of ice-sheets.
Antarctic Ice Sheet: The AGGDB is a Mode 2 glacio-geological database for the Antarctic ice-sheet using information from around 150 published sources, covering glacial activity mainly from the past 30,000 years. It is available online, and aims to be comprehensive to the end of 2007.
British Ice Sheet: BRITICE is a Mode 1 database which aims to map all glacial landforms of Great Britain.
Eurasian Ice Sheet: DATED-1 is a Mode 2 database for the Eurasian ice-sheet. Its sister-project DATED-2 uses the information in DATED-1 to map the retreat of the Eurasian ice-sheet since the Last Glacial Maximum.
See also
Glacial landforms
Sediment
Geology
Ice sheet
Exposure Age Dating
Radio-carbon dating
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Currently, Antarctica is the coldest and driest place on Earth. Paleontologists have found dinosaur fossils in rock layers below the ice. Based on this evidence, which is a reasonable conclusion about the climate of Antarctica in the past?
A. Antarctica has always had the same climate as at present.
B. At some time in the past, Antarctica was a hot, dry desert.
C. At some time in the past, Antarctica was much warmer and wetter.
D. At some time in the past, Antarctica was colder and wetter than it is now.
Answer:
|
|
scienceQA-1919
|
multiple_choice
|
Select the invertebrate.
|
[
"lionfish",
"penguin",
"julia butterfly",
"grizzly bear"
] |
C
|
A lionfish is a fish. Like other fish, a lionfish is a vertebrate. It has a backbone.
A penguin is a bird. Like other birds, a penguin is a vertebrate. It has a backbone.
A grizzly bear is a mammal. Like other mammals, a grizzly bear is a vertebrate. It has a backbone.
A julia butterfly is an insect. Like other insects, a julia butterfly is an invertebrate. It does not have a backbone. It has an exoskeleton.
|
Relavent Documents:
Document 0:::
Animals are multicellular, eukaryotic organisms in the biological kingdom Animalia. With few exceptions, animals consume organic material, breathe oxygen, have myocytes and are able to move, can reproduce sexually, and grow from a hollow sphere of cells, the blastula, during embryonic development. As of 2022, 2.16 million living animal species have been described—of which around 1.05 million are insects, over 85,000 are molluscs, and around 65,000 are vertebrates. It has been estimated there are around 7.77 million animal species. Animals range in length from to . They have complex interactions with each other and their environments, forming intricate food webs. The scientific study of animals is known as zoology.
Most living animal species are in Bilateria, a clade whose members have a bilaterally symmetric body plan. The Bilateria include the protostomes, containing animals such as nematodes, arthropods, flatworms, annelids and molluscs, and the deuterostomes, containing the echinoderms and the chordates, the latter including the vertebrates. Life forms interpreted as early animals were present in the Ediacaran biota of the late Precambrian. Many modern animal phyla became clearly established in the fossil record as marine species during the Cambrian explosion, which began around 539 million years ago. 6,331 groups of genes common to all living animals have been identified; these may have arisen from a single common ancestor that lived 650 million years ago.
Historically, Aristotle divided animals into those with blood and those without. Carl Linnaeus created the first hierarchical biological classification for animals in 1758 with his Systema Naturae, which Jean-Baptiste Lamarck expanded into 14 phyla by 1809. In 1874, Ernst Haeckel divided the animal kingdom into the multicellular Metazoa (now synonymous with Animalia) and the Protozoa, single-celled organisms no longer considered animals. In modern times, the biological classification of animals relies on ad
Document 1:::
International Society for Invertebrate Morphology (ISIM) was founded during the 1st International Congress on Invertebrate Morphology, in Copenhagen, August 2008. The objectives of the society are to promote international collaboration and provide educational opportunities and training on invertebrate morphology, and to organize and promote the international congresses of invertebrate morphology, international meetings and other forms of scientific exchange.
The ISIM has its own Constitution
ISIM board 2014-2017
Gerhard Scholtz (President) Institute of Biology, Humboldt-Universität zu Berlin, Germany. https://www.biologie.hu-berlin.de/de/gruppenseiten/compzool/people/gerhard_scholtz_page
Natalia Biserova (President-Elect) Moscow State University, Moscow, Russia.
Gonzalo Giribet (Past-President) Museum of Comparative Zoology, Harvard University, Cambridge, MA, USA.
Julia Sigwart (Secretary)
Katrina Worsaae (Treasurer)
Greg Edgecombe (2nd term)
Andreas Hejnol (2nd term)
Sally Leys (2nd term)
Fernando Pardos (2nd term)
Katharina Jörger (1st term)
Marymegan Daly (1st term)
Georg Mayer (1st term)
ISIM board 2017-2020
Natalia Biserova (President), Lomonosov Moscow State University, Moscow, Russian Federation http://invert.bio.msu.ru/en/staff-en/33-biserova-en .
Andreas Wanninger (President-elect), Department of Integrative Zoology, University of Vienna, Vienna, Austria.
Gerhard Scholtz (Past-president), Department of Biology, Humboldt-Universität zu Berlin, Germany.
Julia Sigwart (Secretary), School of Biological Sciences, Queen's University Belfast, UK.
Katrine Worsaae (Treasurer), Department of Biology, University of Copenhagen, Copenhagen, Denmark.
Advisory Council:
Ariel Chipman (Israel)
D. Bruce Conn (USA)
Conrad Helm (Germany)
Xiaoya Ma (UK)
Pedro Martinez (Spain)
Ana Riesgo (Spain)
Nadezhda Rimskaya-Korsakova (Russia)
Elected 23-08-2017, Moscow
Former meetings
ICIM 1 (2008) University of Copenhagen, Denmark
ICIM 2 (2011) H
Document 2:::
This is a list of scientific journals which cover the field of zoology.
A
Acta Entomologica Musei Nationalis Pragae
Acta Zoologica Academiae Scientiarum Hungaricae
Acta Zoologica Bulgarica
Acta Zoológica Mexicana
Acta Zoologica: Morphology and Evolution
African Entomology
African Invertebrates
African Journal of Herpetology
African Zoology
Alces
American Journal of Primatology
Animal Biology, formerly Netherlands Journal of Zoology
Animal Cognition
Arctic
Australian Journal of Zoology
Australian Mammalogy
B
Bulgarian Journal of Agricultural Science
Bulletin of the American Museum of Natural History
C
Canadian Journal of Zoology
Caribbean Herpetology
Central European Journal of Biology
Contributions to Zoology
Copeia
Crustaceana
E
Environmental Biology of Fishes
F
Frontiers in Zoology
H
Herpetological Monographs
I
Integrative and Comparative Biology, formerly American Zoologist
International Journal of Acarology
International Journal of Primatology
J
M
Malacologia
N
North-Western Journal of Zoology
P
Physiological and Biochemical Zoology
R
Raffles Bulletin of Zoology
Rangifer
Russian Journal of Nematology
V
The Veliger
W
Worm Runner's Digest
Z
See also
List of biology journals
List of ornithology journals
List of entomology journals
Lists of academic journals
Zoology-related lists
Document 3:::
Invertebrate zoology is the subdiscipline of zoology that consists of the study of invertebrates, animals without a backbone (a structure which is found only in fish, amphibians, reptiles, birds and mammals).
Invertebrates are a vast and very diverse group of animals that includes sponges, echinoderms, tunicates, numerous different phyla of worms, molluscs, arthropods and many additional phyla. Single-celled organisms or protists are usually not included within the same group as invertebrates.
Subdivisions
Invertebrates represent 97% of all named animal species, and because of that fact, this subdivision of zoology has many further
subdivisions, including but not limited to:
Arthropodology - the study of arthropods, which includes
Arachnology - the study of spiders and other arachnids
Entomology - the study of insects
Carcinology - the study of crustaceans
Myriapodology - the study of centipedes, millipedes, and other myriapods
Cnidariology - the study of Cnidaria
Helminthology - the study of parasitic worms.
Malacology - the study of mollusks, which includes
Conchology - the study of Mollusk shells.
Limacology - the study of slugs.
Teuthology - the study of cephalopods.
Invertebrate paleontology - the study of fossil invertebrates
These divisions are sometimes further divided into more specific specialties. For example, within arachnology, acarology is the study of mites and ticks; within entomology, lepidoptery is the study of butterflies and moths, myrmecology is the study of ants and so on. Marine invertebrates are all those invertebrates that exist in marine habitats.
History
Early Modern Era
In the early modern period starting in the late 16th century, invertebrate zoology saw growth in the number of publications made and improvement in the experimental practices associated with the field. (Insects are one of the most diverse groups of organisms on Earth. They play important roles in ecosystems, including pollination, natural enemies, saprophytes, and
Document 4:::
Roshd Biological Education is a quarterly science educational magazine covering recent developments in biology and biology education for a biology teacher Persian -speaking audience. Founded in 1985, it is published by The Teaching Aids Publication Bureau, Organization for Educational Planning and Research, Ministry of Education, Iran. Roshd Biological Education has an editorial board composed of Iranian biologists, experts in biology education, science journalists and biology teachers.
It is read by both biology teachers and students, as a way of launching innovations and new trends in biology education, and helping biology teachers to teach biology in better and more effective ways.
Magazine layout
As of Autumn 2012, the magazine is laid out as follows:
Editorial—often offering a view of point from editor in chief on an educational and/or biological topics.
Explore— New research methods and results on biology and/or education.
World— Reports and explores on biological education worldwide.
In Brief—Summaries of research news and discoveries.
Trends—showing how new technology is altering the way we live our lives.
Point of View—Offering personal commentaries on contemporary topics.
Essay or Interview—often with a pioneer of a biological and/or educational researcher or an influential scientific educational leader.
Muslim Biologists—Short histories of Muslim Biologists.
Environment—An article on Iranian environment and its problems.
News and Reports—Offering short news and reports events on biology education.
In Brief—Short articles explaining interesting facts.
Questions and Answers—Questions about biology concepts and their answers.
Book and periodical Reviews—About new publication on biology and/or education.
Reactions—Letter to the editors.
Editorial staff
Mohammad Karamudini, editor in chief
History
Roshd Biological Education started in 1985 together with many other magazines in other science and art. The first editor was Dr. Nouri-Dalooi, th
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Select the invertebrate.
A. lionfish
B. penguin
C. julia butterfly
D. grizzly bear
Answer:
|
scienceQA-12584
|
multiple_choice
|
Select the amphibian.
|
[
"black howler",
"red-headed poison frog",
"salmon",
"thresher shark"
] |
B
|
A red-headed poison frog is an amphibian. It has moist skin and begins its life in water.
Poison dart frogs come in many bright colors. Their bright color warns other animals that these frogs are poisonous.
A black howler is a mammal. It has hair and feeds its young milk.
Howler monkeys have loud calls, or howls. Their calls can be heard over three miles away!
A thresher shark is a fish. It lives underwater. It has fins, not limbs.
A thresher shark has a long tail. It can use its tail to hit and stun prey.
A salmon is a fish. It lives underwater. It has fins, not limbs.
Unlike most other fish, salmon can live in both fresh water and salt water.
|
Relavent Documents:
Document 0:::
AmphibiaWeb is an American non-profit website that provides information about amphibians. It is run by a group of universities working with the California Academy of Sciences: San Francisco State University, the University of California at Berkeley, University of Florida at Gainesville, and University of Texas at Austin.
AmphibiaWeb's goal is to provide a single page for every species of amphibian in the world so research scientists, citizen scientists and conservationists can collaborate. It added its 7000th animal in 2012, a glass frog from Peru. As of 2022, it hosted more than 8,400 species located worldwide.
Beginning
Scientist David Wake founded AmphibiaWeb in 2000. Wake had been inspired by the decline of amphibian populations across the world. He founded it at the Digital Library Project at the University of California at Berkeley in 2000. Wake came to consider AmphibiaWeb part of his legacy.
Uses
AmphibiaWeb provides information to the IUCN, CalPhotos, Encyclopedia of Life and iNaturalist, and the database is cited in scientific publications.
Document 1:::
Roshd Biological Education is a quarterly science educational magazine covering recent developments in biology and biology education for a biology teacher Persian -speaking audience. Founded in 1985, it is published by The Teaching Aids Publication Bureau, Organization for Educational Planning and Research, Ministry of Education, Iran. Roshd Biological Education has an editorial board composed of Iranian biologists, experts in biology education, science journalists and biology teachers.
It is read by both biology teachers and students, as a way of launching innovations and new trends in biology education, and helping biology teachers to teach biology in better and more effective ways.
Magazine layout
As of Autumn 2012, the magazine is laid out as follows:
Editorial—often offering a view of point from editor in chief on an educational and/or biological topics.
Explore— New research methods and results on biology and/or education.
World— Reports and explores on biological education worldwide.
In Brief—Summaries of research news and discoveries.
Trends—showing how new technology is altering the way we live our lives.
Point of View—Offering personal commentaries on contemporary topics.
Essay or Interview—often with a pioneer of a biological and/or educational researcher or an influential scientific educational leader.
Muslim Biologists—Short histories of Muslim Biologists.
Environment—An article on Iranian environment and its problems.
News and Reports—Offering short news and reports events on biology education.
In Brief—Short articles explaining interesting facts.
Questions and Answers—Questions about biology concepts and their answers.
Book and periodical Reviews—About new publication on biology and/or education.
Reactions—Letter to the editors.
Editorial staff
Mohammad Karamudini, editor in chief
History
Roshd Biological Education started in 1985 together with many other magazines in other science and art. The first editor was Dr. Nouri-Dalooi, th
Document 2:::
Batrachology is the branch of zoology concerned with the study of amphibians including frogs and toads, salamanders, newts, and caecilians. It is a sub-discipline of herpetology, which also includes non-avian reptiles (snakes, lizards, amphisbaenids, turtles, terrapins, tortoises, crocodilians, and the tuatara). Batrachologists may study the evolution, ecology, ethology, or anatomy of amphibians.
Amphibians are cold blooded vertebrates largely found in damp habitats although many species have special behavioural adaptations that allow them to live in deserts, trees, underground and in regions with wide seasonal variations in temperature. There are over 7250 species of amphibians.
Notable batrachologists
Jean Marius René Guibé
Gabriel Bibron
Oskar Boettger
George Albert Boulenger
Edward Drinker Cope
François Marie Daudin
Franz Werner
Leszek Berger
Document 3:::
Toughie was the last known living Rabbs' fringe-limbed treefrog. The species, scientifically known as Ecnomiohyla rabborum, is thought to be extinct, as the last specimen—Toughie—died in captivity on September 26, 2016.
Captivity
Toughie was captured as an adult in Panama in 2005, when researchers went on a conservation mission to rescue species from Batrachochytrium dendrobatidis, a fungus deadly to amphibians. Toughie was one of "several dozen" frogs and tadpoles of the same species to be transported back to the United States.
Toughie lived at the Atlanta Botanical Garden in Georgia. At the Garden, he was placed in a special containment area called the "frogPOD", a biosecure enclosure. Visitors to the Garden are not allowed to visit the frogPOD, as it is used to house critically endangered animals. While in captivity at the Garden, Toughie sired tadpoles with a female, but none survived. After the female died, the only other known specimen in the world was a male, leaving Toughie no other options of reproducing. The other male, who lived at the Zoo Atlanta, was euthanized on February 17, 2012, due to health concerns.
Since Toughie was brought in as an adult to the Garden, they do not know his age but estimated that he was at least 12 years old. On December 15, 2014, Toughie was recorded vocalizing again. It was his first known call since being collected as an adult in 2005.
Toughie died on September 26, 2016, at the Garden.
Personal characteristics
Toughie was given his name by Mark Mandica's son Anthony. Mark Mandica was Toughie's caretaker for many years at the Atlanta Botanical Garden.
Toughie did not like to be handled. He would pinch a handler's hand in an attempt to "say 'let me go'", according to handler Leslie Phillips. She continued with, "For me it is incredibly motivating working with the Rabbs' frog. Having him here is a constant reminder of what can potentially happen to other species if we don't continue the conservation work that we do here a
Document 4:::
Molluscs
Ancylus fluviatilis
Aylacostoma species
Lymnaea ovata
Amphibians
Neurergus strauchii, a newt from Turkey
Pach
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Select the amphibian.
A. black howler
B. red-headed poison frog
C. salmon
D. thresher shark
Answer:
|
sciq-1202
|
multiple_choice
|
Thrombocytes are more commonly known by what name?
|
[
"molecules",
"platelets",
"droplets",
"chromosomes"
] |
B
|
Relavent Documents:
Document 0:::
Platelets or thrombocytes (from Greek θρόμβος, "clot" and κύτος, "cell") are a component of blood whose function (along with the coagulation factors) is to react to bleeding from blood vessel injury by clumping, thereby initiating a blood clot. Platelets have no cell nucleus; they are fragments of cytoplasm derived from the megakaryocytes of the bone marrow or lung, which then enter the circulation. Platelets are found only in mammals, whereas in other vertebrates (e.g. birds, amphibians), thrombocytes circulate as intact mononuclear cells.
One major function of platelets is to contribute to hemostasis: the process of stopping bleeding at the site of interrupted endothelium. They gather at the site and, unless the interruption is physically too large, they plug the hole. First, platelets attach to substances outside the interrupted endothelium: adhesion. Second, they change shape, turn on receptors and secrete chemical messengers: activation. Third, they connect to each other through receptor bridges: aggregation. Formation of this platelet plug (primary hemostasis) is associated with activation of the coagulation cascade, with resultant fibrin deposition and linking (secondary hemostasis). These processes may overlap: the spectrum is from a predominantly platelet plug, or "white clot" to a predominantly fibrin, or "red clot" or the more typical mixture. Some would add the subsequent retraction and platelet inhibition as fourth and fifth steps to the completion of the process and still others would add a sixth step, wound repair. Platelets also participate in both innate and adaptive intravascular immune responses.
Structure
Structure
Structurally the platelet can be divided into four zones, from peripheral to innermost:
Peripheral zone – is rich in glycoproteins required for platelet adhesion, activation and aggregation. For example, GPIb/IX/V; GPVI; GPIIb/IIIa.
Sol-gel zone – is rich in microtubules and microfilaments, allowing the platelets to maintain their
Document 1:::
Lymph node stromal cells are essential to the structure and function of the lymph node whose functions include: creating an internal tissue scaffold for the support of hematopoietic cells; the release of small molecule chemical messengers that facilitate interactions between hematopoietic cells; the facilitation of the migration of hematopoietic cells; the presentation of antigens to immune cells at the initiation of the adaptive immune system; and the homeostasis of lymphocyte numbers. Stromal cells originate from multipotent mesenchymal stem cells.
Structure
Lymph nodes are enclosed in an external fibrous capsule, from which thin walls of sinew called trabeculae penetrate into the lymph node, partially dividing it. Beneath the external capsule and along the courses of the trabeculae, are peritrabecular and subcapsular sinuses. These sinuses are cavities containing macrophages (specialised cells which help to keep the extracellular matrix in order).
The interior of the lymph node has two regions: the cortex and the medulla. In the cortex, lymphoid tissue is organized into nodules. In the nodules, T lymphocytes are located in the T cell zone. B lymphocytes are located in the B cell follicle. The primary B cell follicle matures in germinal centers. In the medulla are hematopoietic cells (which contribute to the formation of the blood) and stromal cells.
Near the medulla is the hilum of lymph node. This is the place where blood vessels enter and leave the lymph node and lymphatic vessels leave the lymph node. Lymph vessels entering the node do so along the perimeter (outer surface).
Function
The lymph nodes, the spleen and Peyer's patches, together are known as secondary lymphoid organs. Lymph nodes are found between lymphatic ducts and blood vessels. Afferent lymphatic vessels bring lymph fluid from the peripheral tissues to the lymph nodes. The lymph tissue in the lymph nodes consists of immune cells (95%), for example lymphocytes, and stromal cells (1% to
Document 2:::
MicrobeLibrary is a permanent collection of over 1400 original peer-reviewed resources for teaching undergraduate microbiology. It is provided by the American Society for Microbiology, Washington DC, United States.
Contents include curriculum activities; images and animations; reviews of books, websites and other resources; and articles from Focus on Microbiology Education, Microbiology Education and Microbe. Around 40% of the materials are free to educators and students, the remainder require a subscription. the service is suspended with the message to:
"Please check back with us in 2017".
External links
MicrobeLibrary
Microbiology
Document 3:::
This is a list of topics in molecular biology. See also index of biochemistry articles.
Document 4:::
The following outline is provided as an overview of and topical guide to biophysics:
Biophysics – interdisciplinary science that uses the methods of physics to study biological systems.
Nature of biophysics
Biophysics is
An academic discipline – branch of knowledge that is taught and researched at the college or university level. Disciplines are defined (in part), and recognized by the academic journals in which research is published, and the learned societies and academic departments or faculties to which their practitioners belong.
A scientific field (a branch of science) – widely recognized category of specialized expertise within science, and typically embodies its own terminology and nomenclature. Such a field will usually be represented by one or more scientific journals, where peer-reviewed research is published.
A natural science – one that seeks to elucidate the rules that govern the natural world using empirical and scientific methods.
A biological science – concerned with the study of living organisms, including their structure, function, growth, evolution, distribution, and taxonomy.
A branch of physics – concerned with the study of matter and its motion through space and time, along with related concepts such as energy and force.
An interdisciplinary field – field of science that overlaps with other sciences
Scope of biophysics research
Biomolecular scale
Biomolecule
Biomolecular structure
Organismal scale
Animal locomotion
Biomechanics
Biomineralization
Motility
Environmental scale
Biophysical environment
Biophysics research overlaps with
Agrophysics
Biochemistry
Biophysical chemistry
Bioengineering
Biogeophysics
Nanotechnology
Systems biology
Branches of biophysics
Astrobiophysics – field of intersection between astrophysics and biophysics concerned with the influence of the astrophysical phenomena upon life on planet Earth or some other planet in general.
Medical biophysics – interdisciplinary field that applies me
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Thrombocytes are more commonly known by what name?
A. molecules
B. platelets
C. droplets
D. chromosomes
Answer:
|
|
sciq-9414
|
multiple_choice
|
What is the term for a close relationship between two species in which at least one species benefits.
|
[
"extosis",
"candiasis",
"symbiosis",
"endosymbiosis"
] |
C
|
Relavent Documents:
Document 0:::
Mutualism describes the ecological interaction between two or more species where each species has a net benefit. Mutualism is a common type of ecological interaction. Prominent examples include most vascular plants engaged in mutualistic interactions with mycorrhizae, flowering plants being pollinated by animals, vascular plants being dispersed by animals, and corals with zooxanthellae, among many others. Mutualism can be contrasted with interspecific competition, in which each species experiences reduced fitness, and exploitation, or parasitism, in which one species benefits at the expense of the other.
The term mutualism was introduced by Pierre-Joseph van Beneden in his 1876 book Animal Parasites and Messmates to mean "mutual aid among species".
Mutualism is often conflated with two other types of ecological phenomena: cooperation and symbiosis. Cooperation most commonly refers to increases in fitness through within-species (intraspecific) interactions, although it has been used (especially in the past) to refer to mutualistic interactions, and it is sometimes used to refer to mutualistic interactions that are not obligate. Symbiosis involves two species living in close physical contact over a long period of their existence and may be mutualistic, parasitic, or commensal, so symbiotic relationships are not always mutualistic, and mutualistic interactions are not always symbiotic. Despite a different definition between mutualistic interactions and symbiosis, mutualistic and symbiosis have been largely used interchangeably in the past, and confusion on their use has persisted.
Mutualism plays a key part in ecology and evolution. For example, mutualistic interactions are vital for terrestrial ecosystem function as about 80% of land plants species rely on mycorrhizal relationships with fungi to provide them with inorganic compounds and trace elements. As another example, the estimate of tropical rainforest plants with seed dispersal mutualisms with animals ranges
Document 1:::
In ecology, a biological interaction is the effect that a pair of organisms living together in a community have on each other. They can be either of the same species (intraspecific interactions), or of different species (interspecific interactions). These effects may be short-term, or long-term, both often strongly influence the adaptation and evolution of the species involved. Biological interactions range from mutualism, beneficial to both partners, to competition, harmful to both partners. Interactions can be direct when physical contact is established or indirect, through intermediaries such as shared resources, territories, ecological services, metabolic waste, toxins or growth inhibitors. This type of relationship can be shown by net effect based on individual effects on both organisms arising out of relationship.
Several recent studies have suggested non-trophic species interactions such as habitat modification and mutualisms can be important determinants of food web structures. However, it remains unclear whether these findings generalize across ecosystems, and whether non-trophic interactions affect food webs randomly, or affect specific trophic levels or functional groups.
History
Although biological interactions, more or less individually, were studied earlier, Edward Haskell (1949) gave an integrative approach to the thematic, proposing a classification of "co-actions", later adopted by biologists as "interactions". Close and long-term interactions are described as symbiosis; symbioses that are mutually beneficial are called mutualistic.
The term symbiosis was subject to a century-long debate about whether it should specifically denote mutualism, as in lichens or in parasites that benefit themselves. This debate created two different classifications for biotic interactions, one based on the time (long-term and short-term interactions), and other based on the magnitud of interaction force (competition/mutualism) or effect of individual fitness, accordi
Document 2:::
Conservation is the maintenance of biological diversity. Conservation can focus on preserving diversity at genetic, species, community or whole ecosystem levels. This article will examine conservation at the species level, because mutualisms involve interactions between species. The ultimate goal of conservation at this level is to prevent the extinction of species. However, species conservation has the broader aim of maintaining the abundance and distribution of all species, not only those threatened with extinction (van Dyke 2008). Determining the value of conserving particular species can be done through the use of evolutionary significant units, which essentially attempt to prioritise the conservation of the species which are rarest, fastest declining, and most distinct genotypically and phenotypically (Moritz 1994, Fraser and Bernatchez 2001).
Mutualisms can be defined as "interspecific interactions in which each of two partner species receives a net benefit" (Bronstein et al. 2004). Here net benefit is defined as, a short-term increase in inclusive fitness (IF). Incorporating the concept of genetic relatedness (through IF) is essential because many mutualisms involve the eusocial insects, where the majority of individuals are not reproductively active. The short-term component is chosen because it is operationally useful, even though the role of long-term adaptation is not considered (de Mazancourt et al. 2005). This definition of mutualism should be suffice for this article, although it neglects discussion of the many subtitles of IF theory applied to mutualisms, and the difficulties of examining short-term compared to long-term benefits, which are discussed in Foster and Wenselneers (2006) and de Mazancourt et al. (2005) respectively. Mutualisms can be broadly divided into two categories. Firstly, obligate mutualism, where two mutualistic partners are completely interdependent for survival and reproduction. Secondly, facultative mutualism, where two mutuali
Document 3:::
Ecological facilitation or probiosis describes species interactions that benefit at least one of the participants and cause harm to neither. Facilitations can be categorized as mutualisms, in which both species benefit, or commensalisms, in which one species benefits and the other is unaffected. This article addresses both the mechanisms of facilitation and the increasing information available concerning the impacts of facilitation on community ecology.
Categories
There are two basic categories of facilitative interactions:
Mutualism is an interaction between species that is beneficial to both. A familiar example of a mutualism is the relationship between flowering plants and their pollinators. The plant benefits from the spread of pollen between flowers, while the pollinator receives some form of nourishment, either from nectar or the pollen itself.
Commensalism is an interaction in which one species benefits and the other species is unaffected. Epiphytes (plants growing on other plants, usually trees) have a commensal relationship with their host plant because the epiphyte benefits in some way (e.g., by escaping competition with terrestrial plants or by gaining greater access to sunlight) while the host plant is apparently unaffected.
Strict categorization, however, is not possible for some complex species interactions. For example, seed germination and survival in harsh environments is often higher under so-called nurse plants than on open ground. A nurse plant is one with an established canopy, beneath which germination and survival are more likely due to increased shade, soil moisture, and nutrients. Thus, the relationship between seedlings and their nurse plants is commensal. However, as the seedlings grow into established plants, they are likely to compete with their former benefactors for resources.
Mechanisms
The beneficial effects of species on one another are realized in various ways, including refuge from physical stress, predation, and competi
Document 4:::
The stress gradient hypothesis (SGH) is an evolutionary theory in microbial ecology and community ecology that provides a framework to predict when positive or negative interactions should be observed in an habitat. The SGH states that facilitation, cooperation or mutualism should be more common in stressful environments, compared with benign environments (i.e nutrient excess) where competition or parasitism should be more common.
The stress gradient hypothesis, in which ecological interactions shift in a positive direction with increasing environmental stress, is controversial among ecologists, in part because of contradictory support, yet a 2021 meta analysis study compared SGH across different organisms with intraspecificity and interspecificity interacrions and conclude that the SGH is indeed a broadly relevant ecological phenomena that is currently held back by cross-disciplinary communication barriers.
SGH is well supported by studies that feature bacteria, plants, terrestrial ecosystems, interspecific negative interactions, adults, survival instead of growth or reproduction, and drought, fire, and nutrient stress.
Drought and nutrient stress, especially when combined, shift ecological interactions positively
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is the term for a close relationship between two species in which at least one species benefits.
A. extosis
B. candiasis
C. symbiosis
D. endosymbiosis
Answer:
|
|
sciq-9156
|
multiple_choice
|
What type of logic can lead to important conclusions from the collecting and analyzing of observations?
|
[
"gravitational reasoning",
"inductive reasoning",
"primitive reasoning",
"conductive reasoning"
] |
B
|
Relavent Documents:
Document 0:::
C
Canonical form --
Canonical form (Boolean algebra) --
Cartesian circle --
Case-based reasoning --
Categorical logic --
Categories (Aristotle) --
Categories (Peirce) --
Category mistake --
Catuṣkoṭi --
Circular definition --
Circular reasoning --
Circular reference --
Circular reporting --
Circumscription (logic) --
Circumscription (taxonomy) --
Classical logic --
Clocked logic --
Cognitive bias --
Cointerpretability --
Colorless green ideas sleep fu
Document 1:::
In mathematical psychology and education theory, a knowledge space is a combinatorial structure used to formulate mathematical models describing the progression of a human learner. Knowledge spaces were introduced in 1985 by Jean-Paul Doignon and Jean-Claude Falmagne, and remain in extensive use in the education theory. Modern applications include two computerized tutoring systems, ALEKS and the defunct RATH.
Formally, a knowledge space assumes that a domain of knowledge is a collection of concepts or skills, each of which must be eventually mastered. Not all concepts are interchangeable; some require other concepts as prerequisites. Conversely, competency at one skill may ease the acquisition of another through similarity. A knowledge space marks out which collections of skills are feasible: they can be learned without mastering any other skills. Under reasonable assumptions, the collection of feasible competencies forms the mathematical structure known as an antimatroid.
Researchers and educators usually explore the structure of a discipline's knowledge space as a latent class model.
Motivation
Knowledge Space Theory attempts to address shortcomings of standardized testing when used in educational psychometry. Common tests, such as the SAT and ACT, compress a student's knowledge into a very small range of ordinal ranks, in the process effacing the conceptual dependencies between questions. Consequently, the tests cannot distinguish between true understanding and guesses, nor can they identify a student's particular weaknesses, only the general proportion of skills mastered. The goal of knowledge space theory is to provide a language by which exams can communicate
What the student can do and
What the student is ready to learn.
Model structure
Knowledge Space Theory-based models presume that an educational subject can be modeled as a finite set of concepts, skills, or topics. Each feasible state of knowledge about is then a subset of ; the set of
Document 2:::
Logic is the formal science of using reason and is considered a branch of both philosophy and mathematics and to a lesser extent computer science. Logic investigates and classifies the structure of statements and arguments, both through the study of formal systems of inference and the study of arguments in natural language. The scope of logic can therefore be very large, ranging from core topics such as the study of fallacies and paradoxes, to specialized analyses of reasoning such as probability, correct reasoning, and arguments involving causality. One of the aims of logic is to identify the correct (or valid) and incorrect (or fallacious) inferences. Logicians study the criteria for the evaluation of arguments.
Foundations of logic
Philosophy of logic
Analytic-synthetic distinction
Antinomy
A priori and a posteriori
Definition
Description
Entailment
Identity (philosophy)
Inference
Logical form
Logical implication
Logical truth
Logical consequence
Name
Necessity
Material conditional
Meaning (linguistic)
Meaning (non-linguistic)
Paradox (list)
Possible world
Presupposition
Probability
Quantification
Reason
Reasoning
Reference
Semantics
Strict conditional
Syntax (logic)
Truth
Truth value
Validity
Branches of logic
Affine logic
Alethic logic
Aristotelian logic
Boolean logic
Buddhist logic
Bunched logic
Categorical logic
Classical logic
Computability logic
Deontic logic
Dependence logic
Description logic
Deviant logic
Doxastic logic
Epistemic logic
First-order logic
Formal logic
Free logic
Fuzzy logic
Higher-order logic
Infinitary logic
Informal logic
Intensional logic
Intermediate logic
Interpretability logic
Intuitionistic logic
Linear logic
Many-valued logic
Mathematical logic
Metalogic
Minimal logic
Modal logic
Non-Aristotelian logic
Non-classical logic
Noncommutative logic
Non-monotonic logic
Ordered logic
Paraconsistent logic
Philosophical logic
Predicate logic
Propositional logic
P
Document 3:::
Discovery is the act of detecting something new, or something previously unrecognized as meaningful. Concerning sciences and academic disciplines, discovery is the observation of new phenomena, new actions, or new events and providing new reasoning to explain the knowledge gathered through such observations with previously acquired knowledge from abstract thought and everyday experiences. A discovery may sometimes be based on earlier discoveries, collaborations, or ideas. Some discoveries represent a radical breakthrough in knowledge or technology.
New discoveries are acquired through various senses and are usually assimilated, merging with pre-existing knowledge and actions. Questioning is a major form of human thought and interpersonal communication, and plays a key role in discovery. Discoveries are often made due to questions. Some discoveries lead to the invention of objects, processes, or techniques. A discovery may sometimes be based on earlier discoveries, collaborations or ideas, and the process of discovery requires at least the awareness that an existing concept or method can be modified or transformed. However, some discoveries also represent a radical breakthrough in knowledge.
Science
Within scientific disciplines, discovery is the observation of new phenomena, actions, or events which help explain the knowledge gathered through previously acquired scientific evidence. In science, exploration is one of three purposes of research, the other two being description and explanation. Discovery is made by providing observational evidence and attempts to develop an initial, rough understanding of some phenomenon.
Discovery within the field of particle physics has an accepted definition for what constitutes a discovery: a five-sigma level of certainty. Such a level defines statistically how unlikely it is that an experimental result is due to chance. The combination of a five-sigma level of certainty, and independent confirmation by other experiments, turn f
Document 4:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What type of logic can lead to important conclusions from the collecting and analyzing of observations?
A. gravitational reasoning
B. inductive reasoning
C. primitive reasoning
D. conductive reasoning
Answer:
|
|
sciq-9773
|
multiple_choice
|
Slime molds have several morphological similarities to fungi, including spore-generation, that are thought to be the result of convergent what?
|
[
"emergence",
"generation",
"phenomenon",
"evolution"
] |
D
|
Relavent Documents:
Document 0:::
In biology, cell theory is a scientific theory first formulated in the mid-nineteenth century, that organisms are made up of cells, that they are the basic structural/organizational unit of all organisms, and that all cells come from pre-existing cells. Cells are the basic unit of structure in all organisms and also the basic unit of reproduction.
The theory was once universally accepted, but now some biologists consider non-cellular entities such as viruses living organisms, and thus disagree with the first tenet. As of 2021: "expert opinion remains divided roughly a third each between yes, no and don’t know". As there is no universally accepted definition of life, discussion still continues.
History
With continual improvements made to microscopes over time, magnification technology became advanced enough to discover cells. This discovery is largely attributed to Robert Hooke, and began the scientific study of cells, known as cell biology. When observing a piece of cork under the scope, he was able to see pores. This was shocking at the time as it was believed no one else had seen these. To further support his theory, Matthias Schleiden and Theodor Schwann both also studied cells of both animal and plants. What they discovered were significant differences between the two types of cells. This put forth the idea that cells were not only fundamental to plants, but animals as well.
Microscopes
The discovery of the cell was made possible through the invention of the microscope. In the first century BC, Romans were able to make glass. They discovered that objects appeared to be larger under the glass. The expanded use of lenses in eyeglasses in the 13th century probably led to wider spread use of simple microscopes (magnifying glasses) with limited magnification. Compound microscopes, which combine an objective lens with an eyepiece to view a real image achieving much higher magnification, first appeared in Europe around 1620. In 1665, Robert Hooke used a microscope
Document 1:::
A clamp connection is a hook-like structure formed by growing hyphal cells of certain fungi. It is a characteristic feature of basidiomycete fungi. It is created to ensure that each cell, or segment of hypha separated by septa (cross walls), receives a set of differing nuclei, which are obtained through mating of hyphae of differing sexual types. It is used to maintain genetic variation within the hypha much like the mechanisms found in croziers (hooks) during the sexual reproduction of ascomycetes.
Formation
Clamp connections are formed by the terminal hypha during elongation. Before the clamp connection is formed this terminal segment contains two nuclei. Once the terminal segment is long enough it begins to form the clamp connection. At the same time, each nucleus undergoes mitotic division to produce two daughter nuclei. As the clamp continues to develop it uptakes one of the daughter (green circle) nuclei and separates it from its sister nucleus. While this is occurring the remaining nuclei (orange circles) begin to migrate from one another to opposite ends of the cell. Once all these steps have occurred a septum forms, separating each set of nuclei.
Use in classification
Clamp connections are structures unique to the phylum Basidiomycota. Many fungi from this phylum produce spores in basidiocarps (fruiting bodies, or mushrooms), above ground. Though clamp connections are exclusive to this phylum, not all species of Basidiomycota possess these structures. As such, the presence or absences of clamp connections has been a tool in categorizing genera and species.
Fossil record
A fungal mycelium containing abundant clamp connections was found that dated to the Pennsylvanian era (298.9–323.2 Mya). This fossil, classified in the form genus Palaeancistrus, has hyphae that compare with extant saprophytic basidiomycetes. The oldest known clamp connections exist in hyphae present in the fossil fern Botryopteris antiqua, which predate Palaeancistrus by about 25 Ma.
Document 2:::
Each species of slime mold has its own specific chemical messenger, which are collectively referred to as acrasins. These chemicals signal that many individual cells aggregate to form a single large cell or plasmodium. One of the earliest acrasins to be identified was cyclic AMP, found in the species Dictyostelium discoideum by Brian Shaffer, which exhibits a complex swirling-pulsating spiral pattern when forming a pseudoplasmodium.
The term acrasin was descriptively named after Acrasia from Edmund Spenser's Faerie Queene, who seduced men against their will and then transformed them into beasts. Acrasia is itself a play on the Greek akrasia that describes loss of free will.
Extraction
Brian Shaffer was the first to purify acrasin, now known to be cyclic AMP, in 1954, using methanol. Glorin, the acrasin of P. violaceum, can be purified by inhibiting the acrasin-degrading enzyme acrasinase with alcohol, extracting with alcohol and separating with column chromatography.
Notes
Evidence for the formation of cell aggregates by chemotaxis in the development of the slime mold Dictyostelium discoideum - J.T.Bonner and L.J.Savage Journal of Experimental Biology Vol. 106, pp. 1, October (1947) Cell Biology
Aggregation in cellular slime moulds: in vitro isolation of acrasin - B.M.Shaffer Nature Vol. 79, pp. 975, (1953) Cell Biology
Identification of a pterin as the acrasin of the cellular slime mold Dictyostelium lacteum - Proceedings of the National Academy of Sciences United States Vol. 79, pp. 6270–6274, October (1982) Cell Biology
Hunting Slime Moulds - Adele Conover, Smithsonian Magazine Online (2001)
Document 3:::
The Neighbour-Sensing mathematical model of hyphal growth is a set of interactive computer models that simulate the way fungi hyphae grow in three-dimensional space. The three-dimensional simulation is an experimental tool which can be used to study the morphogenesis of fungal hyphal networks.
The modelling process starts from the proposition that each hypha in the fungal mycelium generates a certain abstract field that (like known physical fields) decreases with increasing distance. Both scalar and vector fields are included in the models. The field(s) and its (their) gradient(s) are used to inform the algorithm that calculates the likelihood of branching, the angle of branching and the growth direction of each hyphal tip in the simulated mycelium. The growth vector is being informed of its surroundings so, effectively, the virtual hyphal tip is sensing the neighbouring mycelium. This is why we call it the Neighbour-Sensing model.
Cross-walls in living hyphae are formed only at right angles to the long axis of the hypha. A daughter hyphal apex can only arise if a branch is initiated. So, for the fungi, hyphal branch formation is the equivalent of cell division in animals, plants and protists. The position of origin of a branch, and its direction and rate of growth are the main formative events in the development of fungal tissues and organs. Consequently, by simulating the mathematics of the control of hyphal growth and branching the Neighbour-Sensing model provides the user with a way of experimenting with features that may regulate hyphal growth patterns during morphogenesis to arrive at suggestions that could be tested with live fungi.
The model was proposed by Audrius Meškauskas and David Moore in 2004 and developed using the supercomputing facilities of the University of Manchester.
The key idea of this model is that all parts of the fungal mycelium have identical field generation systems, field sensing mechanisms and growth direction altering algorithms.
Document 4:::
A mold () or mould () is one of the structures that certain fungi can form. The dust-like, colored appearance of molds is due to the formation of spores containing fungal secondary metabolites. The spores are the dispersal units of the fungi. Not all fungi form molds. Some fungi form mushrooms; others grow as single cells and are called microfungi (for example yeasts).
A large and taxonomically diverse number of fungal species form molds. The growth of hyphae results in discoloration and a fuzzy appearance, especially on food. The network of these tubular branching hyphae, called a mycelium, is considered a single organism. The hyphae are generally transparent, so the mycelium appears like very fine, fluffy white threads over the surface. Cross-walls (septa) may delimit connected compartments along the hyphae, each containing one or multiple, genetically identical nuclei. The dusty texture of many molds is caused by profuse production of asexual spores (conidia) formed by differentiation at the ends of hyphae. The mode of formation and shape of these spores is traditionally used to classify molds. Many of these spores are colored, making the fungus much more obvious to the human eye at this stage in its life-cycle.
Molds are considered to be microbes and do not form a specific taxonomic or phylogenetic grouping, but can be found in the divisions Zygomycota and Ascomycota. In the past, most molds were classified within the Deuteromycota. Mold had been used as a common name for now non-fungal groups such as water molds or slime molds that were once considered fungi.
Molds cause biodegradation of natural materials, which can be unwanted when it becomes food spoilage or damage to property. They also play important roles in biotechnology and food science in the production of various pigments, foods, beverages, antibiotics, pharmaceuticals and enzymes. Some diseases of animals and humans can be caused by certain molds: disease may result from allergic sensitivity to m
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Slime molds have several morphological similarities to fungi, including spore-generation, that are thought to be the result of convergent what?
A. emergence
B. generation
C. phenomenon
D. evolution
Answer:
|
|
sciq-11066
|
multiple_choice
|
About 20 percent of carbon dioxide is bound by hemoglobin and is transported where?
|
[
"lungs",
"brain tissues",
"intestines",
"bones"
] |
A
|
Relavent Documents:
Document 0:::
Myoglobin (symbol Mb or MB) is an iron- and oxygen-binding protein found in the cardiac and skeletal muscle tissue of vertebrates in general and in almost all mammals. Myoglobin is distantly related to hemoglobin. Compared to hemoglobin, myoglobin has a higher affinity for oxygen and does not have cooperative binding with oxygen like hemoglobin does. Myoglobin consists of non-polar amino acids at the core of the globulin, where the heme group is non-covalently bounded with the surrounding polypeptide of myoglobin. In humans, myoglobin is only found in the bloodstream after muscle injury.
High concentrations of myoglobin in muscle cells allow organisms to hold their breath for a longer period of time. Diving mammals such as whales and seals have muscles with particularly high abundance of myoglobin. Myoglobin is found in Type I muscle, Type II A, and Type II B; although many texts consider myoglobin not to be found in smooth muscle, this has proved erroneous: there is also myoglobin in smooth muscle cells.
Myoglobin was the first protein to have its three-dimensional structure revealed by X-ray crystallography. This achievement was reported in 1958 by John Kendrew and associates. For this discovery, Kendrew shared the 1962 Nobel Prize in chemistry with Max Perutz. Despite being one of the most studied proteins in biology, its physiological function is not yet conclusively established: mice genetically engineered to lack myoglobin can be viable and fertile, but show many cellular and physiological adaptations to overcome the loss. Through observing these changes in myoglobin-depleted mice, it is hypothesised that myoglobin function relates to increased oxygen transport to muscle, and to oxygen storage; as well, it serves as a scavenger of reactive oxygen species.
In humans, myoglobin is encoded by the MB gene.
Myoglobin can take the forms oxymyoglobin (MbO2), carboxymyoglobin (MbCO), and metmyoglobin (met-Mb), analogously to hemoglobin taking the forms oxyhemogl
Document 1:::
In acid base physiology, the Davenport diagram is a graphical tool, developed by Horace W. Davenport, that allows a clinician or investigator to describe blood bicarbonate concentrations and blood pH following a respiratory and/or metabolic acid-base disturbance. The diagram depicts a three-dimensional surface describing all possible states of chemical equilibria between gaseous carbon dioxide, aqueous bicarbonate and aqueous protons at the physiologically complex interface of the alveoli of the lungs and the alveolar capillaries. Although the surface represented in the diagram is experimentally determined, the Davenport diagram is rarely used in the clinical setting, but allows the investigator to envision the effects of physiological changes on blood acid-base chemistry. For clinical use there are two recent innovations: an Acid-Base Diagram which provides Text Descriptions for the abnormalities and a High Altitude Version that provides text descriptions appropriate for the altitude.
Derivation
When a sample of blood is exposed to air, either in the alveoli of the lung or in an in vitro laboratory experiment, carbon dioxide in the air rapidly enters into equilibrium with carbon dioxide derivatives and other species in the aqueous solution. Figure 1 illustrates the most important equilibrium reactions of carbon dioxide in blood relating to acid-base physiology:
Note that in this equation, the HB/B- buffer system represents all non-bicarbonate buffers present in the blood, such as hemoglobin in its various protonated and deprotonated states. Because many different non-bicarbonate buffers are present in human blood, the final equilibrium state reached at any given pCO2 is highly complex and cannot be readily predicted using theory alone. By depicting experimental results, the Davenport diagram provides a simple approach to describing the behavior of this complex system.
Figure 2 shows a Davenport diagram as commonly depicted in textbooks and the literature. To un
Document 2:::
2,3-Bisphosphoglyceric acid (conjugate base 2,3-bisphosphoglycerate) (2,3-BPG), also known as 2,3-diphosphoglyceric acid (conjugate base 2,3-diphosphoglycerate) (2,3-DPG), is a three-carbon isomer of the glycolytic intermediate 1,3-bisphosphoglyceric acid (1,3-BPG).
-2,3-BPG is present in human red blood cells (RBC; erythrocyte) at approximately 5 mmol/L. It binds with greater affinity to deoxygenated hemoglobin (e.g., when the red blood cell is near respiring tissue) than it does to oxygenated hemoglobin (e.g., in the lungs) due to conformational differences: 2,3-BPG (with an estimated size of about 9 Å) fits in the deoxygenated hemoglobin conformation (with an 11-Angstrom pocket), but not as well in the oxygenated conformation (5 Angstroms). It interacts with deoxygenated hemoglobin beta subunits and decreases the affinity for oxygen and allosterically promotes the release of the remaining oxygen molecules bound to the hemoglobin. Therefore, it enhances the ability of RBCs to release oxygen near tissues that need it most. 2,3-BPG is thus an allosteric effector.
Its function was discovered in 1967 by Reinhold Benesch and Ruth Benesch.
Metabolism
2,3-BPG is formed from 1,3-BPG by the enzyme BPG mutase. It can then be broken down by 2,3-BPG phosphatase to form 3-phosphoglycerate. Its synthesis and breakdown are, therefore, a way around a step of glycolysis, with the net expense of one ATP per molecule of 2,3-BPG generated as the high-energy carboxylic acid-phosphate mixed anhydride bond is cleaved by bisphosphoglycerate mutase.
Document 3:::
The Haldane effect is a property of hemoglobin first described by John Scott Haldane, within which oxygenation of blood in the lungs displaces carbon dioxide from hemoglobin, increasing the removal of carbon dioxide. Consequently, oxygenated blood has a reduced affinity for carbon dioxide. Thus, the Haldane effect describes the ability of hemoglobin to carry increased amounts of carbon dioxide (CO2) in the deoxygenated state as opposed to the oxygenated state. Vice versa, it is true that a high concentration of CO2 facilitates dissociation of oxyhemoglobin, though this is the result of two distinct processes (Bohr effect and Margaria-Green effect) and should be distinguished from Haldane effect.
Carbaminohemoglobin
Carbon dioxide travels through the blood in three different ways. One of these ways is by binding to amino groups, creating carbamino compounds. Amino groups are available for binding at the N-terminals and at side-chains of arginine and lysine residues in hemoglobin. When carbon dioxide binds to these residues carbaminohemoglobin is formed. This amount of carbaminohemoglobin formed is inversely proportional to the amount of oxygen attached to hemoglobin. Thus, at lower oxygen saturation, more carbaminohemoglobin is formed. These dynamics explain the relative difference in hemoglobin's affinity for carbon dioxide depending on oxygen levels known as the Haldane effect.
Buffering
Histidine residues in hemoglobin can accept protons and act as buffers. Deoxygenated hemoglobin is a better proton acceptor than the oxygenated form.
In red blood cells, the enzyme carbonic anhydrase catalyzes the conversion of dissolved carbon dioxide to carbonic acid, which rapidly dissociates to bicarbonate and a free proton:
CO2 + H2O → H2CO3 → H+ + HCO3−
By Le Chatelier's principle, anything that stabilizes the proton produced will cause the reaction to shift to the right, thus the enhanced affinity of deoxyhemoglobin for protons enhances synthesis of bicarbonate and accord
Document 4:::
Blood is a body fluid in the circulatory system of humans and other vertebrates that delivers necessary substances such as nutrients and oxygen to the cells, and transports metabolic waste products away from those same cells. Blood in the circulatory system is also known as peripheral blood, and the blood cells it carries, peripheral blood cells.
Blood is composed of blood cells suspended in blood plasma. Plasma, which constitutes 55% of blood fluid, is mostly water (92% by volume), and contains proteins, glucose, mineral ions, hormones, carbon dioxide (plasma being the main medium for excretory product transportation), and blood cells themselves. Albumin is the main protein in plasma, and it functions to regulate the colloidal osmotic pressure of blood. The blood cells are mainly red blood cells (also called RBCs or erythrocytes), white blood cells (also called WBCs or leukocytes), and in mammals platelets (also called thrombocytes). The most abundant cells in vertebrate blood are red blood cells. These contain hemoglobin, an iron-containing protein, which facilitates oxygen transport by reversibly binding to this respiratory gas thereby increasing its solubility in blood. In contrast, carbon dioxide is mostly transported extracellularly as bicarbonate ion transported in plasma.
Vertebrate blood is bright red when its hemoglobin is oxygenated and dark red when it is deoxygenated.
Some animals, such as crustaceans and mollusks, use hemocyanin to carry oxygen, instead of hemoglobin. Insects and some mollusks use a fluid called hemolymph instead of blood, the difference being that hemolymph is not contained in a closed circulatory system. In most insects, this "blood" does not contain oxygen-carrying molecules such as hemoglobin because their bodies are small enough for their tracheal system to suffice for supplying oxygen.
Jawed vertebrates have an adaptive immune system, based largely on white blood cells. White blood cells help to resist infections and parasite
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
About 20 percent of carbon dioxide is bound by hemoglobin and is transported where?
A. lungs
B. brain tissues
C. intestines
D. bones
Answer:
|
|
sciq-1316
|
multiple_choice
|
What makes the realized niche of chthamalus much smaller than its fundamental niche?
|
[
"microscopic competition",
"instantaneous competition",
"asexual competition",
"interspecific competition"
] |
D
|
Relavent Documents:
Document 0:::
In ecology, a niche is the match of a species to a specific environmental condition. It describes how an organism or population responds to the distribution of resources and competitors (for example, by growing when resources are abundant, and when predators, parasites and pathogens are scarce) and how it in turn alters those same factors (for example, limiting access to resources by other organisms, acting as a food source for predators and a consumer of prey). "The type and number of variables comprising the dimensions of an environmental niche vary from one species to another [and] the relative importance of particular environmental variables for a species may vary according to the geographic and biotic contexts".
A Grinnellian niche is determined by the habitat in which a species lives and its accompanying behavioral adaptations. An Eltonian niche emphasizes that a species not only grows in and responds to an environment, it may also change the environment and its behavior as it grows. The Hutchinsonian niche uses mathematics and statistics to try to explain how species coexist within a given community.
The concept of ecological niche is central to ecological biogeography, which focuses on spatial patterns of ecological communities. "Species distributions and their dynamics over time result from properties of the species, environmental variation..., and interactions between the two—in particular the abilities of some species, especially our own, to modify their environments and alter the range dynamics of many other species." Alteration of an ecological niche by its inhabitants is the topic of niche construction.
The majority of species exist in a standard ecological niche, sharing behaviors, adaptations, and functional traits similar to the other closely related species within the same broad taxonomic class, but there are exceptions. A premier example of a non-standard niche filling species is the flightless, ground-dwelling kiwi bird of New Zealand, which
Document 1:::
In ecology, the competition–colonization trade-off is a stabilizing mechanism that has been proposed to explain species diversity in some biological systems, especially those that are not in equilibrium. In which case some species are particularly good at colonizing and others have well-established survival abilities.
The concept of the competition-colonization trade-off was originally proposed by Levins and Culver, the model indicated that two species could coexist if one had impeccable competition skill and the other was excellent at colonizing. The model indicates that there is typically a trade-off, in which a species is typically better at either competing or colonizing. A later model, labelled The Lottery Model was also proposed, in which interspecific competition is accounted for within the population.
Mathematical models
Levins and Culver model
Where:
fraction of patches that are occupied by species ;
colonization rate of species ;
mortality rate of species (independent of patch density).
Species 1 = competitor, can colonize in area that is uninhabited or inhabited by species 2 .
Species 2 = colonizer, can only colonize in uninhabited areas .
Species 2 is subject to displacement by its competitor .
If species 2 has a higher colonization rate it can coexist with species 1: .
This model is described as the displacement competition model, it has been observed in marine mollusks and fungi. This model makes two large assumptions:
1. "a propagule of a superior competitor takes over a patch from an adult of the inferior competitor".
2. The adult must be displaced fast enough to ensure that it does not reproduce while it is being displaced.
Lottery model
Colonization rate is now described by interspecific competition.
and . Both f and g > 0.
An increase in p1 is related to a decrease in the colonization rate of species 2.
g < f implies a competitive advantage of species 1 and c2 > c1 implies a colonization advantage for species 2.
In plants
Document 2:::
In ecological theory, the Hutchinson's ratio is the ratio of the size differences between similar species when they are living together as compared to when they are isolated. It is named after G. Evelyn Hutchinson who concluded that various key attributes in species varied according to the ratio of 1:1.1 to 1:1.4. The mean ratio 1.3 can be interpreted as the amount of separation necessary to obtain coexistence of species at the same trophic level.
The variation in trophic structures of sympatric congeneric species is presumed to lead to niche differentiation, and allowing coexistence of multiple similar species in the same habitat by the partitioning of food resources. Hutchinson concluded that this size ratio could be used as an indicator of the kind of difference necessary to permit two species to co-occur in different niches but at the same level of the food web. The rule's legitimacy has been questioned, as other categories of objects also exhibit size ratios of roughly 1.3.
Studies done on interspecific competition and niche changes in Tits (Parus spp.) show that when there are multiple species in the same community there is an expected change in foraging when they are of similar size (size ratio 1-1.2). There was no change found among the less similar species. In this paper this was strong evidence for niche differentiation for interspecific competition, and would also be a good argument for Hutchinson's rule.
The simplest and perhaps the most effective way to differentiate the ecological niches of coexisting species is their morphological differentiation (in particular, size differentiation).
Hutchinson showed that the average body size ratio in species of the same genus that belong to the same community and use the same resource is about 1.3 (from 1.1 to 1.4) and the respective body weight ratio is 2. This empirical pattern tells us that this rule does not apply to all organisms and ecological situations. And, therefore, it would be of particular
Document 3:::
This glossary of biology terms is a list of definitions of fundamental terms and concepts used in biology, the study of life and of living organisms. It is intended as introductory material for novices; for more specific and technical definitions from sub-disciplines and related fields, see Glossary of cell biology, Glossary of genetics, Glossary of evolutionary biology, Glossary of ecology, Glossary of environmental science and Glossary of scientific naming, or any of the organism-specific glossaries in :Category:Glossaries of biology.
A
B
C
D
E
F
G
H
I
J
K
L
M
N
O
P
R
S
T
U
V
W
X
Y
Z
Related to this search
Index of biology articles
Outline of biology
Glossaries of sub-disciplines and related fields:
Glossary of botany
Glossary of ecology
Glossary of entomology
Glossary of environmental science
Glossary of genetics
Glossary of ichthyology
Glossary of ornithology
Glossary of scientific naming
Glossary of speciation
Glossary of virology
Document 4:::
The Eltonian niche is an ecological niche that emphasizes the functional attributes of animals and their corresponding trophic position. This was the definition Eugene Odum popularized in his analogy of the niche of a species with its profession in the ecosystem as opposed to the habitat being its address. The definition is attributed to Charles Elton in his 1927 now classic book Animal Ecology. Elton used the two African Rhinoceros species to exemplify the definition. The White Rhinoceros has broad (wide, hence its name) mouthparts, which are efficient in harvesting grass, while the Black Rhinoceros has narrow pointed lips enabling it to feed selectively on the foliage of thorny bushes.
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What makes the realized niche of chthamalus much smaller than its fundamental niche?
A. microscopic competition
B. instantaneous competition
C. asexual competition
D. interspecific competition
Answer:
|
|
ai2_arc-107
|
multiple_choice
|
Geologists frequently use mass spectrometers in determining the relative age of rocks. A spectrometer can differentiate between the ratios of isotopes in rock samples and calculates the radioactive decay involving what two isotopes?
|
[
"Uranium - Lead",
"Rubidium - Strontium",
"Potassium - Argon",
"Uranium - Strontium"
] |
A
|
Relavent Documents:
Document 0:::
Isotopic reference materials are compounds (solids, liquids, gasses) with well-defined isotopic compositions and are the ultimate sources of accuracy in mass spectrometric measurements of isotope ratios. Isotopic references are used because mass spectrometers are highly fractionating. As a result, the isotopic ratio that the instrument measures can be very different from that in the sample's measurement. Moreover, the degree of instrument fractionation changes during measurement, often on a timescale shorter than the measurement's duration, and can depend on the characteristics of the sample itself. By measuring a material of known isotopic composition, fractionation within the mass spectrometer can be removed during post-measurement data processing. Without isotope references, measurements by mass spectrometry would be much less accurate and could not be used in comparisons across different analytical facilities. Due to their critical role in measuring isotope ratios, and in part, due to historical legacy, isotopic reference materials define the scales on which isotope ratios are reported in the peer-reviewed scientific literature.
Isotope reference materials are generated, maintained, and sold by the International Atomic Energy Agency (IAEA), the National Institute of Standards and Technology (NIST), the United States Geologic Survey (USGS), the Institute for Reference Materials and Measurements (IRMM), and a variety of universities and scientific supply companies. Each of the major stable isotope systems (hydrogen, carbon, oxygen, nitrogen, and sulfur) has a wide variety of references encompassing distinct molecular structures. For example, nitrogen isotope reference materials include N-bearing molecules such ammonia (NH3), atmospheric dinitrogen (N2), and nitrate (NO3−). Isotopic abundances are commonly reported using the δ notation, which is the ratio of two isotopes (R) in a sample relative to the same ratio in a reference material, often reported in per mill
Document 1:::
Isotope-ratio mass spectrometry (IRMS) is a specialization of mass spectrometry, in which mass spectrometric methods are used to measure the relative abundance of isotopes in a given sample.
This technique has two different applications in the earth and environmental sciences. The analysis of 'stable isotopes' is normally concerned with measuring isotopic variations arising from mass-dependent isotopic fractionation in natural systems. On the other hand, radiogenic isotope analysis involves measuring the abundances of decay-products of natural radioactivity, and is used in most long-lived radiometric dating methods.
Introduction
The isotope-ratio mass spectrometer (IRMS) allows the precise measurement of mixtures of naturally occurring isotopes. Most instruments used for precise determination of isotope ratios are of the magnetic sector type. This type of analyzer is superior to the quadrupole type in this field of research for two reasons. First, it can be set up for multiple-collector analysis, and second, it gives high-quality 'peak shapes'. Both of these considerations are important for isotope-ratio analysis at very high precision and accuracy.
The sector-type instrument designed by Alfred Nier was such an advance in mass spectrometer design that this type of instrument is often called the 'Nier type'. In the most general terms the instrument operates by ionizing the sample of interest, accelerating it over a potential in the kilo-volt range, and separating the resulting stream of ions according to their mass-to-charge ratio (m/z). Beams with lighter ions bend at a smaller radius than beams with heavier ions. The current of each ion beam is then measured using a 'Faraday cup' or multiplier detector.
Many radiogenic isotope measurements are made by ionization of a solid source, whereas stable isotope measurements of light elements (e.g. H, C, O) are usually made in an instrument with a gas source. In a "multicollector" instrument, the ion collector typica
Document 2:::
Isotope geochemistry is an aspect of geology based upon the study of natural variations in the relative abundances of isotopes of various elements. Variations in isotopic abundance are measured by isotope ratio mass spectrometry, and can reveal information about the ages and origins of rock, air or water bodies, or processes of mixing between them.
Stable isotope geochemistry is largely concerned with isotopic variations arising from mass-dependent isotope fractionation, whereas radiogenic isotope geochemistry is concerned with the products of natural radioactivity.
Stable isotope geochemistry
For most stable isotopes, the magnitude of fractionation from kinetic and equilibrium fractionation is very small; for this reason, enrichments are typically reported in "per mil" (‰, parts per thousand). These enrichments (δ) represent the ratio of heavy isotope to light isotope in the sample over the ratio of a standard. That is,
‰
Hydrogen
Carbon
Carbon has two stable isotopes, 12C and 13C, and one radioactive isotope, 14C.
The stable carbon isotope ratio, δ13C, is measured against Vienna Pee Dee Belemnite (VPDB). The stable carbon isotopes are fractionated primarily by photosynthesis (Faure, 2004). The 13C/12C ratio is also an indicator of paleoclimate: a change in the ratio in the remains of plants indicates a change in the amount of photosynthetic activity, and thus in how favorable the environment was for the plants. During photosynthesis, organisms using the C3 pathway show different enrichments compared to those using the C4 pathway, allowing scientists not only to distinguish organic matter from abiotic carbon, but also what type of photosynthetic pathway the organic matter was using. Occasional spikes in the global 13C/12C ratio have also been useful as stratigraphic markers for chemostratigraphy, especially during the Paleozoic.
The 14C ratio has been used to track ocean circulation, among other things.
Nitrogen
Nitrogen has two stable isotopes, 14N
Document 3:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 4:::
The term stable isotope has a meaning similar to stable nuclide, but is preferably used when speaking of nuclides of a specific element. Hence, the plural form stable isotopes usually refers to isotopes of the same element. The relative abundance of such stable isotopes can be measured experimentally (isotope analysis), yielding an isotope ratio that can be used as a research tool. Theoretically, such stable isotopes could include the radiogenic daughter products of radioactive decay, used in radiometric dating. However, the expression stable-isotope ratio is preferably used to refer to isotopes whose relative abundances are affected by isotope fractionation in nature. This field is termed stable isotope geochemistry.
Stable-isotope ratios
Measurement of the ratios of naturally occurring stable isotopes (isotope analysis) plays an important role in isotope geochemistry, but stable isotopes (mostly hydrogen, carbon, nitrogen, oxygen and sulfur) are also finding uses in ecological and biological studies. Other workers have used oxygen isotope ratios to reconstruct historical atmospheric temperatures, making them important tools for paleoclimatology.
These isotope systems for lighter elements that exhibit more than one primordial isotope for each element have been under investigation for many years in order to study processes of isotope fractionation in natural systems. The long history of study of these elements is in part because the proportions of stable isotopes in these light and volatile elements is relatively easy to measure. However, recent advances in isotope ratio mass spectrometry (i.e. multiple-collector inductively coupled plasma mass spectrometry) now enable the measurement of isotope ratios in heavier stable elements, such as iron, copper, zinc, molybdenum, etc.
Applications
The variations in oxygen and hydrogen isotope ratios have applications in hydrology since most samples lie between two extremes, ocean water and Arctic/Antarctic snow. Given a sa
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Geologists frequently use mass spectrometers in determining the relative age of rocks. A spectrometer can differentiate between the ratios of isotopes in rock samples and calculates the radioactive decay involving what two isotopes?
A. Uranium - Lead
B. Rubidium - Strontium
C. Potassium - Argon
D. Uranium - Strontium
Answer:
|
|
sciq-5595
|
multiple_choice
|
Nerves of the hands, arms, feet, legs, and trunk are part of what system?
|
[
"circulatory system",
"cardiovascular system",
"urinary system",
"peripheral nervous system"
] |
D
|
Relavent Documents:
Document 0:::
The following diagram is provided as an overview of and topical guide to the human nervous system:
Human nervous system – the part of the human body that coordinates a person's voluntary and involuntary actions and transmits signals between different parts of the body. The human nervous system consists of two main parts: the central nervous system (CNS) and the peripheral nervous system (PNS). The CNS contains the brain and spinal cord. The PNS consists mainly of nerves, which are long fibers that connect the CNS to every other part of the body. The PNS includes motor neurons, mediating voluntary movement; the autonomic nervous system, comprising the sympathetic nervous system and the parasympathetic nervous system and regulating involuntary functions; and the enteric nervous system, a semi-independent part of the nervous system whose function is to control the gastrointestinal system.
Evolution of the human nervous system
Evolution of nervous systems
Evolution of human intelligence
Evolution of the human brain
Paleoneurology
Some branches of science that study the human nervous system
Neuroscience
Neurology
Paleoneurology
Central nervous system
The central nervous system (CNS) is the largest part of the nervous system and includes the brain and spinal cord.
Spinal cord
Brain
Brain – center of the nervous system.
Outline of the human brain
List of regions of the human brain
Principal regions of the vertebrate brain:
Peripheral nervous system
Peripheral nervous system (PNS) – nervous system structures that do not lie within the CNS.
Sensory system
A sensory system is a part of the nervous system responsible for processing sensory information. A sensory system consists of sensory receptors, neural pathways, and parts of the brain involved in sensory perception.
List of sensory systems
Sensory neuron
Perception
Visual system
Auditory system
Somatosensory system
Vestibular system
Olfactory system
Taste
Pain
Components of the nervous system
Neuron
I
Document 1:::
Body reactivity is usually understood as the body's ability to react in a proper way to influence the environment. Resistance of an organism is its stability under the influence of pathogenic factors. The body reactivity can range from homeostasis to a fight or flight response. Ultimately, they are all governed by the nervous system.
Nervous system divisions
The central nervous system (CNS) consists of parts that are encased by the bones of the skull and spinal column: the brain and spinal cord. The peripheral nervous system (PNS) is found outside those bones and consists of the nerves and most of the sensory organs.
Central nervous system
The CNS can be divided into the brain and spinal cord. The CNS processes many different kinds of incoming sensory information. It is also the source of thoughts, emotions, and memories. Most signals that stimulate muscles to contract and glands to secrete originate in the CNS. The spinal cord and spinal nerves contribute to homeostasis by providing quick reflexive responses to many stimuli. The spinal cord is the pathway for sensory input to the brain and motor output from the brain. The brain is responsible for integrating most sensory information and coordinating body function, both consciously and unconsciously.
Peripheral nervous system
The PNS can be divided into the autonomic and somatic nervous system. The autonomic nervous system can be divided into the parasympathetic, sympathetic, and enteric nervous system. The sympathetic nervous system regulates the “fight or flight” responses. The parasympathetic nervous system regulates the “rest and digest” responses. The enteric nervous system innervates the viscera (gastrointestinal tract, pancreas, and gall bladder). The somatic nervous system consists of peripheral nerve fibers that send sensory information to the central nervous system and motor nerve fibers that project to skeletal muscle. The somatic nervous system engages in voluntary reactions, and the autonomic nervous
Document 2:::
Cutaneous innervation refers to an area of the skin which is supplied by a specific cutaneous nerve.
Dermatomes are similar; however, a dermatome only specifies the area served by a spinal nerve. In some cases, the dermatome is less specific (when a spinal nerve is the source for more than one cutaneous nerve), and in other cases it is more specific (when a cutaneous nerve is derived from multiple spinal nerves.)
Modern texts are in agreement about which areas of the skin are served by which nerves, but there are minor variations in some of the details. The borders designated by the diagrams in the 1918 edition of Gray's Anatomy are similar, but not identical, to those generally accepted today.
Importance of the peripheral nervous system
The peripheral nervous system (PNS) is divided into the somatic nervous system, the autonomic nervous system, and the enteric nervous system. However, it is the somatic nervous system, responsible for body movement and the reception of external stimuli, which allows one to understand how cutaneous innervation is made possible by the action of specific sensory fibers located on the skin, as well as the distinct pathways they take to the central nervous system. The skin, which is part of the integumentary system, plays an important role in the somatic nervous system because it contains a range of nerve endings that react to heat and cold, touch, pressure, vibration, and tissue injury.
Importance of the central nervous system
The central nervous system (CNS) works with the peripheral nervous system in cutaneous innervation. The CNS is responsible for processing the information it receives from the cutaneous nerves that detect a given stimulus, and then identifying the kind of sensory inputs which project to a specific region of the primary somatosensory cortex.
The role of nerve endings on the surface of the skin
Groups of nerve terminals located in the different layers of the skin are categorized depending on whether the skin
Document 3:::
A biological system is a complex network which connects several biologically relevant entities. Biological organization spans several scales and are determined based different structures depending on what the system is. Examples of biological systems at the macro scale are populations of organisms. On the organ and tissue scale in mammals and other animals, examples include the circulatory system, the respiratory system, and the nervous system. On the micro to the nanoscopic scale, examples of biological systems are cells, organelles, macromolecular complexes and regulatory pathways. A biological system is not to be confused with a living system, such as a living organism.
Organ and tissue systems
These specific systems are widely studied in human anatomy and are also present in many other animals.
Respiratory system: the organs used for breathing, the pharynx, larynx, bronchi, lungs and diaphragm.
Digestive system: digestion and processing food with salivary glands, oesophagus, stomach, liver, gallbladder, pancreas, intestines, rectum and anus.
Cardiovascular system (heart and circulatory system): pumping and channeling blood to and from the body and lungs with heart, blood and blood vessels.
Urinary system: kidneys, ureters, bladder and urethra involved in fluid balance, electrolyte balance and excretion of urine.
Integumentary system: skin, hair, fat, and nails.
Skeletal system: structural support and protection with bones, cartilage, ligaments and tendons.
Endocrine system: communication within the body using hormones made by endocrine glands such as the hypothalamus, pituitary gland, pineal body or pineal gland, thyroid, parathyroid and adrenals, i.e., adrenal glands.
Lymphatic system: structures involved in the transfer of lymph between tissues and the blood stream; includes the lymph and the nodes and vessels. The lymphatic system includes functions including immune responses and development of antibodies.
Immune system: protects the organism from
Document 4:::
The sensory nervous system is a part of the nervous system responsible for processing sensory information. A sensory system consists of sensory neurons (including the sensory receptor cells), neural pathways, and parts of the brain involved in sensory perception and interoception. Commonly recognized sensory systems are those for vision, hearing, touch, taste, smell, balance and visceral sensation. Sense organs are transducers that convert data from the outer physical world to the realm of the mind where people interpret the information, creating their perception of the world around them.
The receptive field is the area of the body or environment to which a receptor organ and receptor cells respond. For instance, the part of the world an eye can see, is its receptive field; the light that each rod or cone can see, is its receptive field. Receptive fields have been identified for the visual system, auditory system and somatosensory system.
Stimulus
Organisms need information to solve at least three kinds of problems: (a) to maintain an appropriate environment, i.e., homeostasis; (b) to time activities (e.g., seasonal changes in behavior) or synchronize activities with those of conspecifics; and (c) to locate and respond to resources or threats (e.g., by moving towards resources or evading or attacking threats). Organisms also need to transmit information in order to influence another's behavior: to identify themselves, warn conspecifics of danger, coordinate activities, or deceive.
Sensory systems code for four aspects of a stimulus; type (modality), intensity, location, and duration. Arrival time of a sound pulse and phase differences of continuous sound are used for sound localization. Certain receptors are sensitive to certain types of stimuli (for example, different mechanoreceptors respond best to different kinds of touch stimuli, like sharp or blunt objects). Receptors send impulses in certain patterns to send information about the intensity of a stimul
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Nerves of the hands, arms, feet, legs, and trunk are part of what system?
A. circulatory system
B. cardiovascular system
C. urinary system
D. peripheral nervous system
Answer:
|
|
sciq-1978
|
multiple_choice
|
What do you call any process in which excess water or wastes are removed from the body?
|
[
"excretion",
"diffusion",
"depletion",
"digestion"
] |
A
|
Relavent Documents:
Document 0:::
The excretory system is a passive biological system that removes excess, unnecessary materials from the body fluids of an organism, so as to help maintain internal chemical homeostasis and prevent damage to the body. The dual function of excretory systems is the elimination of the waste products of metabolism and to drain the body of used up and broken down components in a liquid and gaseous state. In humans and other amniotes (mammals, birds and reptiles) most of these substances leave the body as urine and to some degree exhalation, mammals also expel them through sweating.
Only the organs specifically used for the excretion are considered a part of the excretory system. In the narrow sense, the term refers to the urinary system. However, as excretion involves several functions that are only superficially related, it is not usually used in more formal classifications of anatomy or function.
As most healthy functioning organs produce metabolic and other wastes, the entire organism depends on the function of the system. Breaking down of one of more of the systems is a serious health condition, for example kidney failure.
Systems
Urinary system
The kidneys are large, bean-shaped organs which are present on each side of the vertebral column in the abdominal cavity. Humans have two kidneys and each kidney is supplied with blood from the renal artery. The kidneys remove from the blood the nitrogenous wastes such as urea, as well as salts and excess water, and excrete them in the form of urine. This is done with the help of millions of nephrons present in the kidney. The filtrated blood is carried away from the kidneys by the renal vein (or kidney vein). The urine from the kidney is collected by the ureter (or excretory tubes), one from each kidney, and is passed to the urinary bladder. The urinary bladder collects and stores the urine until urination. The urine collected in the bladder is passed into the external environment from the body through an opening called
Document 1:::
Urine is a liquid by-product of metabolism in humans and in many other animals. Urine flows from the kidneys through the ureters to the urinary bladder. Urination results in urine being excreted from the body through the urethra.
Cellular metabolism generates many by-products that are rich in nitrogen and must be cleared from the bloodstream, such as urea, uric acid, and creatinine. These by-products are expelled from the body during urination, which is the primary method for excreting water-soluble chemicals from the body. A urinalysis can detect nitrogenous wastes of the mammalian body.
Urine plays an important role in the earth's nitrogen cycle. In balanced ecosystems, urine fertilizes the soil and thus helps plants to grow. Therefore, urine can be used as a fertilizer. Some animals use it to mark their territories. Historically, aged or fermented urine (known as lant) was also used for gunpowder production, household cleaning, tanning of leather and dyeing of textiles.
Human urine and feces are collectively referred to as human waste or human excreta, and are managed via sanitation systems. Livestock urine and feces also require proper management if the livestock population density is high.
Physiology
Most animals have excretory systems for elimination of soluble toxic wastes. In humans, soluble wastes are excreted primarily by the urinary system and, to a lesser extent in terms of urea, removed by perspiration. The urinary system consists of the kidneys, ureters, urinary bladder, and urethra. The system produces urine by a process of filtration, reabsorption, and tubular secretion. The kidneys extract the soluble wastes from the bloodstream, as well as excess water, sugars, and a variety of other compounds. The resulting urine contains high concentrations of urea and other substances, including toxins. Urine flows from the kidneys through the ureter, bladder, and finally the urethra before passing from the body.
Duration
Research looking at the duration
Document 2:::
Drinking is the act of ingesting water or other liquids into the body through the mouth, proboscis, or elsewhere. Humans drink by swallowing, completed by peristalsis in the esophagus. The physiological processes of drinking vary widely among other animals.
Most animals drink water to maintain bodily hydration, although many can survive on the water gained from their food. Water is required for many physiological processes. Both inadequate and (less commonly) excessive water intake are associated with health problems.
Methods of drinking
In humans
When a liquid enters a human mouth, the swallowing process is completed by peristalsis which delivers the liquid through the esophagus to the stomach; much of the activity is abetted by gravity. The liquid may be poured from the hands or drinkware may be used as vessels. Drinking can also be performed by acts of inhalation, typically when imbibing hot liquids or drinking from a spoon. Infants employ a method of suction wherein the lips are pressed tight around a source, as in breastfeeding: a combination of breath and tongue movement creates a vacuum which draws in liquid.
In other land mammals
By necessity, terrestrial animals in captivity become accustomed to drinking water, but most free-roaming animals stay hydrated through the fluids and moisture in fresh food, and learn to actively seek foods with high fluid content. When conditions impel them to drink from bodies of water, the methods and motions differ greatly among species.
Cats, canines, and ruminants all lower the neck and lap in water with their powerful tongues. Cats and canines lap up water with the tongue in a spoon-like shape. Canines lap water by scooping it into their mouth with a tongue which has taken the shape of a ladle. However, with cats, only the tip of their tongue (which is smooth) touches the water, and then the cat quickly pulls its tongue back into its mouth which soon closes; this results in a column of liquid being pulled into the ca
Document 3:::
Metabolic wastes or excrements are substances left over from metabolic processes (such as cellular respiration) which cannot be used by the organism (they are surplus or toxic), and must therefore be excreted. This includes nitrogen compounds, water, CO2, phosphates, sulphates, etc. Animals treat these compounds as excretes. Plants have metabolic pathways which transforms some of them (primarily the oxygen compounds) into useful substances..
All the metabolic wastes are excreted in a form of water solutes through the excretory organs (nephridia, Malpighian tubules, kidneys), with the exception of CO2, which is excreted together with the water vapor throughout the lungs. The elimination of these compounds enables the chemical homeostasis of the organism.
Nitrogen wastes
The nitrogen compounds through which excess nitrogen is eliminated from organisms are called nitrogenous wastes () or nitrogen wastes. They are ammonia, urea, uric acid, and creatinine. All of these substances are produced from protein metabolism. In many animals, the urine is the main route of excretion for such wastes; in some, it is the feces.
Ammonotelism
Ammonotelism is the excretion of ammonia and ammonium ions. Ammonia (NH3) forms with the oxidation of amino groups.(-NH2), which are removed from the proteins when they convert into carbohydrates. It is a very toxic substance to tissues and extremely soluble in water. Only one nitrogen atom is removed with it. A lot of water is needed for the excretion of ammonia, about 0.5 L of water is needed per 1 g of nitrogen to maintain ammonia levels in the excretory fluid below the level in body fluids to prevent toxicity. Thus, the marine organisms excrete ammonia directly into the water and are called ammonotelic. Ammonotelic animals include crustaceans, platyhelminths, cnidarians, poriferans, echinoderms, and other aquatic invertebrates.
Ureotelism
The excretion of urea is called ureotelism. Land animals, mainly amphibians and mammals, convert
Document 4:::
Evapotranspiration (ET) is the combined processes which move water from the Earth's surface into the atmosphere. It covers both water evaporation (movement of water to the air directly from soil, canopies, and water bodies) and transpiration (evaporation that occurs through the stomata, or openings, in plant leaves). Evapotranspiration is an important part of the local water cycle and climate, and measurement of it plays a key role in agricultural irrigation and water resource management.
Definition of evapotranspiration
Evapotranspiration is a combination of evaporation and transpiration, measured in order to better understand crop water requirements, irrigation scheduling, and watershed management. The two key components of evapotranspiration are:
Evaporation: the movement of water directly to the air from sources such as the soil and water bodies. It can be affected by factors including heat, humidity, solar radiation and wind speed.
Transpiration: the movement of water from root systems, through a plant, and exit into the air as water vapor. This exit occurs through stomata in the plant. Rate of transpiration can be influenced by factors including plant type, soil type, weather conditions and water content, and also cultivation practices.
Evapotranspiration is typically measured in millimeters of water (i.e. volume of water moved per unit area of the Earth's surface) in a set unit of time. Globally, it is estimated that on average between three-fifths and three-quarters of land precipitation is returned to the atmosphere via evapotranspiration.
Evapotranspiration does not, in general, account for other mechanisms which are involved in returning water to the atmosphere, though some of these, such as snow and ice sublimation in regions of high elevation or high latitude, can make a large contribution to atmospheric moisture even under standard conditions.
Factors that impact evapotranspiration levels
Primary factors
Because evaporation and transpiration
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What do you call any process in which excess water or wastes are removed from the body?
A. excretion
B. diffusion
C. depletion
D. digestion
Answer:
|
|
ai2_arc-580
|
multiple_choice
|
Which process is most directly related to constructing new land?
|
[
"erosion",
"polluting",
"deposition",
"weathering"
] |
C
|
Relavent Documents:
Document 0:::
Plant ecology is a subdiscipline of ecology that studies the distribution and abundance of plants, the effects of environmental factors upon the abundance of plants, and the interactions among plants and between plants and other organisms. Examples of these are the distribution of temperate deciduous forests in North America, the effects of drought or flooding upon plant survival, and competition among desert plants for water, or effects of herds of grazing animals upon the composition of grasslands.
A global overview of the Earth's major vegetation types is provided by O.W. Archibold. He recognizes 11 major vegetation types: tropical forests, tropical savannas, arid regions (deserts), Mediterranean ecosystems, temperate forest ecosystems, temperate grasslands, coniferous forests, tundra (both polar and high mountain), terrestrial wetlands, freshwater ecosystems and coastal/marine systems. This breadth of topics shows the complexity of plant ecology, since it includes plants from floating single-celled algae up to large canopy forming trees.
One feature that defines plants is photosynthesis. Photosynthesis is the process of a chemical reactions to create glucose and oxygen, which is vital for plant life. One of the most important aspects of plant ecology is the role plants have played in creating the oxygenated atmosphere of earth, an event that occurred some 2 billion years ago. It can be dated by the deposition of banded iron formations, distinctive sedimentary rocks with large amounts of iron oxide. At the same time, plants began removing carbon dioxide from the atmosphere, thereby initiating the process of controlling Earth's climate. A long term trend of the Earth has been toward increasing oxygen and decreasing carbon dioxide, and many other events in the Earth's history, like the first movement of life onto land, are likely tied to this sequence of events.
One of the early classic books on plant ecology was written by J.E. Weaver and F.E. Clements. It
Document 1:::
Ian Gordon Simmons (born 22 January 1937) is a British geographer. He retired as Professor of Geography from the University of Durham in 2001. He has made significant contributions to environmental history and prehistoric archaeology.
Background
Simmons grew up in East London and then East Lincolnshire until the age of 12. He studied physical geography (BSc) and holds a PhD from the University of London (early 1960s) on the vegetation history of Dartmoor. He began university lecturing in his early 20s and was Lecturer and then Reader in Geography at the University of Durham from 1962 to 1977, then Professor of Geography at the University of Bristol from 1977 to 1981 before returning to a Chair in Geography at Durham, where he worked until retiring in 2001.
In 1972–73, he taught biogeography for a year at York University, Canada and has held other appointments including Visiting Scholar, St. John's College, University of Oxford in the 1990s. Previously, he had been an ACLS postdoctoral fellow at the University of California, Berkeley.
Scholarship
His research includes the study of the later Mesolithic and early Neolithic in their environmental setting on English uplands, where he has demonstrated the role of these early human communities in initiating some of Britain's characteristic landscape elements. His work also encompasses the long-term effects of human manipulation of the natural environment and its consequences for resource use and environmental change. This line of work resulted in his last three books, which looked at environmental history on three nested scales: the moorlands of England and Wales, Great Britain, and the Globe. Each dealt with the last 10,000 years and tried to encompassboth conventional science-based data with the insights of the social sciences and humanities.
Simmons has authored several books on environmental thought and culture over the ages as well as contemporary resource management and environmental problems. Since retireme
Document 2:::
The Géotechnique lecture is an biennial lecture on the topic of soil mechanics, organised by the British Geotechnical Association named after its major scientific journal Géotechnique.
This should not be confused with the annual BGA Rankine Lecture.
List of Géotechnique Lecturers
See also
Named lectures
Rankine Lecture
Terzaghi Lecture
External links
ICE Géotechnique journal
British Geotechnical Association
Document 3:::
Primary succession is the beginning step of ecological succession after an extreme disturbance, which usually occurs in an environment devoid of vegetation and other organisms. These environments are typically lacking in soil, as disturbances like lava flow or retreating glaciers scour the environment clear of nutrients.
In contrast, secondary succession occurs on substrates that previously supported vegetation before an ecological disturbance. This occurs when smaller disturbances like floods, hurricanes, tornadoes, and fires destroy only the local plant life and leave soil nutrients for immediate establishment by intermediate community species.
Occurrence
In primary succession pioneer species like lichen, algae and fungi as well as abiotic factors like wind and water start to "normalise" the habitat or in other words start to develop soil and other important mechanisms for greater diversity to flourish. Primary succession begins on rock formations, such as volcanoes or mountains, or in a place with no organisms or soil. Primary succession leads to conditions nearer optimum for vascular plant growth; pedogenesis or the formation of soil, and the increased amount of shade are the most important processes.
These pioneer lichen, algae, and fungi are then dominated and often replaced by plants that are better adapted to less harsh conditions, these plants include vascular plants like grasses and some shrubs that are able to live in thin soils that are often mineral-based. Water and nutrient levels increase with the amount of succession exhibited.
The early stages of primary succession are dominated by species with small propagules (seed and spores) which can be dispersed long distances. The early colonizers—often algae, fungi, and lichens—stabilize the substrate. Nitrogen supplies are limited in new soils, and nitrogen-fixing species tend to play an important role early in primary succession. Unlike in primary succession, the species that dominate secondary success
Document 4:::
Land cover is the physical material at the surface of Earth. Land covers include grass, asphalt, trees, bare ground, water, etc. Earth cover is the expression used by ecologist Frederick Edward Clements that has its closest modern equivalent being vegetation. The expression continues to be used by the United States Bureau of Land Management.
There are two primary methods for capturing information on land cover: field survey, and analysis of remotely sensed imagery. Land change models can be built from these types of data to assess changes in land cover over time.
One of the major land cover issues (as with all natural resource inventories) is that every survey defines similarly named categories in different ways. For instance, there are many definitions of "forest"—sometimes within the same organisation—that may or may not incorporate a number of different forest features (e.g., stand height, canopy cover, strip width, inclusion of grasses, and rates of growth for timber production). Areas without trees may be classified as forest cover "if the intention is to re-plant" (UK and Ireland), while areas with many trees may not be labelled as forest "if the trees are not growing fast enough" (Norway and Finland).
Distinction from "land use"
"Land cover" is distinct from "land use", despite the two terms often being used interchangeably. Land use is a description of how people utilize the land and of socio-economic activity. Urban and agricultural land uses are two of the most commonly known land use classes. At any one point or place, there may be multiple and alternate land uses, the specification of which may have a political dimension. The origins of the "land cover/land use" couplet and the implications of their confusion are discussed in Fisher et al. (2005).
Types
Following table is Land Cover statistics by Food and Agriculture Organization (FAO) with 14 classes.
Mapping
Land cover change detection using remote sensing and geospatial data provides baselin
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Which process is most directly related to constructing new land?
A. erosion
B. polluting
C. deposition
D. weathering
Answer:
|
|
sciq-7834
|
multiple_choice
|
What is the term for a gas that completely abides by the kinetic-molecular theory?
|
[
"cold gas",
"fast gas",
"heat gas",
"ideal gas"
] |
D
|
Relavent Documents:
Document 0:::
In physics and chemistry, "monatomic" is a combination of the words "mono" and "atomic", and means "single atom". It is usually applied to gases: a monatomic gas is a gas in which atoms are not bound to each other. Examples at standard conditions of temperature and pressure include all the noble gases (helium, neon, argon, krypton, xenon, and radon), though all chemical elements will be monatomic in the gas phase at sufficiently high temperature (or very low pressure). The thermodynamic behavior of a monatomic gas is much simpler when compared to polyatomic gases because it is free of any rotational or vibrational energy.
Noble gases
The only chemical elements that are stable single atoms (so they are not molecules) at standard temperature and pressure (STP) are the noble gases. These are helium, neon, argon, krypton, xenon, and radon. Noble gases have a full outer valence shell making them rather non-reactive species. While these elements have been described historically as completely inert, chemical compounds have been synthesized with all but neon and helium.
When grouped together with the homonuclear diatomic gases such as nitrogen (N2), the noble gases are called "elemental gases" to distinguish them from molecules that are also chemical compounds.
Thermodynamic properties
The only possible motion of an atom in a monatomic gas is translation (electronic excitation is not important at room temperature). Thus by the equipartition theorem, the kinetic energy of a single atom of a monatomic gas at thermodynamic temperature T is given by , where kb is Boltzmann's constant. One mole of atoms contains an Avogadro number () of atoms, so that the energy of one mole of atoms of a monatomic gas is , where R is the gas constant.
In an adiabatic process, monatomic gases have an idealised γ-factor (Cp/Cv) of 5/3, as opposed to 7/5 for ideal diatomic gases where rotation (but not vibration at room temperature) also contributes. Also, for ideal monatomic gases:
Document 1:::
A Knudsen gas is a gas in a state of such low density that the average distance travelled by the gas molecules between collisions (mean free path) is greater than the diameter of the receptacle that contains it. If the mean free path is much greater than the diameter, the flow regime is dominated by collisions between the gas molecules and the walls of the receptacle, rather than intermolecular collisions with each other. It is named after Martin Knudsen.
Knudsen number
For a Knudsen gas, the Knudsen number must be greater than 1. The Knudsen number can be defined as:
where
is the mean free path [m]
is the diameter of the receptacle [m].
When , the flow regime of the gas is transitional flow. In this regime the intermolecular collisions between gas particles are not yet negligible compared to collisions with the wall. However when , the flow regime is free molecular flow, so the intermolecular collisions between the particles are negligible compared to the collisions with the wall.
Example
For example, consider a receptacle of air at room temperature and pressure with a mean free path of 68nm. If the diameter of the receptacle is less than 68nm, the Knudsen number would greater than 1, and this sample of air would be considered a Knudsen gas. It would not be a Knudsen gas if the diameter of the receptacle is greater than 68nm.
Document 2:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 3:::
In physics and engineering, a perfect gas is a theoretical gas model that differs from real gases in specific ways that makes certain calculations easier to handle. In all perfect gas models, intermolecular forces are neglected. This means that one can neglect many complications that may arise from the Van der Waals forces. All perfect gas models are ideal gas models in the sense that they all follow the ideal gas equation of state. However, the idea of a perfect gas model is often invoked as a combination of the ideal gas equation of state with specific additional assumptions regarding the variation (or nonvariation) of the heat capacity with temperature.
Perfect gas nomenclature
The terms perfect gas and ideal gas are sometimes used interchangeably, depending on the particular field of physics and engineering. Sometimes, other distinctions are made, such as between thermally perfect gas and calorically perfect gas, or between imperfect, semi-perfect, and perfect gases, and as well as the characteristics of ideal gases. Two of the common sets of nomenclatures are summarized in the following table.
Thermally and calorically perfect gas
Along with the definition of a perfect gas, there are also two more simplifications that can be made although various textbooks either omit or combine the following simplifications into a general "perfect gas" definition.
For a fixed number of moles of gas , a thermally perfect gas
is in thermodynamic equilibrium
is not chemically reacting
has internal energy , enthalpy , and constant volume / constant pressure heat capacities , that are solely functions of temperature and not of pressure or volume , i.e., , , , . These latter expressions hold for all tiny property changes and are not restricted to constant- or constant- variations.
A calorically perfect gas
is in thermodynamic equilibrium
is not chemically reacting
has internal energy , and enthalpy that are functions of temperature only, i.e., ,
has heat capacities
Document 4:::
The kinetic theory of gases is a simple, historically significant classical model of the thermodynamic behavior of gases, with which many principal concepts of thermodynamics were established. The model describes a gas as a large number of identical submicroscopic particles (atoms or molecules), all of which are in constant, rapid, random motion. Their size is assumed to be much smaller than the average distance between the particles. The particles undergo random elastic collisions between themselves and with the enclosing walls of the container. The basic version of the model describes the ideal gas, and considers no other interactions between the particles.
The kinetic theory of gases explains the macroscopic properties of gases, such as volume, pressure, and temperature, as well as transport properties such as viscosity, thermal conductivity and mass diffusivity. Due to the time reversibility of microscopic dynamics (microscopic reversibility), the kinetic theory is also connected to the principle of detailed balance, in terms of the fluctuation-dissipation theorem (for Brownian motion) and the Onsager reciprocal relations.
Historically, the kinetic theory of gases was the first explicit exercise of the ideas of statistical mechanics.
History
In about 50 BCE, the Roman philosopher Lucretius proposed that apparently static macroscopic bodies were composed on a small scale of rapidly moving atoms all bouncing off each other. This Epicurean atomistic point of view was rarely considered in the subsequent centuries, when Aristotlean ideas were dominant.
In 1738 Daniel Bernoulli published Hydrodynamica, which laid the basis for the kinetic theory of gases. In this work, Bernoulli posited the argument, that gases consist of great numbers of molecules moving in all directions, that their impact on a surface causes the pressure of the gas, and that their average kinetic energy determines the temperature of the gas. The theory was not immediately accepted, in part be
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is the term for a gas that completely abides by the kinetic-molecular theory?
A. cold gas
B. fast gas
C. heat gas
D. ideal gas
Answer:
|
|
sciq-10661
|
multiple_choice
|
What system is designed to facilitate the transformation of food matter into the nutrient components that sustain organisms?
|
[
"respiratory",
"digestive",
"tissue",
"skeletal"
] |
B
|
Relavent Documents:
Document 0:::
A nutrient is a substance used by an organism to survive, grow, and reproduce. The requirement for dietary nutrient intake applies to animals, plants, fungi, and protists. Nutrients can be incorporated into cells for metabolic purposes or excreted by cells to create non-cellular structures, such as hair, scales, feathers, or exoskeletons. Some nutrients can be metabolically converted to smaller molecules in the process of releasing energy, such as for carbohydrates, lipids, proteins, and fermentation products (ethanol or vinegar), leading to end-products of water and carbon dioxide. All organisms require water. Essential nutrients for animals are the energy sources, some of the amino acids that are combined to create proteins, a subset of fatty acids, vitamins and certain minerals. Plants require more diverse minerals absorbed through roots, plus carbon dioxide and oxygen absorbed through leaves. Fungi live on dead or living organic matter and meet nutrient needs from their host.
Different types of organisms have different essential nutrients. Ascorbic acid (vitamin C) is essential, meaning it must be consumed in sufficient amounts, to humans and some other animal species, but some animals and plants are able to synthesize it. Nutrients may be organic or inorganic: organic compounds include most compounds containing carbon, while all other chemicals are inorganic. Inorganic nutrients include nutrients such as iron, selenium, and zinc, while organic nutrients include, among many others, energy-providing compounds and vitamins.
A classification used primarily to describe nutrient needs of animals divides nutrients into macronutrients and micronutrients. Consumed in relatively large amounts (grams or ounces), macronutrients (carbohydrates, fats, proteins, water) are primarily used to generate energy or to incorporate into tissues for growth and repair. Micronutrients are needed in smaller amounts (milligrams or micrograms); they have subtle biochemical and physiologi
Document 1:::
The soil food web is the community of organisms living all or part of their lives in the soil. It describes a complex living system in the soil and how it interacts with the environment, plants, and animals.
Food webs describe the transfer of energy between species in an ecosystem. While a food chain examines one, linear, energy pathway through an ecosystem, a food web is more complex and illustrates all of the potential pathways. Much of this transferred energy comes from the sun. Plants use the sun’s energy to convert inorganic compounds into energy-rich, organic compounds, turning carbon dioxide and minerals into plant material by photosynthesis. Plant flowers exude energy-rich nectar above ground and plant roots exude acids, sugars, and ectoenzymes into the rhizosphere, adjusting the pH and feeding the food web underground.
Plants are called autotrophs because they make their own energy; they are also called producers because they produce energy available for other organisms to eat. Heterotrophs are consumers that cannot make their own food. In order to obtain energy they eat plants or other heterotrophs.
Above ground food webs
In above ground food webs, energy moves from producers (plants) to primary consumers (herbivores) and then to secondary consumers (predators). The phrase, trophic level, refers to the different levels or steps in the energy pathway. In other words, the producers, consumers, and decomposers are the main trophic levels. This chain of energy transferring from one species to another can continue several more times, but eventually ends. At the end of the food chain, decomposers such as bacteria and fungi break down dead plant and animal material into simple nutrients.
Methodology
The nature of soil makes direct observation of food webs difficult. Since soil organisms range in size from less than 0.1 mm (nematodes) to greater than 2 mm (earthworms) there are many different ways to extract them. Soil samples are often taken using a metal
Document 2:::
Consumer–resource interactions are the core motif of ecological food chains or food webs, and are an umbrella term for a variety of more specialized types of biological species interactions including prey-predator (see predation), host-parasite (see parasitism), plant-herbivore and victim-exploiter systems. These kinds of interactions have been studied and modeled by population ecologists for nearly a century. Species at the bottom of the food chain, such as algae and other autotrophs, consume non-biological resources, such as minerals and nutrients of various kinds, and they derive their energy from light (photons) or chemical sources. Species higher up in the food chain survive by consuming other species and can be classified by what they eat and how they obtain or find their food.
Classification of consumer types
The standard categorization
Various terms have arisen to define consumers by what they eat, such as meat-eating carnivores, fish-eating piscivores, insect-eating insectivores, plant-eating herbivores, seed-eating granivores, and fruit-eating frugivores and omnivores are meat eaters and plant eaters. An extensive classification of consumer categories based on a list of feeding behaviors exists.
The Getz categorization
Another way of categorizing consumers, proposed by South African American ecologist Wayne Getz, is based on a biomass transformation web (BTW) formulation that organizes resources into five components: live and dead animal, live and dead plant, and particulate (i.e. broken down plant and animal) matter. It also distinguishes between consumers that gather their resources by moving across landscapes from those that mine their resources by becoming sessile once they have located a stock of resources large enough for them to feed on during completion of a full life history stage.
In Getz's scheme, words for miners are of Greek etymology and words for gatherers are of Latin etymology. Thus a bestivore, such as a cat, preys on live animal
Document 3:::
The trophic level of an organism is the position it occupies in a food web. A food chain is a succession of organisms that eat other organisms and may, in turn, be eaten themselves. The trophic level of an organism is the number of steps it is from the start of the chain. A food web starts at trophic level 1 with primary producers such as plants, can move to herbivores at level 2, carnivores at level 3 or higher, and typically finish with apex predators at level 4 or 5. The path along the chain can form either a one-way flow or a food "web". Ecological communities with higher biodiversity form more complex trophic paths.
The word trophic derives from the Greek τροφή (trophē) referring to food or nourishment.
History
The concept of trophic level was developed by Raymond Lindeman (1942), based on the terminology of August Thienemann (1926): "producers", "consumers", and "reducers" (modified to "decomposers" by Lindeman).
Overview
The three basic ways in which organisms get food are as producers, consumers, and decomposers.
Producers (autotrophs) are typically plants or algae. Plants and algae do not usually eat other organisms, but pull nutrients from the soil or the ocean and manufacture their own food using photosynthesis. For this reason, they are called primary producers. In this way, it is energy from the sun that usually powers the base of the food chain. An exception occurs in deep-sea hydrothermal ecosystems, where there is no sunlight. Here primary producers manufacture food through a process called chemosynthesis.
Consumers (heterotrophs) are species that cannot manufacture their own food and need to consume other organisms. Animals that eat primary producers (like plants) are called herbivores. Animals that eat other animals are called carnivores, and animals that eat both plants and other animals are called omnivores.
Decomposers (detritivores) break down dead plant and animal material and wastes and release it again as energy and nutrients into
Document 4:::
Relatively speaking, the brain consumes an immense amount of energy in comparison to the rest of the body. The mechanisms involved in the transfer of energy from foods to neurons are likely to be fundamental to the control of brain function. Human bodily processes, including the brain, all require both macronutrients, as well as micronutrients.
Insufficient intake of selected vitamins, or certain metabolic disorders, may affect cognitive processes by disrupting the nutrient-dependent processes within the body that are associated with the management of energy in neurons, which can subsequently affect synaptic plasticity, or the ability to encode new memories.
Macronutrients
The human brain requires nutrients obtained from the diet to develop and sustain its physical structure and cognitive functions. Additionally, the brain requires caloric energy predominately derived from the primary macronutrients to operate. The three primary macronutrients include carbohydrates, proteins, and fats. Each macronutrient can impact cognition through multiple mechanisms, including glucose and insulin metabolism, neurotransmitter actions, oxidative stress and inflammation, and the gut-brain axis. Inadequate macronutrient consumption or proportion could impair optimal cognitive functioning and have long-term health implications.
Carbohydrates
Through digestion, dietary carbohydrates are broken down and converted into glucose, which is the sole energy source for the brain. Optimal brain function relies on adequate carbohydrate consumption, as carbohydrates provide the quickest source of glucose for the brain. Glucose deficiencies such as hypoglycaemia reduce available energy for the brain and impair all cognitive processes and performance. Additionally, situations with high cognitive demand, such as learning a new task, increase brain glucose utilization, depleting blood glucose stores and initiating the need for supplementation.
Complex carbohydrates, especially those with high d
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What system is designed to facilitate the transformation of food matter into the nutrient components that sustain organisms?
A. respiratory
B. digestive
C. tissue
D. skeletal
Answer:
|
|
sciq-11098
|
multiple_choice
|
What group of simple organisms recycles nutrients and has been on earth since long before multicellular life appeared?
|
[
"protists",
"prokaryotes",
"eukaryotes",
"bacteria"
] |
B
|
Relavent Documents:
Document 0:::
The eukaryotes () constitute the domain of Eukarya, organisms whose cells have a membrane-bound nucleus. All animals, plants, fungi, and many unicellular organisms are eukaryotes. They constitute a major group of life forms alongside the two groups of prokaryotes: the Bacteria and the Archaea. Eukaryotes represent a small minority of the number of organisms, but due to their generally much larger size, their collective global biomass is much larger than that of prokaryotes.
The eukaryotes seemingly emerged in the Archaea, within the Asgard archaea. This implies that there are only two domains of life, Bacteria and Archaea, with eukaryotes incorporated among the Archaea. Eukaryotes emerged approximately 2.2 billion years ago, during the Proterozoic eon, likely as flagellated cells. The leading evolutionary theory is they were created by symbiogenesis between an anaerobic Asgard archaean and an aerobic proteobacterium, which formed the mitochondria. A second episode of symbiogenesis with a cyanobacterium created the plants, with chloroplasts. The oldest-known eukaryote fossils, multicellular planktonic organisms belonging to the Gabonionta, were discovered in Gabon in 2023, dating back to 2.1 billion years ago.
Eukaryotic cells contain membrane-bound organelles such as the nucleus, the endoplasmic reticulum, and the Golgi apparatus. Eukaryotes may be either unicellular or multicellular. In comparison, prokaryotes are typically unicellular. Unicellular eukaryotes are sometimes called protists. Eukaryotes can reproduce both asexually through mitosis and sexually through meiosis and gamete fusion (fertilization).
Diversity
Eukaryotes are organisms that range from microscopic single cells, such as picozoans under 3 micrometres across, to animals like the blue whale, weighing up to 190 tonnes and measuring up to long, or plants like the coast redwood, up to tall. Many eukaryotes are unicellular; the informal grouping called protists includes many of these, with some
Document 1:::
Marine prokaryotes are marine bacteria and marine archaea. They are defined by their habitat as prokaryotes that live in marine environments, that is, in the saltwater of seas or oceans or the brackish water of coastal estuaries. All cellular life forms can be divided into prokaryotes and eukaryotes. Eukaryotes are organisms whose cells have a nucleus enclosed within membranes, whereas prokaryotes are the organisms that do not have a nucleus enclosed within a membrane. The three-domain system of classifying life adds another division: the prokaryotes are divided into two domains of life, the microscopic bacteria and the microscopic archaea, while everything else, the eukaryotes, become the third domain.
Prokaryotes play important roles in ecosystems as decomposers recycling nutrients. Some prokaryotes are pathogenic, causing disease and even death in plants and animals. Marine prokaryotes are responsible for significant levels of the photosynthesis that occurs in the ocean, as well as significant cycling of carbon and other nutrients.
Prokaryotes live throughout the biosphere. In 2018 it was estimated the total biomass of all prokaryotes on the planet was equivalent to 77 billion tonnes of carbon (77 Gt C). This is made up of 7 Gt C for archaea and 70 Gt C for bacteria. These figures can be contrasted with the estimate for the total biomass for animals on the planet, which is about 2 Gt C, and the total biomass of humans, which is 0.06 Gt C. This means archaea collectively have over 100 times the collective biomass of humans, and bacteria over 1000 times.
There is no clear evidence of life on Earth during the first 600 million years of its existence. When life did arrive, it was dominated for 3,200 million years by the marine prokaryotes. More complex life, in the form of crown eukaryotes, didn't appear until the Cambrian explosion a mere 500 million years ago.
Evolution
The Earth is about 4.54 billion years old. The earliest undisputed evidence of life on Eart
Document 2:::
Eukaryogenesis, the process which created the eukaryotic cell and lineage, is a milestone in the evolution of life, since eukaryotes include all complex cells and almost all multicellular organisms. The process is widely agreed to have involved symbiogenesis, in which archaea and bacteria came together to create the first eukaryotic common ancestor (FECA). This cell had a new level of complexity and capability, with a nucleus, at least one centriole and cilium, facultatively aerobic mitochondria, sex (meiosis and syngamy), a dormant cyst with a cell wall of chitin and/or cellulose and peroxisomes. It evolved into a population of single-celled organisms that included the last eukaryotic common ancestor (LECA), gaining capabilities along the way, though the sequence of the steps involved has been disputed, and may not have started with symbiogenesis. In turn, the LECA gave rise to the eukaryotes' crown group, containing the ancestors of animals, fungi, plants, and a diverse range of single-celled organisms.
Context
Life arose on Earth once it had cooled enough for oceans to form. The last universal common ancestor (LUCA) was an organism which had ribosomes and the genetic code; it lived some 4 billion years ago. It gave rise to two main branches of prokaryotic life, the bacteria and the archaea. From among these small-celled, rapidly-dividing ancestors arose the Eukaryotes, with much larger cells, nuclei, and distinctive biochemistry. The eukaryotes form a domain that contains all complex cells and most types of multicellular organism, including the animals, plants, and fungi.
Symbiogenesis
According to the theory of symbiogenesis (also known as the endosymbiotic theory) championed by Lynn Margulis, a member of the archaea gained a bacterial cell as a component. The archaeal cell was a member of the Asgard group. The bacterium was one of the Alphaproteobacteria, which had the ability to use oxygen in its respiration. This enabled it – and the archaeal cells that
Document 3:::
Microorganisms engage in a wide variety of social interactions, including cooperation. A cooperative behavior is one that benefits an individual (the recipient) other than the one performing the behavior (the actor). This article outlines the various forms of cooperative interactions (mutualism and altruism) seen in microbial systems, as well as the benefits that might have driven the evolution of these complex behaviors.
Introduction
Microorganisms, or microbes, span all three domains of life – bacteria, archaea, and many unicellular eukaryotes including some fungi and protists. Typically defined as unicellular life forms that can only be observed with a microscope, microorganisms were the first cellular life forms, and were critical for creating the conditions for the evolution of more complex multicellular forms.
Although microbes are too small to see with the naked eye, they represent the overwhelming majority of biological diversity, and thus serve as an excellent system to study evolutionary questions. One such topic that scientists have examined in microbes is the evolution of social behaviors, including cooperation. A cooperative interaction benefits a recipient, and is selected for on that basis. In microbial systems, cells belonging to the same taxa have been documented partaking in cooperative interactions to perform a wide range of complex multicellular behaviors such as dispersal, foraging, construction of biofilms, reproduction, chemical warfare, and signaling. This article will outline the various forms of cooperative interactions seen in microbial systems, as well as the benefits that might have driven the evolution of these complex behaviors.
History
From an evolutionary point of view, a behavior is social if it has fitness consequences for both the individual that performs that behavior (the actor) and another individual (the recipient). Hamilton first categorized social behaviors according to whether the consequences they entail for the actor
Document 4:::
The smallest organisms found on Earth can be determined according to various aspects of organism size, including volume, mass, height, length, or genome size.
Given the incomplete nature of scientific knowledge, it is possible that the smallest organism is undiscovered. Furthermore, there is some debate over the definition of life, and what entities qualify as organisms; consequently the smallest known organism (microorganism) is debatable.
Microorganisms
Obligate endosymbiotic bacteria
The genome of Nasuia deltocephalinicola, a symbiont of the European pest leafhopper, Macrosteles quadripunctulatus, consists of a circular chromosome of 112,031 base pairs.
The genome of Nanoarchaeum equitans is 491 Kbp nucleotides long.
Pelagibacter ubique
Pelagibacter ubique is one of the smallest known free-living bacteria, with a length of and an average cell diameter of . They also have the smallest free-living bacterium genome: 1.3 Mbp, 1354 protein genes, 35 RNA genes. They are one of the most common and smallest organisms in the ocean, with their total weight exceeding that of all fish in the sea.
Mycoplasma genitalium
Mycoplasma genitalium, a parasitic bacterium which lives in the primate bladder, waste disposal organs, genital, and respiratory tracts, is thought to be the smallest known organism capable of independent growth and reproduction. With a size of approximately 200 to 300 nm, M. genitalium is an ultramicrobacterium, smaller than other small bacteria, including rickettsia and chlamydia. However, the vast majority of bacterial strains have not been studied, and the marine ultramicrobacterium Sphingomonas sp. strain RB2256 is reported to have passed through a ultrafilter. A complicating factor is nutrient-downsized bacteria, bacteria that become much smaller due to a lack of available nutrients.
Nanoarchaeum
Nanoarchaeum equitans is a species of microbe in diameter. It was discovered in 2002 in a hydrothermal vent off the coast of Iceland by Karl Stet
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What group of simple organisms recycles nutrients and has been on earth since long before multicellular life appeared?
A. protists
B. prokaryotes
C. eukaryotes
D. bacteria
Answer:
|
|
sciq-2100
|
multiple_choice
|
Involved in balance, the semicircular canals are parts of what organs?
|
[
"eyes",
"ears",
"lungs",
"ovaries"
] |
B
|
Relavent Documents:
Document 0:::
The semicircular canals or semicircular ducts are three semicircular, interconnected tubes located in the innermost part of each ear, the inner ear. The three canals are the horizontal, superior and posterior semicircular canals.
Structure
The semicircular canals are a component of the bony labyrinth that are at right angles from each other. At one end of each of the semicircular canals is a dilated sac called an osseous ampulla, which is more than twice the diameter of the canal. Each ampulla contains an ampullary crest, the crista ampullaris which consists of a thick gelatinous cap called a cupula and many hair cells. The superior and posterior semicircular canals are oriented vertically at right angles to each other. The lateral semicircular canal is about a 30-degree angle from the horizontal plane. The orientations of the canals cause a different canal to be stimulated by movement of the head in different planes, and more than one canal is stimulated at once if the movement is off those planes. The lateral canal detects angular acceleration of the head when the head is turned and the superior and posterior canals detect vertical head movements when the head is moved up or down. When the head changes position, the endolymph in the canals lags behind due to inertia and this acts on the cupula which bends the cilia of the hair cells. The stimulation of the hair cells sends the message to the brain that acceleration is taking place. The ampullae open into the vestibule by five orifices, one of the apertures being common to two of the canals.
Among species of mammals, the size of the semicircular canals is correlated with their type of locomotion. Specifically, species that are agile and have fast, jerky locomotion have larger canals relative to their body size than those that move more cautiously.
Horizontal semicircular canal
The lateral or horizontal canal (external semicircular canal) is the shortest of the three canals. Movement of fluid within this canal c
Document 1:::
The vestibular system, in vertebrates, is a sensory system that creates the sense of balance and spatial orientation for the purpose of coordinating movement with balance. Together with the cochlea, a part of the auditory system, it constitutes the labyrinth of the inner ear in most mammals.
As movements consist of rotations and translations, the vestibular system comprises two components: the semicircular canals, which indicate rotational movements; and the otoliths, which indicate linear accelerations. The vestibular system sends signals primarily to the neural structures that control eye movement; these provide the anatomical basis of the vestibulo-ocular reflex, which is required for clear vision. Signals are also sent to the muscles that keep an animal upright and in general control posture; these provide the anatomical means required to enable an animal to maintain its desired position in space.
The brain uses information from the vestibular system in the head and from proprioception throughout the body to enable the animal to understand its body's dynamics and kinematics (including its position and acceleration) from moment to moment. How these two perceptive sources are integrated to provide the underlying structure of the sensorium is unknown.
Semicircular canal system
The semicircular canal system detects rotational movements. Semicircular canals are its main tools to achieve this detection.
Structure
Since the world is three-dimensional, the vestibular system contains three semicircular canals in each labyrinth. They are approximately orthogonal (at right angles) to each other, and are the horizontal (or lateral), the anterior semicircular canal (or superior), and the posterior (or inferior) semicircular canal. Anterior and posterior canals may collectively be called vertical semicircular canals.
Movement of fluid within the horizontal semicircular canal corresponds to rotation of the head around a vertical axis (i.e. the neck), as when doing a pi
Document 2:::
Medial to the opening for the carotid canal and close to its posterior border, in front of the jugular fossa, is a triangular depression; at the apex of this is a small opening, the aquaeductus cochleae (or cochlear aqueduct, or aqueduct of cochlea), which lodges a tubular prolongation of the dura mater establishing a communication between the perilymphatic space and the subarachnoid space, and transmits a vein from the cochlea to join the internal jugular vein.
Additional images
Document 3:::
The ampullary cupula, or cupula, is a structure in the vestibular system, providing the sense of spatial orientation.
The cupula is located within the ampullae of each of the three semicircular canals. Part of the crista ampullaris, the cupula has embedded within it hair cells that have several stereocilia associated with each kinocilium. The cupula itself is the gelatinous component of the crista ampullaris that extends from the crista to the roof of the ampullae. When the head rotates, the endolymph filling the semicircular ducts initially lags behind due to inertia. As a result, the cupula is deflected opposite the direction of head movement. As the endolymph pushes the cupula, the stereocilia is bent as well, stimulating the hair cells within the crista ampullaris. After a short time of continual rotation however, the endolymph's acceleration normalizes with the rate of rotation of the semicircular ducts. As a result, the cupula returns to its resting position and the hair cells cease to be stimulated. This continues until the head stops rotating which simultaneously halts semicircular duct rotation. Due to inertia, however, the endolymph continues on. As the endolymph continues to move, the cupula is once again deflected resulting in the compensatory movements of the body when spun. In only the first situation, as fluid rushes by the cupula, the hair cells stimulated transmit the corresponding signal to the brain through the vestibulocochlear nerve (CN VIII). In the second one, there is no stimulation as the kinocilium can only be bent in one direction.
In their natural orientation within the head, the cupulae are located on the medial aspect of the semicircular canals. In this orientation, the kinocilia rest on the posterior aspect of the cupula.
Effects of alcohol
The Buoyancy Hypothesis posits that alcohol causes vertigo by affecting the neutral buoyancy of the cupula within the surrounding fluid called the endolymph. Linear accelerations (such as that
Document 4:::
The leans is the most common type of spatial disorientation for aviators. Through stabilization of the fluid in the semicircular canals, a pilot may perceive straight and level flight while actually in a banked turn. This is caused by a quick return to level flight after a gradual, prolonged turn that the pilot failed to notice. The phenomenon consists of a false perception of angular displacement about the roll axis and therefore becomes an illusion of bank. This illusion is often associated with a vestibulospinal reflex that results in the pilot actually leaning in the direction of the falsely perceived vertical. Other common explanations of the leans are due to deficiencies of both otolith-organ and semicircular-duct sensory mechanisms.
Physiology
The leans is a type of vestibular illusion in flight which causes spatial disorientation. The process involves the semicircular canals of the vestibular system. The semicircular canals detect angular acceleration. In total, there are three semicircular canals: the anterior, posterior, and lateral canals. Each canal is filled with a fluid called endolymph and each canal arises from a small bag-like structure called a utricle. At the ends of each duct, there is a saclike portion called the ampulla. Inside are hair cells and supporting cells known as the crista ampullaris.
Changing a person's orientation will cause specific ducts to be stimulated due to these hair cells. When the head turns, the canals move but because of its inertia, the endolymph fluid tends to lag and thereby stimulates the hair cells. This stimulation results in awareness of angular acceleration in that plane. After approximately 10–20 seconds, the endolymph velocity matches that of the canal, which stops stimulation of the hair cells, and the person's awareness of rotation is reduced or stopped. In addition, the canals cannot detect rotational acceleration of approximately 2 degrees per second or lower; this is the detection threshold of the semici
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Involved in balance, the semicircular canals are parts of what organs?
A. eyes
B. ears
C. lungs
D. ovaries
Answer:
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.