id
stringlengths 6
15
| question_type
stringclasses 1
value | question
stringlengths 15
683
| choices
listlengths 4
4
| answer
stringclasses 5
values | explanation
stringclasses 481
values | prompt
stringlengths 1.75k
10.9k
|
---|---|---|---|---|---|---|
sciq-11514
|
multiple_choice
|
What is the leading cause of lung cancer?
|
[
"chewing tobacco",
"tuberculosis",
"heredity",
"tobacco smoke"
] |
D
|
Relavent Documents:
Document 0:::
Margaret L. Kripke is an American immunologist. She is an expert in photoimmunology and the immunology of skin cancers. She earned a BS and MS in bacteriology, and a Ph.D in immunology, at the University of California at Berkeley.
She founded the department of immunology at The University of Texas M. D. Anderson Cancer Center in 1983, and served as the cancer center's executive vice president and chief academic officer until her retirement in 2007. After her retirement, Kripke served as special advisor to the provost.
From 1993 to 1994, Kripke served as president of the American Association for Cancer Research.
In 2008, M. D. Anderson established the Margaret Kripke Legend Award "to honor individuals who have enhanced the careers of women in cancer medicine and cancer science".
She served on the President's Cancer Panel from 2003 to 2011. The panel's 2006-2007 report, Promoting Healthy Lifestyles, urged "that the influence of the tobacco industry – particularly on America’s children – be weakened
through strict Federal regulation of tobacco product sales and marketing". The panel's 2008-2009 report, Reducing Environmental Cancer Risk: What We Can Do Now, "for the first time highlights the contribution of environmental contaminants to the development of cancer". A 2021 video describes how Dr. Kripke came to rethink her assumptions about the causes of cancer.
In 2013, she was named a Fellow of the American Association for Cancer Research Academy.
From 2012 through 2016, she was the chief scientific officer of the Cancer Prevention and Research Institute of Texas.
She has served on the board of directors of Silent Spring Institute.
In 2020, Kripke called upon the National Cancer Institute to publish information about cancer risks from exposure to chemicals in the environment.
Bibliography
Publication Lists
JSTOR.org
Researchgate.net
Google Scholar
Books
Google Books list
Document 1:::
Cancer is caused by genetic changes leading to uncontrolled cell growth and tumor formation. The basic cause of sporadic (non-familial) cancers is DNA damage and genomic instability. A minority of cancers are due to inherited genetic mutations. Most cancers are related to environmental, lifestyle, or behavioral exposures. Cancer is generally not contagious in humans, though it can be caused by oncoviruses and cancer bacteria. The term "environmental", as used by cancer researchers, refers to everything outside the body that interacts with humans. The environment is not limited to the biophysical environment (e.g. exposure to factors such as air pollution or sunlight), but also includes lifestyle and behavioral factors.
Over one third of cancer deaths worldwide (and about 75–80% in the United States) are potentially avoidable by reducing exposure to known factors. Common environmental factors that contribute to cancer death include exposure to different chemical and physical agents (tobacco use accounts for 25–30% of cancer deaths), environmental pollutants, diet and obesity (30–35%), infections (15–20%), and radiation (both ionizing and non-ionizing, up to 10%). These factors act, at least partly, by altering the function of genes within cells. Typically many such genetic changes are required before cancer develops. Aging has been repeatedly and consistently regarded as an important aspect to consider when evaluating the risk factors for the development of particular cancers. Many molecular and cellular changes involved in the development of cancer accumulate during the aging process and eventually manifest as cancer.
Genetics
Although there are over 50 identifiable hereditary forms of cancer, less than 0.3% of the population are carriers of a cancer-related genetic mutation and these make up less than 3–10% of all cancer cases. The vast majority of cancers are non-hereditary ("sporadic cancers"). Hereditary cancers are primarily caused by an inherited genetic d
Document 2:::
Single Best Answer (SBA or One Best Answer) is a written examination form of multiple choice questions used extensively in medical education.
Structure
A single question is posed with typically five alternate answers, from which the candidate must choose the best answer. This method avoids the problems of past examinations of a similar form described as Single Correct Answer. The older form can produce confusion where more than one of the possible answers has some validity. The newer form makes it explicit that more than one answer may have elements that are correct, but that one answer will be superior.
Prior to the widespread introduction of SBAs into medical education, the typical form of examination was true-false multiple choice questions. But during the 2000s, educators found that SBAs would be superior.
Document 3:::
The Science, Technology, Engineering and Mathematics Network or STEMNET is an educational charity in the United Kingdom that seeks to encourage participation at school and college in science and engineering-related subjects (science, technology, engineering, and mathematics) and (eventually) work.
History
It is based at Woolgate Exchange near Moorgate tube station in London and was established in 1996. The chief executive is Kirsten Bodley. The STEMNET offices are housed within the Engineering Council.
Function
Its chief aim is to interest children in science, technology, engineering and mathematics. Primary school children can start to have an interest in these subjects, leading secondary school pupils to choose science A levels, which will lead to a science career. It supports the After School Science and Engineering Clubs at schools. There are also nine regional Science Learning Centres.
STEM ambassadors
To promote STEM subjects and encourage young people to take up jobs in these areas, STEMNET have around 30,000 ambassadors across the UK. these come from a wide selection of the STEM industries and include TV personalities like Rob Bell.
Funding
STEMNET used to receive funding from the Department for Education and Skills. Since June 2007, it receives funding from the Department for Children, Schools and Families and Department for Innovation, Universities and Skills, since STEMNET sits on the chronological dividing point (age 16) of both of the new departments.
See also
The WISE Campaign
Engineering and Physical Sciences Research Council
National Centre for Excellence in Teaching Mathematics
Association for Science Education
Glossary of areas of mathematics
Glossary of astronomy
Glossary of biology
Glossary of chemistry
Glossary of engineering
Glossary of physics
Document 4:::
The STEM (Science, Technology, Engineering, and Mathematics) pipeline is a critical infrastructure for fostering the development of future scientists, engineers, and problem solvers. It's the educational and career pathway that guides individuals from early childhood through to advanced research and innovation in STEM-related fields.
Description
The "pipeline" metaphor is based on the idea that having sufficient graduates requires both having sufficient input of students at the beginning of their studies, and retaining these students through completion of their academic program. The STEM pipeline is a key component of workplace diversity and of workforce development that ensures sufficient qualified candidates are available to fill scientific and technical positions.
The STEM pipeline was promoted in the United States from the 1970s onwards, as “the push for STEM (science, technology, engineering, and mathematics) education appears to have grown from a concern for the low number of future professionals to fill STEM jobs and careers and economic and educational competitiveness.”
Today, this metaphor is commonly used to describe retention problems in STEM fields, called “leaks” in the pipeline. For example, the White House reported in 2012 that 80% of minority groups and women who enroll in a STEM field switch to a non-STEM field or drop out during their undergraduate education. These leaks often vary by field, gender, ethnic and racial identity, socioeconomic background, and other factors, drawing attention to structural inequities involved in STEM education and careers.
Current efforts
The STEM pipeline concept is a useful tool for programs aiming at increasing the total number of graduates, and is especially important in efforts to increase the number of underrepresented minorities and women in STEM fields. Using STEM methodology, educational policymakers can examine the quantity and retention of students at all stages of the K–12 educational process and beyo
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is the leading cause of lung cancer?
A. chewing tobacco
B. tuberculosis
C. heredity
D. tobacco smoke
Answer:
|
|
sciq-3823
|
multiple_choice
|
Ion drives have low thrust but high what?
|
[
"tolerance",
"acceleration",
"efficiency",
"power"
] |
C
|
Relavent Documents:
Document 0:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 1:::
Engineering mathematics is a branch of applied mathematics concerning mathematical methods and techniques that are typically used in engineering and industry. Along with fields like engineering physics and engineering geology, both of which may belong in the wider category engineering science, engineering mathematics is an interdisciplinary subject motivated by engineers' needs both for practical, theoretical and other considerations outside their specialization, and to deal with constraints to be effective in their work.
Description
Historically, engineering mathematics consisted mostly of applied analysis, most notably: differential equations; real and complex analysis (including vector and tensor analysis); approximation theory (broadly construed, to include asymptotic, variational, and perturbative methods, representations, numerical analysis); Fourier analysis; potential theory; as well as linear algebra and applied probability, outside of analysis. These areas of mathematics were intimately tied to the development of Newtonian physics, and the mathematical physics of that period. This history also left a legacy: until the early 20th century subjects such as classical mechanics were often taught in applied mathematics departments at American universities, and fluid mechanics may still be taught in (applied) mathematics as well as engineering departments.
The success of modern numerical computer methods and software has led to the emergence of computational mathematics, computational science, and computational engineering (the last two are sometimes lumped together and abbreviated as CS&E), which occasionally use high-performance computing for the simulation of phenomena and the solution of problems in the sciences and engineering. These are often considered interdisciplinary fields, but are also of interest to engineering mathematics.
Specialized branches include engineering optimization and engineering statistics.
Engineering mathematics in tertiary educ
Document 2:::
Advanced Placement (AP) Physics C: Electricity and Magnetism (also known as AP Physics C: E&M or AP E&M) is an introductory physics course administered by the College Board as part of its Advanced Placement program. It is intended to proxy a second-semester calculus-based university course in electricity and magnetism. The content of Physics C: E&M overlaps with that of AP Physics 2, but Physics 2 is algebra-based and covers other topics outside of electromagnetism, while Physics C is calculus-based and only covers electromagnetism. Physics C: E&M may be combined with its mechanics counterpart to form a year-long course that prepares for both exams.
Course content
E&M is equivalent to an introductory college course in electricity and magnetism for physics or engineering majors. The course modules are:
Electrostatics
Conductors, capacitors, and dielectrics
Electric circuits
Magnetic fields
Electromagnetism.
Methods of calculus are used wherever appropriate in formulating physical principles and in applying them to physical problems. Therefore, students should have completed or be concurrently enrolled in a calculus class.
AP test
The course culminates in an optional exam for which high-performing students may receive some credit towards their college coursework, depending on the institution.
Registration
The AP examination for AP Physics C: Electricity and Magnetism is separate from the AP examination for AP Physics C: Mechanics. Before 2006, test-takers paid only once and were given the choice of taking either one or two parts of the Physics C test.
Format
The exam is typically administered on a Monday afternoon in May. The exam is configured in two categories: a 35-question multiple choice section and a 3-question free response section. Test takers are allowed to use an approved calculator during the entire exam. The test is weighted such that each section is worth half of the final score. This and AP Physics C: Mechanics are the shortest AP exams, with
Document 3:::
Female education in STEM refers to child and adult female representation in the educational fields of science, technology, engineering, and mathematics (STEM). In 2017, 33% of students in STEM fields were women.
The organization UNESCO has stated that this gender disparity is due to discrimination, biases, social norms and expectations that influence the quality of education women receive and the subjects they study. UNESCO also believes that having more women in STEM fields is desirable because it would help bring about sustainable development.
Current status of girls and women in STEM education
Overall trends in STEM education
Gender differences in STEM education participation are already visible in early childhood care and education in science- and math-related play, and become more pronounced at higher levels of education. Girls appear to lose interest in STEM subjects with age, particularly between early and late adolescence. This decreased interest affects participation in advanced studies at the secondary level and in higher education. Female students represent 35% of all students enrolled in STEM-related fields of study at this level globally. Differences are also observed by disciplines, with female enrollment lowest in engineering, manufacturing and construction, natural science, mathematics and statistics and ICT fields. Significant regional and country differences in female representation in STEM studies can be observed, though, suggesting the presence of contextual factors affecting girls’ and women's engagement in these fields. Women leave STEM disciplines in disproportionate numbers during their higher education studies, in their transition to the world of work and even in their career cycle.
Learning achievement in STEM education
Data on gender differences in learning achievement present a complex picture, depending on what is measured (subject, knowledge acquisition against knowledge application), the level of education/age of students, and
Document 4:::
Computer science and engineering (CSE) is an academic program at many universities which comprises computer science classes (e.g. data structures and algorithms) and computer engineering classes (e.g computer architecture). There is no clear division in computing between science and engineering, just like in the field of materials science and engineering. CSE is also a term often used in Europe to translate the name of engineering informatics academic programs. It is offered in both undergraduate as well postgraduate with specializations.
Academic courses
Academic programs vary between colleges, but typically include a combination of topics in computer science, computer engineering, and electrical engineering. Undergraduate courses usually include programming, algorithms and data structures, computer architecture, operating systems, computer networks, parallel computing, embedded systems, algorithms design, circuit analysis and electronics, digital logic and processor design, computer graphics, scientific computing, software engineering, database systems, digital signal processing, virtualization, computer simulations and games programming. CSE programs also include core subjects of theoretical computer science such as theory of computation, numerical methods, machine learning, programming theory and paradigms. Modern academic programs also cover emerging computing fields like image processing, data science, robotics, bio-inspired computing, computational biology, autonomic computing and artificial intelligence. Most CSE programs require introductory mathematical knowledge, hence the first year of study is dominated by mathematical courses, primarily discrete mathematics, mathematical analysis, linear algebra, probability, and statistics, as well as the basics of electrical and electronic engineering, physics, and electromagnetism.
Example universities with CSE majors and departments
APJ Abdul Kalam Technological University
American International University-B
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Ion drives have low thrust but high what?
A. tolerance
B. acceleration
C. efficiency
D. power
Answer:
|
|
sciq-8973
|
multiple_choice
|
The tensor fascia latae acts as what, in relation to the gluteus medius and iliopsoas, for the purpose of flexing and abducting the thigh?
|
[
"synergist",
"antagonist",
"spicule",
"symbiotic"
] |
A
|
Relavent Documents:
Document 0:::
The latissimus dorsi () is a large, flat muscle on the back that stretches to the sides, behind the arm, and is partly covered by the trapezius on the back near the midline. The word latissimus dorsi (plural: latissimi dorsi) comes from Latin and means "broadest [muscle] of the back", from "latissimus" ()' and "dorsum" (). The pair of muscles are commonly known as "lats", especially among bodybuilders. The latissimus dorsi is the largest muscle in the upper body.
The latissimus dorsi is responsible for extension, adduction, transverse extension also known as horizontal abduction (or horizontal extension), flexion from an extended position, and (medial) internal rotation of the shoulder joint. It also has a synergistic role in extension and lateral flexion of the lumbar spine.
Due to bypassing the scapulothoracic joints and attaching directly to the spine, the actions the latissimi dorsi have on moving the arms can also influence the movement of the scapulae, such as their downward rotation during a pull up.
Structure
Variations
The number of dorsal vertebrae to which it is attached varies from four to eight; the number of costal attachments varies; muscle fibers may or may not reach the crest of the ilium.
A muscle slip, the axillary arch, varying from 7 to 10 cm in length, and from 5 to 15 mm in breadth, occasionally springs from the upper edge of the latissimus dorsi about the middle of the posterior fold of the axilla, and crosses the axilla in front of the axillary vessels and nerves, to join the under surface of the tendon of the pectoralis major, the coracobrachialis, or the fascia over the biceps brachii. This axillary arch crosses the axillary artery, just above the spot usually selected for the application of a ligature, and may mislead a surgeon. It is present in about 7% of the population and may be easily recognized by the transverse direction of its fibers. Guy et al. extensively described this muscular variant using MRI data and positively corre
Document 1:::
The rectus femoris muscle is one of the four quadriceps muscles of the human body. The others are the vastus medialis, the vastus intermedius (deep to the rectus femoris), and the vastus lateralis. All four parts of the quadriceps muscle attach to the patella (knee cap) by the quadriceps tendon.
The rectus femoris is situated in the middle of the front of the thigh; it is fusiform in shape, and its superficial fibers are arranged in a bipenniform manner, the deep fibers running straight () down to the deep aponeurosis. Its functions are to flex the thigh at the hip joint and to extend the leg at the knee joint.
Structure
It arises by two tendons: one, the anterior or straight, from the anterior inferior iliac spine; the other, the posterior or reflected, from a groove above the rim of the acetabulum.
The two unite at an acute angle and spread into an aponeurosis that is prolonged downward on the anterior surface of the muscle, and from this the muscular fibers arise.
The muscle ends in a broad and thick aponeurosis that occupies the lower two-thirds of its posterior surface, and, gradually becoming narrowed into a flattened tendon, is inserted into the base of the patella.
Nerve supply
The neurons for voluntary thigh contraction originate near the summit of the medial side of the precentral gyrus (the primary motor area of the brain). These neurons send a nerve signal that is carried by the corticospinal tract down the brainstem and spinal cord. The signal starts with the upper motor neurons carrying the signal from the precentral gyrus down through the internal capsule, through the cerebral peduncle, and into the medulla. In the medullary pyramid, the corticospinal tract decussates and becomes the lateral corticospinal tract. The nerve signal will continue down the lateral corticospinal tract until it reaches spinal nerve L4. At this point, the nerve signal will synapse from the upper motor neurons to the lower motor neurons. The signal will travel through the
Document 2:::
The gluteus maximus is the main extensor muscle of the hip in humans. It is the largest and outermost of the three gluteal muscles and makes up a large part of the shape and appearance of each side of the hips. It is the single largest muscle in the human body. Its thick fleshy mass, in a quadrilateral shape, forms the prominence of the buttocks. The other gluteal muscles are the medius and minimus, and sometimes informally these are collectively referred to as the glutes.
Its large size is one of the most characteristic features of the muscular system in humans, connected as it is with the power of maintaining the trunk in the erect posture. Other primates have much flatter hips and cannot sustain standing erectly.
The muscle is made up of muscle fascicles lying parallel with one another, and are collected together into larger bundles separated by fibrous septa.
Structure
The gluteus maximus (or buttock) is the outermost muscle of the buttocks. It arises from connections to nearby structures in this area. It arises from the posterior gluteal line of the inner upper ilium, a bone of the pelvis, as well as above it to the iliac crest and slightly below it; from the lower part of the sacrum and the side of the coccyx, the tailbone; from the aponeurosis of the erector spinae (lumbodorsal fascia), the sacrotuberous ligament, and the fascia covering the gluteus medius (gluteal aponeurosis).
The fibers are directed obliquely inferiorly and laterally;
The gluteus maximus ends in two main areas:
those forming the upper and larger portion of the muscle, together with the superficial fibers of the lower portion, end in a thick tendinous lamina, which passes across the greater trochanter, and inserts into the iliotibial band of the fascia lata;
the deeper fibers of the lower portion are inserted into the gluteal tuberosity of the linea aspera, between the vastus lateralis and adductor magnus. If present, the third trochanter also serves as an attachment.
Bursae
Three bu
Document 3:::
The tensor fasciae latae (or tensor fasciæ latæ or, formerly, tensor vaginae femoris) is a muscle of the thigh. Together with the gluteus maximus, it acts on the iliotibial band and is continuous with the iliotibial tract, which attaches to the tibia. The muscle assists in keeping the balance of the pelvis while standing, walking, or running.
Structure
It arises from the anterior part of the outer lip of the iliac crest; from the outer surface of the anterior superior iliac spine, and part of the outer border of the notch below it, between the gluteus medius and sartorius; and from the deep surface of the fascia lata.
It is inserted between the two layers of the iliotibial tract of the fascia lata about the junction of the middle and upper thirds of the thigh. The tensor fasciae latae tautens the iliotibial tract and braces the knee, especially when the opposite foot is lifted. The terminal insertion point lies on the lateral condyle of the tibia.
Nerve supply
Tensor fasciae latae is innervated by the superior gluteal nerve, L5 and S1. At its origins of the anterior rami of L4, L5, and S1 nerves, the superior gluteal nerve exits the pelvis via greater sciatic foramen superior to the piriformis. The nerve also courses between the gluteus medius and minimus. The superior gluteal artery also supplies the tensor fasciae latae. The superior gluteal nerve arises from the sacral plexus and only has muscular innervation associated with it. There is no cutaneous innervation for sensation that stems from the superior gluteal nerve.
Function
The tensor fasciae latae is a tensor of the fascia lata; continuing its action, the oblique direction of its fibers enables it to stabilize the hip in extension (assists gluteus maximus during hip extension). The fascia lata is a fibrous sheath that encircles the thigh like a subcutaneous stocking and tightly binds its muscles. On the lateral surface, it combines with the tendons of the gluteus maximus and tensor fasciae latae to
Document 4:::
The deep fascia of leg, or crural fascia forms a complete investment to the muscles, and is fused with the periosteum over the subcutaneous surfaces of the bones.
The deep fascia of the leg is continuous above with the fascia lata (deep fascia of the thigh), and is attached around the knee to the patella, the patellar ligament, the tuberosity and condyles of the tibia, and the head of the fibula.
Behind, it forms the popliteal fascia, covering in the popliteal fossa; here it is strengthened by transverse fibers, and perforated by the small saphenous vein.
It receives an expansion from the tendon of the biceps femoris laterally, and from the tendons of the sartorius, gracilis, semitendinosus, and semimembranosus medially; in front, it blends with the periosteum covering the subcutaneous surface of the tibia, and with that covering the head and malleolus of the fibula; below, it is continuous with the transverse crural and laciniate ligaments.
It is thick and dense in the upper and anterior part of the leg, and gives attachment, by its deep surface, to the tibialis anterior and extensor digitorum longus; but thinner behind, where it covers the gastrocnemius and soleus.
It gives off from its deep surface, on the lateral side of the leg, two strong intermuscular septa, the anterior and posterior peroneal septa, which enclose the fibularis (peroneus) longus and brevis muscles and separate them from the muscles of the anterior and posterior crural regions, and several more slender processes which enclose the individual muscles in each region.
A broad transverse intermuscular septum, called the deep transverse fascia of the leg, intervenes between the superficial and deep posterior crural muscles.
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
The tensor fascia latae acts as what, in relation to the gluteus medius and iliopsoas, for the purpose of flexing and abducting the thigh?
A. synergist
B. antagonist
C. spicule
D. symbiotic
Answer:
|
|
sciq-1376
|
multiple_choice
|
Alternation of generations is characteristic of the life cycle of all what?
|
[
"cells",
"plants",
"animals",
"organs"
] |
B
|
Relavent Documents:
Document 0:::
Evolutionary biology is the subfield of biology that studies the evolutionary processes (natural selection, common descent, speciation) that produced the diversity of life on Earth. It is also defined as the study of the history of life forms on Earth. Evolution holds that all species are related and gradually change over generations. In a population, the genetic variations affect the phenotypes (physical characteristics) of an organism. These changes in the phenotypes will be an advantage to some organisms, which will then be passed onto their offspring. Some examples of evolution in species over many generations are the peppered moth and flightless birds. In the 1930s, the discipline of evolutionary biology emerged through what Julian Huxley called the modern synthesis of understanding, from previously unrelated fields of biological research, such as genetics and ecology, systematics, and paleontology.
The investigational range of current research has widened to encompass the genetic architecture of adaptation, molecular evolution, and the different forces that contribute to evolution, such as sexual selection, genetic drift, and biogeography. Moreover, the newer field of evolutionary developmental biology ("evo-devo") investigates how embryogenesis is controlled, thus yielding a wider synthesis that integrates developmental biology with the fields of study covered by the earlier evolutionary synthesis.
Subfields
Evolution is the central unifying concept in biology. Biology can be divided into various ways. One way is by the level of biological organization, from molecular to cell, organism to population. Another way is by perceived taxonomic group, with fields such as zoology, botany, and microbiology, reflecting what was once seen as the major divisions of life. A third way is by approaches, such as field biology, theoretical biology, experimental evolution, and paleontology. These alternative ways of dividing up the subject have been combined with evolution
Document 1:::
A cell type is a classification used to identify cells that share morphological or phenotypical features. A multicellular organism may contain cells of a number of widely differing and specialized cell types, such as muscle cells and skin cells, that differ both in appearance and function yet have identical genomic sequences. Cells may have the same genotype, but belong to different cell types due to the differential regulation of the genes they contain. Classification of a specific cell type is often done through the use of microscopy (such as those from the cluster of differentiation family that are commonly used for this purpose in immunology). Recent developments in single cell RNA sequencing facilitated classification of cell types based on shared gene expression patterns. This has led to the discovery of many new cell types in e.g. mouse cortex, hippocampus, dorsal root ganglion and spinal cord.
Animals have evolved a greater diversity of cell types in a multicellular body (100–150 different cell types), compared
with 10–20 in plants, fungi, and protists. The exact number of cell types is, however, undefined, and the Cell Ontology, as of 2021, lists over 2,300 different cell types.
Multicellular organisms
All higher multicellular organisms contain cells specialised for different functions. Most distinct cell types arise from a single totipotent cell that differentiates into hundreds of different cell types during the course of development. Differentiation of cells is driven by different environmental cues (such as cell–cell interaction) and intrinsic differences (such as those caused by the uneven distribution of molecules during division). Multicellular organisms are composed of cells that fall into two fundamental types: germ cells and somatic cells. During development, somatic cells will become more specialized and form the three primary germ layers: ectoderm, mesoderm, and endoderm. After formation of the three germ layers, cells will continue to special
Document 2:::
This glossary of developmental biology is a list of definitions of terms and concepts commonly used in the study of developmental biology and related disciplines in biology, including embryology and reproductive biology, primarily as they pertain to vertebrate animals and particularly to humans and other mammals. The developmental biology of invertebrates, plants, fungi, and other organisms is treated in other articles; e.g. terms relating to the reproduction and development of insects are listed in Glossary of entomology, and those relating to plants are listed in Glossary of botany.
This glossary is intended as introductory material for novices; for more specific and technical detail, see the article corresponding to each term. Additional terms relevant to vertebrate reproduction and development may also be found in Glossary of biology, Glossary of cell biology, Glossary of genetics, and Glossary of evolutionary biology.
A
B
C
D
E
F
G
H
I
J
K
L
M
N
O
P
Q
R
S
T
U
V
W
X
Y
Z
See also
Introduction to developmental biology
Outline of developmental biology
Outline of cell biology
Glossary of biology
Glossary of cell biology
Glossary of genetics
Glossary of evolutionary biology
Document 3:::
In biology, evolution is the process of change in all forms of life over generations, and evolutionary biology is the study of how evolution occurs. Biological populations evolve through genetic changes that correspond to changes in the organisms' observable traits. Genetic changes include mutations, which are caused by damage or replication errors in organisms' DNA. As the genetic variation of a population drifts randomly over generations, natural selection gradually leads traits to become more or less common based on the relative reproductive success of organisms with those traits.
The age of the Earth is about 4.5 billion years. The earliest undisputed evidence of life on Earth dates from at least 3.5 billion years ago. Evolution does not attempt to explain the origin of life (covered instead by abiogenesis), but it does explain how early lifeforms evolved into the complex ecosystem that we see today. Based on the similarities between all present-day organisms, all life on Earth is assumed to have originated through common descent from a last universal ancestor from which all known species have diverged through the process of evolution.
All individuals have hereditary material in the form of genes received from their parents, which they pass on to any offspring. Among offspring there are variations of genes due to the introduction of new genes via random changes called mutations or via reshuffling of existing genes during sexual reproduction. The offspring differs from the parent in minor random ways. If those differences are helpful, the offspring is more likely to survive and reproduce. This means that more offspring in the next generation will have that helpful difference and individuals will not have equal chances of reproductive success. In this way, traits that result in organisms being better adapted to their living conditions become more common in descendant populations. These differences accumulate resulting in changes within the population. This proce
Document 4:::
Evolution of cells refers to the evolutionary origin and subsequent evolutionary development of cells. Cells first emerged at least 3.8 billion years ago approximately 750 million years after Earth was formed.
The first cells
The initial development of the cell marked the passage from prebiotic chemistry to partitioned units resembling modern cells. The final transition to living entities that fulfill all the definitions of modern cells depended on the ability to evolve effectively by natural selection. This transition has been called the Darwinian transition.
If life is viewed from the point of view of replicator molecules, cells satisfy two fundamental conditions: protection from the outside environment and confinement of biochemical activity. The former condition is needed to keep complex molecules stable in a varying and sometimes aggressive environment; the latter is fundamental for the evolution of biocomplexity. If freely floating molecules that code for enzymes are not enclosed in cells, the enzymes will automatically benefit neighboring replicator molecules as well. Thus, the consequences of diffusion in non-partitioned lifeforms would result in "parasitism by default." Therefore, the selection pressure on replicator molecules will be lower, as the 'lucky' molecule that produces the better enzyme does not fully leverage its advantage over its close neighbors. In contrast, if the molecule is enclosed in a cell membrane, the enzymes coded will be available only to itself. That molecule will uniquely benefit from the enzymes it codes for, increasing individuality and thus accelerating natural selection.
Partitioning may have begun from cell-like spheroids formed by proteinoids, which are observed by heating amino acids with phosphoric acid as a catalyst. They bear much of the basic features provided by cell membranes. Proteinoid-based protocells enclosing RNA molecules could have been the first cellular life forms on Earth.
Another possibility is that the
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Alternation of generations is characteristic of the life cycle of all what?
A. cells
B. plants
C. animals
D. organs
Answer:
|
|
sciq-2833
|
multiple_choice
|
The first plants probably evolved from what?
|
[
"mould",
"dry green algae",
"aquatic green algae",
"moss"
] |
C
|
Relavent Documents:
Document 0:::
Plant life-form schemes constitute a way of classifying plants alternatively to the ordinary species-genus-family scientific classification. In colloquial speech, plants may be classified as trees, shrubs, herbs (forbs and graminoids), etc. The scientific use of life-form schemes emphasizes plant function in the ecosystem and that the same function or "adaptedness" to the environment may be achieved in a number of ways, i.e. plant species that are closely related phylogenetically may have widely different life-form, for example Adoxa moschatellina and Sambucus nigra are from the same family, but the former is a small herbaceous plant and the latter is a shrub or tree. Conversely, unrelated species may share a life-form through convergent evolution.
While taxonomic classification is concerned with the production of natural classifications (being natural understood either in philosophical basis for pre-evolutionary thinking, or phylogenetically as non-polyphyletic), plant life form classifications uses other criteria than naturalness, like morphology, physiology and ecology.
Life-form and growth-form are essentially synonymous concepts, despite attempts to restrict the meaning of growth-form to types differing in shoot architecture. Most life form schemes are concerned with vascular plants only. Plant construction types may be used in a broader sense to encompass planktophytes, benthophytes (mainly algae) and terrestrial plants.
A popular life-form scheme is the Raunkiær system.
History
One of the earliest attempts to classify the life-forms of plants and animals was made by Aristotle, whose writings are lost. His pupil, Theophrastus, in Historia Plantarum (c. 350 BC), was the first who formally recognized plant habits: trees, shrubs and herbs.
Some earlier authors (e.g., Humboldt, 1806) did classify species according to physiognomy, but were explicit about the entities being merely practical classes without any relation to plant function. A marked exception was
Document 1:::
Plants are the eukaryotes that form the kingdom Plantae; they are predominantly photosynthetic. This means that they obtain their energy from sunlight, using chloroplasts derived from endosymbiosis with cyanobacteria to produce sugars from carbon dioxide and water, using the green pigment chlorophyll. Exceptions are parasitic plants that have lost the genes for chlorophyll and photosynthesis, and obtain their energy from other plants or fungi.
Historically, as in Aristotle's biology, the plant kingdom encompassed all living things that were not animals, and included algae and fungi. Definitions have narrowed since then; current definitions exclude the fungi and some of the algae. By the definition used in this article, plants form the clade Viridiplantae (green plants), which consists of the green algae and the embryophytes or land plants (hornworts, liverworts, mosses, lycophytes, ferns, conifers and other gymnosperms, and flowering plants). A definition based on genomes includes the Viridiplantae, along with the red algae and the glaucophytes, in the clade Archaeplastida.
There are about 380,000 known species of plants, of which the majority, some 260,000, produce seeds. They range in size from single cells to the tallest trees. Green plants provide a substantial proportion of the world's molecular oxygen; the sugars they create supply the energy for most of Earth's ecosystems; other organisms, including animals, either consume plants directly or rely on organisms which do so.
Grain, fruit, and vegetables are basic human foods and have been domesticated for millennia. People use plants for many purposes, such as building materials, ornaments, writing materials, and, in great variety, for medicines. The scientific study of plants is known as botany, a branch of biology.
Definition
Taxonomic history
All living things were traditionally placed into one of two groups, plants and animals. This classification dates from Aristotle (384–322 BC), who distinguished d
Document 2:::
The history of phycology is the history of the scientific study of algae. Human interest in plants as food goes back into the origins of the species, and knowledge of algae can be traced back more than two thousand years. However, only in the last three hundred years has that knowledge evolved into a rapidly developing science.
Early days
The study of botany goes back into pre-history, as plants have been eaten since the beginning of the human race. The first attempts at plant cultivation are believed to have been made shortly before 10,000 BC in Western Asia (Morton, 1981) and the first references to algae are to be found in early Chinese literature. Records as far back as 3000 BC indicate that algae were used by the emperor of China as food (Huisman, 2000 p. 13). The use of Porphyra in China dates back to at least AD 533–544 (Mumfard and Miura, 1988); there are also references in Roman and Greek literature. The Greek word for algae was phycos whilst in Latin the name became fucus. There are early references to the use of algae for manure. The first coralline algae to be recognized as living organisms were probably Corallina, by Pliny the Elder in the 1st century AD (Irvine and Chamberlain, 1994 p. 11).
The classification of plants suffered many changes since Theophrastus (372–287 BC) and Aristotle (384–322 BC) grouped them as "trees", "shrubs" and "herbs" (Smith, 1955 p. 1).
Little is known of botany during the Middle Ages — it was the Dark Ages of botany.
The development of the study of phycology runs in a pattern comparable with, and parallel to, other biological fields but at a different rate. After the invention of the printing-press in the 15th century education enabled people to read and knowledge to spread.
Exploration of the world and the advance of knowledge
Written accounts of the algae of South Africa were made by the Portuguese explorers of the 15th and 16th centuries; however it is not clear to which species they refer. (Huisman, 2000 p. 7)
17
Document 3:::
The evolution of plants has resulted in a wide range of complexity, from the earliest algal mats of unicellular archaeplastids evolved through endosymbiosis, through multicellular marine and freshwater green algae, to spore-bearing terrestrial bryophytes, lycopods and ferns, and eventually to the complex seed-bearing gymnosperms and angiosperms (flowering plants) of today. While many of the earliest groups continue to thrive, as exemplified by red and green algae in marine environments, more recently derived groups have displaced previously ecologically dominant ones; for example, the ascendance of flowering plants over gymnosperms in terrestrial environments.
There is evidence that cyanobacteria and multicellular photosynthetic eukaryotes lived in freshwater communities on land as early as 1 billion years ago, and that communities of complex, multicellular photosynthesizing organisms existed on land in the late Precambrian, around .
Evidence of the emergence of embryophyte land plants first occurs in the mid-Ordovician (~), and by the middle of the Devonian (~), many of the features recognised in land plants today were present, including roots and leaves. By the late Devonian (~) some free-sporing plants such as Archaeopteris had secondary vascular tissue that produced wood and had formed forests of tall trees. Also by the late Devonian, Elkinsia, an early seed fern, had evolved seeds.
Evolutionary innovation continued throughout the rest of the Phanerozoic eon and still continues today. Most plant groups were relatively unscathed by the Permo-Triassic extinction event, although the structures of communities changed. This may have set the scene for the appearance of the flowering plants in the Triassic (~), and their later diversification in the Cretaceous and Paleogene. The latest major group of plants to evolve were the grasses, which became important in the mid-Paleogene, from around . The grasses, as well as many other groups, evolved new mechanisms of metab
Document 4:::
Douglas Houghton Campbell (December 19, 1859 – February 24, 1953) was an American botanist and university professor. He was one of the 15 founding professors at Stanford University. His death was described as "the end of an era of a group of great plant morphologists."
Campbell was born and raised in Detroit, Michigan. His father, James V. Campbell, was a member of the Supreme Court of the state of Michigan and a law professor at the University of Michigan. Douglas Campbell graduated from Detroit High School in 1878, going on to study at the University of Michigan. He studied botany, learning new microscopy techniques, and becoming interested in cryptogrammic (deciduous) ferns. He received his master's degree in 1882, and taught botany at Detroit High School while he completed his PhD research. He received his PhD in 1886, then travelled to Germany to learn more microscopy techniques. He developed a technique to embed plant material in paraffin to make fine cross-sections; he was one of the first if not the first to study plant specimens using this technique, which had been newly developed by zoologists. He was also a pioneer in the study of microscopic specimens using vital stains.
When Campbell returned to the United States he took up a professorship at Indiana University (1888 to 1891), writing the textbook Elements of Structural and Systematic Botany. In 1891 he became the founding head of the botany department at Stanford University and remained at Stanford for the remainder of his career, retiring in 1925. He studied mosses and liverworts, producing The Structure and Development of Mosses and Ferns in 1895. This book, together with its subsequent editions in 1905 and 1918, became the authoritative work on the subject and "firmly established Campbell's reputation as one of the leading botanists of the United States." His Lectures on the Evolution of Plants was published in 1899, and became widely used as a botany textbook. University Textbook of Botany was
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
The first plants probably evolved from what?
A. mould
B. dry green algae
C. aquatic green algae
D. moss
Answer:
|
|
sciq-7363
|
multiple_choice
|
Why do some dry fruits split open at maturity?
|
[
"to release toxins",
"to release berries",
"to release seeds",
"to regenerate"
] |
C
|
Relavent Documents:
Document 0:::
Generally, fleshy fruits can be divided into two groups based on the presence or absence of a respiratory increase at the onset of ripening. This respiratory increase—which is preceded, or accompanied, by a rise in ethylene—is called a climacteric, and there are marked differences in the development of climacteric and non-climacteric fruits. Climacteric fruit can be either monocots or dicots and the ripening of these fruits can still be achieved even if the fruit has been harvested at the end of their growth period (prior to ripening on the parent plant). Non-climacteric fruits ripen without ethylene and respiration bursts, the ripening process is slower, and for the most part they will not be able to ripen if the fruit is not attached to the parent plant. Examples of climacteric fruits include apples, bananas, melons, apricots, tomatoes, as well as most stone fruits. Non-climacteric fruits on the other hand include citrus fruits, grapes, and strawberries (However, non-climacteric melons and apricots do exist, and grapes and strawberries harbor several active ethylene receptors.) Essentially, a key difference between climacteric and non-climacteric fruits (particularly for commercial production) is that climacteric fruits continue to ripen following their harvest, whereas non-climacteric fruits do not. The accumulation of starch over the early stages of climacteric fruit development may be a key issue, as starch can be converted to sugars after harvest.
Overview
The climacteric stage of fruit ripening is associated with increased ethylene production and a rise in cellular respiration and is the final physiological process that marks the end of fruit maturation and the beginning of fruit senescence. Its defining point is a sudden rise in respiration of the fruit, and normally takes place without any external influences. After the climacteric period, respiration rates (noted by carbon dioxide production) return to or dip below the pre-climacteric rates. The climacte
Document 1:::
In agriculture, shattering is the dispersal of a crop's seeds upon their becoming ripe. From an agricultural perspective this is generally an undesirable process, and in the history of crop domestication several important advances have involved a mutation in a crop plant that reduced shattering—instead of the seeds being dispersed as soon as they were ripe, the mutant plants retained the seeds for longer, which made harvesting much more effective. Non-shattering phenotype is one of the prerequisites for plant breeding especially when introgressing valuable traits from wild varieties of domesticated crops.
A particularly important mutation that was selected very early in the history of agriculture removed the "brittle rachis" problem from wheat. A ripe head ("ear") of wild-type wheat is easily shattered into dispersal units when touched, or blown by the wind, because during ripening a series of abscission layers forms that divides the rachis into short segments, each attached to a single spikelet (which contains 2–3 grains along with chaff).
A different class of shattering mechanisms involves dehiscence of the mature fruit, which releases the seeds.
Current research priorities to understand the genetics of shattering include the following crops:
Barley
Buckwheat
Grain Amaranth
Oilseed rape (Brassica napus)
Sesame and rapeseed are harvested before the seed is fully mature, so that the pods do not split and drop the seeds.
Document 2:::
A seedless fruit is a fruit developed to possess no mature seeds. Since eating seedless fruits is generally easier and more convenient, they are considered commercially valuable.
Most commercially produced seedless fruits have been developed from plants whose fruits normally contain numerous relatively large hard seeds distributed throughout the flesh of the fruit.
Varieties
Common varieties of seedless fruits include watermelons, tomatoes, and grapes (such as Termarina rossa). Additionally, there are numerous seedless citrus fruits, such as oranges, lemons and limes.
A recent development over the last twenty years has been that of seedless sweet peppers (Capsicum annuum). The seedless plant combines male sterility in the pepper plant (commonly occurring) with the ability to set seedless fruits (a natural fruit-setting without fertilization). In male sterile plants, the parthenocarpy expresses itself only sporadically on the plant with deformed fruits. It has been reported that plant hormones provided by the ovary seed (such as auxins and gibberellins) promote fruit set and growth to produce seedless fruits. Initially, without seeds in the fruit, vegetative propagation was essential. However, now – as with seedless watermelon – seedless peppers can be grown from seeds.
Biological description
Seedless fruits can develop in one of two ways: either the fruit develops without fertilization (parthenocarpy), or pollination triggers fruit development, but the ovules or embryos abort without producing mature seeds (stenospermocarpy). Seedless banana and watermelon fruits are produced on triploid plants, whose three sets of chromosomes make it very unlikely for meiosis to successfully produce spores and gametophytes. This is because one of the three copies of each chromosome cannot pair with another appropriate chromosome before separating into daughter cells, so these extra third copies end up randomly distributed between the two daughter cells from meiosis 1, resul
Document 3:::
Fruit tree propagation is usually carried out vegetatively (non-sexually) by grafting or budding a desired variety onto a suitable rootstock.
Perennial plants can be propagated either by sexual or vegetative means. Sexual reproduction begins when a male germ cell (pollen) from one flower fertilises a female germ cell (ovule, incipient seed) of the same species, initiating the development of a fruit containing seeds. Each seed, when germinated, can grow to become a new specimen tree. However, the new tree inherits characteristics of both its parents, and it will not grow true to the variety of either parent from which it came. That is, it will be a fresh individual with an unpredictable combination of characteristics of its own. Although this is desirable in terms of producing novel combinations from the richness of the gene pool of the two parent plants (such sexual recombination is the source of new cultivars), only rarely will the resulting new fruit tree be directly useful or attractive to the tastes of humankind. Most new plants will have characteristics that lie somewhere between those of the two parents.
Therefore, from the orchard grower or gardener's point of view, it is preferable to propagate fruit cultivars vegetatively in order to ensure reliability. This involves taking a cutting (or scion) of wood from a desirable parent tree which is then grown on to produce a new plant or "clone" of the original. In effect this means that the original Bramley apple tree, for example, was a successful variety grown from a pip, but that every Bramley since then has been propagated by taking cuttings of living matter from that tree, or one of its descendants.
Methods
The simplest method of propagating a tree vegetatively is rooting or taking cuttings. A cutting (usually a piece of stem of the parent plant) is cut off and stuck into soil. Artificial rooting hormones are sometimes used to improve chances of success. If the cutting does not die from rot-inducing fungi o
Document 4:::
The raspberry is the edible fruit of a multitude of plant species in the genus Rubus of the rose family, most of which are in the subgenus Idaeobatus. The name also applies to these plants themselves. Raspberries are perennial with woody stems. World production of raspberries in 2020 was 895,771 tonnes, led by Russia with 20% of the total.
Description
A raspberry is an aggregate fruit, developing from the numerous distinct carpels of a single flower. What distinguishes the raspberry from its blackberry relatives is whether or not the torus (receptacle or stem) "picks with" (i.e., stays with) the fruit. When picking a blackberry fruit, the torus stays with the fruit. With a raspberry, the torus remains on the plant, leaving a hollow core in the raspberry fruit.
Raspberries are grown for the fresh fruit market and for commercial processing into individually quick frozen (IQF) fruit, purée, juice, or as dried fruit used in a variety of grocery products such as raspberry pie. Raspberries need ample sun and water for optimal development. Raspberries thrive in well-drained soil with a pH between 6 and 7 with ample organic matter to assist in retaining water. While moisture is essential, wet and heavy soils or excess irrigation can bring on Phytophthora root rot, which is one of the most serious pest problems facing the red raspberry. As a cultivated plant in moist, temperate regions, it is easy to grow and has a tendency to spread unless pruned. Escaped raspberries frequently appear as garden weeds, spread by seeds found in bird droppings.
An individual raspberry weighs , and is made up of around 100 drupelets, each of which consists of a juicy pulp and a single central seed. A raspberry bush can yield several hundred berries a year.
Etymology
Raspberry derives its name from raspise, "a sweet rose-colored wine" (mid-15th century), from the Anglo-Latin vinum raspeys, or from raspoie, meaning "thicket", of Germanic origin. The name may have been influenced by its appe
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Why do some dry fruits split open at maturity?
A. to release toxins
B. to release berries
C. to release seeds
D. to regenerate
Answer:
|
|
sciq-2902
|
multiple_choice
|
What is the most common type of cancer in adult males?
|
[
"prostate cancer",
"lung cancer",
"liver cancer",
"stomach cancer"
] |
A
|
Relavent Documents:
Document 0:::
Single Best Answer (SBA or One Best Answer) is a written examination form of multiple choice questions used extensively in medical education.
Structure
A single question is posed with typically five alternate answers, from which the candidate must choose the best answer. This method avoids the problems of past examinations of a similar form described as Single Correct Answer. The older form can produce confusion where more than one of the possible answers has some validity. The newer form makes it explicit that more than one answer may have elements that are correct, but that one answer will be superior.
Prior to the widespread introduction of SBAs into medical education, the typical form of examination was true-false multiple choice questions. But during the 2000s, educators found that SBAs would be superior.
Document 1:::
The SAT Subject Test in Biology was the name of a one-hour multiple choice test given on biology by the College Board. A student chose whether to take the test depending upon college entrance requirements for the schools in which the student is planning to apply. Until 1994, the SAT Subject Tests were known as Achievement Tests; and from 1995 until January 2005, they were known as SAT IIs. Of all SAT subject tests, the Biology E/M test was the only SAT II that allowed the test taker a choice between the ecological or molecular tests. A set of 60 questions was taken by all test takers for Biology and a choice of 20 questions was allowed between either the E or M tests. This test was graded on a scale between 200 and 800. The average for Molecular is 630 while Ecological is 591.
On January 19 2021, the College Board discontinued all SAT Subject tests, including the SAT Subject Test in Biology E/M. This was effective immediately in the United States, and the tests were to be phased out by the following summer for international students. This was done as a response to changes in college admissions due to the impact of the COVID-19 pandemic on education.
Format
This test had 80 multiple-choice questions that were to be answered in one hour. All questions had five answer choices. Students received one point for each correct answer, lost ¼ of a point for each incorrect answer, and received 0 points for questions left blank. The student's score was based entirely on his or her performance in answering the multiple-choice questions.
The questions covered a broad range of topics in general biology. There were more specific questions related respectively on ecological concepts (such as population studies and general Ecology) on the E test and molecular concepts such as DNA structure, translation, and biochemistry on the M test.
Preparation
The College Board suggested a year-long course in biology at the college preparatory level, as well as a one-year course in algebra, a
Document 2:::
Several universities have designed interdisciplinary courses with a focus on human biology at the undergraduate level. There is a wide variation in emphasis ranging from business, social studies, public policy, healthcare and pharmaceutical research.
Americas
Human Biology major at Stanford University, Palo Alto (since 1970)
Stanford's Human Biology Program is an undergraduate major; it integrates the natural and social sciences in the study of human beings. It is interdisciplinary and policy-oriented and was founded in 1970 by a group of Stanford faculty (Professors Dornbusch, Ehrlich, Hamburg, Hastorf, Kennedy, Kretchmer, Lederberg, and Pittendrigh). It is a very popular major and alumni have gone to post-graduate education, medical school, law, business and government.
Human and Social Biology (Caribbean)
Human and Social Biology is a Level 4 & 5 subject in the secondary and post-secondary schools in the Caribbean and is optional for the Caribbean Secondary Education Certification (CSEC) which is equivalent to Ordinary Level (O-Level) under the British school system. The syllabus centers on structure and functioning (anatomy, physiology, biochemistry) of human body and the relevance to human health with Caribbean-specific experience. The syllabus is organized under five main sections: Living organisms and the environment, life processes, heredity and variation, disease and its impact on humans, the impact of human activities on the environment.
Human Biology Program at University of Toronto
The University of Toronto offers an undergraduate program in Human Biology that is jointly offered by the Faculty of Arts & Science and the Faculty of Medicine. The program offers several major and specialist options in: human biology, neuroscience, health & disease, global health, and fundamental genetics and its applications.
Asia
BSc (Honours) Human Biology at All India Institute of Medical Sciences, New Delhi (1980–2002)
BSc (honours) Human Biology at AIIMS (New
Document 3:::
Cancer is caused by genetic changes leading to uncontrolled cell growth and tumor formation. The basic cause of sporadic (non-familial) cancers is DNA damage and genomic instability. A minority of cancers are due to inherited genetic mutations. Most cancers are related to environmental, lifestyle, or behavioral exposures. Cancer is generally not contagious in humans, though it can be caused by oncoviruses and cancer bacteria. The term "environmental", as used by cancer researchers, refers to everything outside the body that interacts with humans. The environment is not limited to the biophysical environment (e.g. exposure to factors such as air pollution or sunlight), but also includes lifestyle and behavioral factors.
Over one third of cancer deaths worldwide (and about 75–80% in the United States) are potentially avoidable by reducing exposure to known factors. Common environmental factors that contribute to cancer death include exposure to different chemical and physical agents (tobacco use accounts for 25–30% of cancer deaths), environmental pollutants, diet and obesity (30–35%), infections (15–20%), and radiation (both ionizing and non-ionizing, up to 10%). These factors act, at least partly, by altering the function of genes within cells. Typically many such genetic changes are required before cancer develops. Aging has been repeatedly and consistently regarded as an important aspect to consider when evaluating the risk factors for the development of particular cancers. Many molecular and cellular changes involved in the development of cancer accumulate during the aging process and eventually manifest as cancer.
Genetics
Although there are over 50 identifiable hereditary forms of cancer, less than 0.3% of the population are carriers of a cancer-related genetic mutation and these make up less than 3–10% of all cancer cases. The vast majority of cancers are non-hereditary ("sporadic cancers"). Hereditary cancers are primarily caused by an inherited genetic d
Document 4:::
Malignant transformation is the process by which cells acquire the properties of cancer. This may occur as a primary process in normal tissue, or secondarily as malignant degeneration of a previously existing benign tumor.
Causes
There are many causes of primary malignant transformation, or tumorigenesis. Most human cancers in the United States are caused by external factors, and these factors are largely avoidable. These factors were summarized by Doll and Peto in 1981, and were still considered to be valid in 2015. These factors are listed in the table.
a Reproductive and sexual behaviors include: number of partners; age at first menstruation; zero versus one or more live births
Examples of diet-related malignant transformation
Diet and colon cancer
Colon cancer provides one example of the mechanisms by which diet, the top factor listed in the table, is an external factor in cancer. The Western diet of African Americans in the United States is associated with a yearly colon cancer rate of 65 per 100,000 individuals, while the high fiber/low fat diet of rural Native Africans in South Africa is associated with a yearly colon cancer rate of <5 per 100,000. Feeding the Western diet for two weeks to Native Africans increased their secondary bile acids, including carcinogenic deoxycholic acid, by 400%, and also changed the colonic microbiota. Evidence reviewed by Sun and Kato indicates that differences in human colonic microbiota play an important role in the progression of colon cancer.
Diet and lung cancer
A second example, relating a dietary component to a cancer, is illustrated by lung cancer. Two large population-based studies were performed, one in Italy and one in the United States. In Italy, the study population consisted of two cohorts: the first, 1721 individuals diagnosed with lung cancer and no severe disease, and the second, 1918 control individuals with absence of lung cancer history or any advanced diseases. All individuals filled out a
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is the most common type of cancer in adult males?
A. prostate cancer
B. lung cancer
C. liver cancer
D. stomach cancer
Answer:
|
|
sciq-2712
|
multiple_choice
|
How thick is the earth's continental crust, on average?
|
[
"35 kilometers",
"60 kilometers",
"25 kilometers",
"15 kilometers"
] |
A
|
Relavent Documents:
Document 0:::
The internal structure of Earth is the layers of the Earth, excluding its atmosphere and hydrosphere. The structure consists of an outer silicate solid crust, a highly viscous asthenosphere and solid mantle, a liquid outer core whose flow generates the Earth's magnetic field, and a solid inner core.
Scientific understanding of the internal structure of Earth is based on observations of topography and bathymetry, observations of rock in outcrop, samples brought to the surface from greater depths by volcanoes or volcanic activity, analysis of the seismic waves that pass through Earth, measurements of the gravitational and magnetic fields of Earth, and experiments with crystalline solids at pressures and temperatures characteristic of Earth's deep interior.
Global properties
"Note: In chondrite model (1), the light element in the core is assumed to be Si. Chondrite model (2) is a model of chemical composition of the mantle corresponding to the model of core shown in chondrite model (1)."Measurements of the force exerted by Earth's gravity can be used to calculate its mass. Astronomers can also calculate Earth's mass by observing the motion of orbiting satellites. Earth's average density can be determined through gravimetric experiments, which have historically involved pendulums. The mass of Earth is about . The average density of Earth is .
Layers
The structure of Earth can be defined in two ways: by mechanical properties such as rheology, or chemically. Mechanically, it can be divided into lithosphere, asthenosphere, mesospheric mantle, outer core, and the inner core. Chemically, Earth can be divided into the crust, upper mantle, lower mantle, outer core, and inner core. The geologic component layers of Earth are at increasing depths below the surface:
Crust and lithosphere
Earth's crust ranges from in depth and is the outermost layer. The thin parts are the oceanic crust, which underlie the ocean basins (5–10 km) and is mafic-rich (dense iron-magnesium silic
Document 1:::
The core–mantle boundary (CMB) of Earth lies between the planet's silicate mantle and its liquid iron–nickel outer core, at a depth of below Earth's surface. The boundary is observed via the discontinuity in seismic wave velocities at that depth due to the differences between the acoustic impedances of the solid mantle and the molten outer core. P-wave velocities are much slower in the outer core than in the deep mantle while S-waves do not exist at all in the liquid portion of the core. Recent evidence suggests a distinct boundary layer directly above the CMB possibly made of a novel phase of the basic perovskite mineralogy of the deep mantle named post-perovskite. Seismic tomography studies have shown significant irregularities within the boundary zone and appear to be dominated by the African and Pacific Large Low-Shear-Velocity Provinces (LLSVP).
The uppermost section of the outer core is thought to be about 500–1,800 K hotter than the overlying mantle, creating a thermal boundary layer. The boundary is thought to harbor topography, much like Earth's surface, that is supported by solid-state convection within the overlying mantle. Variations in the thermal properties of the core-mantle boundary may affect how the outer core's iron-rich fluids flow, which are ultimately responsible for Earth's magnetic field.
The D″ region
The approx. 200 km thick layer of the lower mantle directly above the boundary is referred to as the D″ region ("D double-prime" or "D prime prime") and is sometimes included in discussions regarding the core–mantle boundary zone. The D″ name originates from geophysicist Keith Bullen's designations for the Earth's layers. His system was to label each layer alphabetically, A through G, with the crust as 'A' and the inner core as 'G'. In his 1942 publication of his model, the entire lower mantle was the D layer. In 1949, Bullen found his 'D' layer to actually be two different layers. The upper part of the D layer, about 1800 km thick, was r
Document 2:::
Ricardo Antonio Olea () is a Chilean American who was a research mathematical statistician with the United States Geological Survey (2006–21). Previously, he spent most of his career with the National Oil Company of Chile (ENAP) in Punta Arenas and Santiago, and with the Kansas Geological Survey in Lawrence. He received the William Christian Krumbein Medal in 2004 from the International Association for Mathematical Geosciences. He served as Secretary-General (1992−1996) and President (1996–2000) for the International Association for Mathematical Geosciences; and Secretary General (2019–21) of the Compositional Data Association.
Research
He has been active in geostatistics since the early 1970s, with briefer involvements in various other forms of quantitative modeling, including geophysics, petrophysics, reservoir engineering, energy resources assessment, compositional data analysis, lithostratigraphy, statistics, enhanced oil recovery, coastal processes, economic analysis, coal mining, geohydrology, marine geology, epidemiology ichnology and chemometrics.
Selected books
Ricardo A. Olea, ed., Geostatistical Glossary and Multilingual Dictionary, Oxford, 1991, 177p.
Ricardo A. Olea, Geostatistics for Engineers and Earth Scientists, Kluwer, 1999, 313 p.
Vera Pawlowsky-Glahn, Ricardo A. Olea, Geostatistical Analysis of Compositional Data, Oxford, 2004, 181 p.
George Christakos, Ricardo A. Olea, Marc L. Serre, Hwa-Lung Yu, Lin-Lin Wang, Interdisciplinary Public Health Reasoning and Epidemic Modelling: The Case of Black Death, Springer, 2005, 319 p.
Document 3:::
Dr. Nahid Khazenie is a mechanical engineer who served as president of the IEEE Geoscience and Remote Sensing Society from 1998 to 1999.
Khazenie completed her undergraduate education at the Michigan Technological University before going on to receive several graduate degrees from the University of Texas at Austin, including a Ph.D. in 1987 for Mechanical Engineering and Operations Research. She joined the faculty and was a research scientist, specializing in remote sensing applications in agriculture and ocean studies. Because of her work there, she became a Senior Scientist appointment to the Naval Research Laboratory and then NASA as Earth Science Enterprise Education Programs Manager.
Document 4:::
Earth's crustal evolution involves the formation, destruction and renewal of the rocky outer shell at that planet's surface.
The variation in composition within the Earth's crust is much greater than that of other terrestrial planets. Mars, Venus, Mercury and other planetary bodies have relatively quasi-uniform crusts unlike that of the Earth which contains both oceanic and continental plates. This unique property reflects the complex series of crustal processes that have taken place throughout the planet's history, including the ongoing process of plate tectonics.
The proposed mechanisms regarding Earth's crustal evolution take a theory-orientated approach. Fragmentary geologic evidence and observations provide the basis for hypothetical solutions to problems relating to the early Earth system. Therefore, a combination of these theories creates both a framework of current understanding and also a platform for future study.
Early crust
Mechanisms of early crust formation
The early Earth was entirely molten. This was due to high temperatures created and maintained by the following processes:
Compression of the early atmosphere
Rapid axial rotation
Regular impacts with neighbouring planetesimals.
The mantle remained hotter than modern day temperatures throughout the Archean. Over time the Earth began to cool as planetary accretion slowed and heat stored within the magma ocean was lost to space through radiation.
A theory for the initiation of magma solidification states that once cool enough, the cooler base of the magma ocean would begin to crystallise first. This is because pressure of 25 GPa at the surface cause the solidus to lower. The formation of a thin 'chill-crust' at the extreme surface would provide thermal insulation to the shallow sub surface, keeping it warm enough to maintain the mechanism of crystallisation from the deep magma ocean.
The composition of the crystals produced during the crystallisation of the magma ocean varied with depth. Ex
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
How thick is the earth's continental crust, on average?
A. 35 kilometers
B. 60 kilometers
C. 25 kilometers
D. 15 kilometers
Answer:
|
|
sciq-9188
|
multiple_choice
|
What type of electrons are electrons that are not confined to the bond between two atoms?
|
[
"virtualized",
"internalized",
"detached",
"delocalized"
] |
D
|
Relavent Documents:
Document 0:::
A non-bonding electron is an electron not involved in chemical bonding. This can refer to:
Lone pair, with the electron localized on one atom.
Non-bonding orbital, with the electron delocalized throughout the molecule.
Chemical bonding
Document 1:::
Steudel R 2020, Chemistry of the Non-metals: Syntheses - Structures - Bonding - Applications, in collaboration with D Scheschkewitz, Berlin, Walter de Gruyter, . ▲
An updated translation of the 5th German edition of 2013, incorporating the literature up to Spring 2019. Twenty-three nonmetals, including B, Si, Ge, As, Se, Te, and At but not Sb (nor Po). The nonmetals are identified on the basis of their electrical conductivity at absolute zero putatively being close to zero, rather than finite as in the case of metals. That does not work for As however, which has the electronic structure of a semimetal (like Sb).
Halka M & Nordstrom B 2010, "Nonmetals", Facts on File, New York,
A reading level 9+ book covering H, C, N, O, P, S, Se. Complementary books by the same authors examine (a) the post-transition metals (Al, Ga, In, Tl, Sn, Pb and Bi) and metalloids (B, Si, Ge, As, Sb, Te and Po); and (b) the halogens and noble gases.
Woolins JD 1988, Non-Metal Rings, Cages and Clusters, John Wiley & Sons, Chichester, .
A more advanced text that covers H; B; C, Si, Ge; N, P, As, Sb; O, S, Se and Te.
Steudel R 1977, Chemistry of the Non-metals: With an Introduction to Atomic Structure and Chemical Bonding, English edition by FC Nachod & JJ Zuckerman, Berlin, Walter de Gruyter, . ▲
Twenty-four nonmetals, including B, Si, Ge, As, Se, Te, Po and At.
Powell P & Timms PL 1974, The Chemistry of the Non-metals, Chapman & Hall, London, . ▲
Twenty-two nonmetals including B, Si, Ge, As and Te. Tin and antimony are shown as being intermediate between metals and nonmetals; they are later shown as either metals or nonmetals. Astatine is counted as a metal.
Document 2:::
This page deals with the electron affinity as a property of isolated atoms or molecules (i.e. in the gas phase). Solid state electron affinities are not listed here.
Elements
Electron affinity can be defined in two equivalent ways. First, as the energy that is released by adding an electron to an isolated gaseous atom. The second (reverse) definition is that electron affinity is the energy required to remove an electron from a singly charged gaseous negative ion. The latter can be regarded as the ionization energy of the –1 ion or the zeroth ionization energy. Either convention can be used.
Negative electron affinities can be used in those cases where electron capture requires energy, i.e. when capture can occur only if the impinging electron has a kinetic energy large enough to excite a resonance of the atom-plus-electron system. Conversely electron removal from the anion formed in this way releases energy, which is carried out by the freed electron as kinetic energy. Negative ions formed in these cases are always unstable. They may have lifetimes of the order of microseconds to milliseconds, and invariably autodetach after some time.
Molecules
The electron affinities Eea of some molecules are given in the table below, from the lightest to the heaviest. Many more have been listed by . The electron affinities of the radicals OH and SH are the most precisely known of all molecular electron affinities.
Second and third electron affinity
Bibliography
.
.
Updated values can be found in the NIST chemistry webbook for around three dozen elements and close to 400 compounds.
Specific molecules
Document 3:::
Stannide ions,
Some examples of stannide Zintl ions are listed below. Some of them contain 2-centre 2-electron bonds (2c-2e), others are "electron deficient" and bonding sometimes can be described using polyhedral skeletal electron pair theory (Wade's rules) where the number of valence electrons contributed by each tin atom is considered to be 2 (the s electrons do not contribute). There are some examples of silicide and plumbide ions with similar structures, for example tetrahedral , the chain anion (Si2−)n, and .
Sn4− found for example in Mg2Sn.
, tetrahedral with 2c-2e bonds e.g. in CsSn.
, tetrahedral closo-cluster with 10 electrons (2n + 2).
(Sn2−)n zig-zag chain polymeric anion with 2c-2e bonds found for example in BaSn.
closo-
Document 4:::
Electron deficiency (and electron-deficient) is jargon that is used in two contexts: species that violate the octet rule because they have too few valence electrons and species that happen to follow the octet rule but have electron-acceptor properties, forming donor-acceptor charge-transfer salts.
Octet rule violations
Traditionally, "electron-deficiency" is used as a general descriptor for boron hydrides and other molecules which do not have enough valence electrons to form localized (2-centre 2-electron) bonds joining all atoms. For example, diborane (B2H6) would require a minimum of 7 localized bonds with 14 electrons to join all 8 atoms, but there are only 12 valence electrons. A similar situation exists in trimethylaluminium. The electron deficiency in such compounds is similar to metallic bonding.
Electron-acceptor molecules
Alternatively, electron-deficiency describes molecules or ions that function as electron acceptors. Such electron-deficient species obey the octet rule, but they have (usually mild) oxidizing properties. 1,3,5-Trinitrobenzene and related polynitrated aromatic compounds are often described as electron-deficient. Electron deficiency can be measured by linear free-energy relationships: "a strongly negative ρ value indicates a large electron demand at the reaction center, from which it may be concluded that a highly electron-deficient center, perhaps an incipient carbocation, is involved."
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What type of electrons are electrons that are not confined to the bond between two atoms?
A. virtualized
B. internalized
C. detached
D. delocalized
Answer:
|
|
sciq-7305
|
multiple_choice
|
More common in developing countries, parasitic diseases caused by roundworms often result from poor practice of what?
|
[
"preventative medicine",
"agriculture",
"personal hygiene",
"education"
] |
C
|
Relavent Documents:
Document 0:::
Helminthiasis, also known as worm infection, is any macroparasitic disease of humans and other animals in which a part of the body is infected with parasitic worms, known as helminths. There are numerous species of these parasites, which are broadly classified into tapeworms, flukes, and roundworms. They often live in the gastrointestinal tract of their hosts, but they may also burrow into other organs, where they induce physiological damage.
Soil-transmitted helminthiasis and schistosomiasis are the most important helminthiases, and are among the neglected tropical diseases. These group of helminthiases have been targeted under the joint action of the world's leading pharmaceutical companies and non-governmental organizations through a project launched in 2012 called the London Declaration on Neglected Tropical Diseases, which aimed to control or eradicate certain neglected tropical diseases by 2020.
Helminthiasis has been found to result in poor birth outcome, poor cognitive development, poor school and work performance, poor socioeconomic development, and poverty. Chronic illness, malnutrition, and anemia are further examples of secondary effects.
Soil-transmitted helminthiases are responsible for parasitic infections in as much as a quarter of the human population worldwide. One well-known example of soil-transmitted helminthiases is ascariasis.
Types of parasitic helminths
Of all the known helminth species, the most important helminths with respect to understanding their transmission pathways, their control, inactivation and enumeration in samples of human excreta from dried feces, faecal sludge, wastewater, and sewage sludge are:
soil-transmitted helminths, including Ascaris lumbricoides (the most common worldwide), Trichuris trichiura, Necator americanus, Strongyloides stercoralis and Ancylostoma duodenale
Hymenolepis nana
Taenia saginata
Enterobius
Fasciola hepatica
Schistosoma mansoni
Toxocara canis
Toxocara cati
Helminthiases are classified a
Document 1:::
Alternative medicine degrees include academic degrees, first professional degrees, qualifications or diplomas issued by accredited and legally recognised academic institutions in alternative medicine or related areas, either human or animal.
Examples
Examples of alternative medicine degrees include:
Ayurveda - BSc, MSc, BAMC, MD(Ayurveda), M.S.(Ayurveda), Ph.D(Ayurveda)
Siddha medicine - BSMS, MD(Siddha), Ph.D(Siddha)
Acupuncture - BSc, LAc, DAc, AP, DiplAc, MAc
Herbalism - Acs, BSc, Msc.
Homeopathy - BSc, MSc, DHMs, BHMS, M.D. (HOM), PhD in homoeopathy
Naprapathy - DN
Naturopathic medicine - BSc, MSc, BNYS, MD (Naturopathy), ND, NMD
Oriental Medicine - BSc, MSOM, MSTOM, KMD (Korea), BCM (Hong Kong), MCM (Hong Kong), BChinMed (Hong Kong), MChinMed (Hong Kong), MD (Taiwan), MB (China), TCM-Traditional Chinese medicine master (China)
Osteopathy - BOst, BOstMed, BSc (Osteo), DipOsteo
Document 2:::
The Science, Technology, Engineering and Mathematics Network or STEMNET is an educational charity in the United Kingdom that seeks to encourage participation at school and college in science and engineering-related subjects (science, technology, engineering, and mathematics) and (eventually) work.
History
It is based at Woolgate Exchange near Moorgate tube station in London and was established in 1996. The chief executive is Kirsten Bodley. The STEMNET offices are housed within the Engineering Council.
Function
Its chief aim is to interest children in science, technology, engineering and mathematics. Primary school children can start to have an interest in these subjects, leading secondary school pupils to choose science A levels, which will lead to a science career. It supports the After School Science and Engineering Clubs at schools. There are also nine regional Science Learning Centres.
STEM ambassadors
To promote STEM subjects and encourage young people to take up jobs in these areas, STEMNET have around 30,000 ambassadors across the UK. these come from a wide selection of the STEM industries and include TV personalities like Rob Bell.
Funding
STEMNET used to receive funding from the Department for Education and Skills. Since June 2007, it receives funding from the Department for Children, Schools and Families and Department for Innovation, Universities and Skills, since STEMNET sits on the chronological dividing point (age 16) of both of the new departments.
See also
The WISE Campaign
Engineering and Physical Sciences Research Council
National Centre for Excellence in Teaching Mathematics
Association for Science Education
Glossary of areas of mathematics
Glossary of astronomy
Glossary of biology
Glossary of chemistry
Glossary of engineering
Glossary of physics
Document 3:::
MicrobeLibrary is a permanent collection of over 1400 original peer-reviewed resources for teaching undergraduate microbiology. It is provided by the American Society for Microbiology, Washington DC, United States.
Contents include curriculum activities; images and animations; reviews of books, websites and other resources; and articles from Focus on Microbiology Education, Microbiology Education and Microbe. Around 40% of the materials are free to educators and students, the remainder require a subscription. the service is suspended with the message to:
"Please check back with us in 2017".
External links
MicrobeLibrary
Microbiology
Document 4:::
Neglected tropical diseases (NTDs) are a diverse group of tropical infections that are common in low-income populations in developing regions of Africa, Asia, and the Americas. They are caused by a variety of pathogens, such as viruses, bacteria, protozoa, and parasitic worms (helminths). These diseases are contrasted with the "big three" infectious diseases (HIV/AIDS, tuberculosis, and malaria), which generally receive greater treatment and research funding. In sub-Saharan Africa, the effect of neglected tropical diseases as a group is comparable to that of malaria and tuberculosis. NTD co-infection can also make HIV/AIDS and tuberculosis more deadly.
Some treatments for NTDs are relatively inexpensive. For example, treatment for schistosomiasis costs US$0.20 per child per year. Nevertheless, in 2010 it was estimated that control of neglected diseases would require funding of between US$2 billion and $3 billion over the subsequent five to seven years. Some pharmaceutical companies have committed to donating all the drug therapies required, and mass drug administration efforts (for example, mass deworming) have been successful in several countries. While preventive measures are often more accessible in the developed world, they are not universally available in poorer areas.
Within developed countries, neglected tropical diseases affect the very poorest in society. In the United States, there are up to 1.46 million families, including 2.8 million children, living on less than two dollars per day. In developed countries, the burdens of neglected tropical diseases are often overshadowed by other public health issues. However, many of the same issues put populations at risk in developed as well as developing nations. For example, other problems stemming from poverty, such as lack of adequate housing, can expose individuals to the vectors of these diseases.
Twenty neglected tropical diseases are prioritized by the World Health Organization (WHO), though other organiz
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
More common in developing countries, parasitic diseases caused by roundworms often result from poor practice of what?
A. preventative medicine
B. agriculture
C. personal hygiene
D. education
Answer:
|
|
sciq-9540
|
multiple_choice
|
What term is used to describe muscle fibers that allows muscles to contract?
|
[
"organisms",
"fluorescence",
"organelle",
"ligaments"
] |
C
|
Relavent Documents:
Document 0:::
Muscle contraction is the activation of tension-generating sites within muscle cells. In physiology, muscle contraction does not necessarily mean muscle shortening because muscle tension can be produced without changes in muscle length, such as when holding something heavy in the same position. The termination of muscle contraction is followed by muscle relaxation, which is a return of the muscle fibers to their low tension-generating state.
For the contractions to happen, the muscle cells must rely on the interaction of two types of filaments: thin and thick filaments.
The major constituent of thin filaments is a chain formed by helical coiling of two strands of actin, and thick filaments dominantly consist of chains of the motor-protein myosin. Together, these two filaments form myofibrils - the basic functional organelles in the skeletal muscle system.
In vertebrates, skeletal muscle contractions are neurogenic as they require synaptic input from motor neurons. A single motor neuron is able to innervate multiple muscle fibers, thereby causing the fibers to contract at the same time. Once innervated, the protein filaments within each skeletal muscle fiber slide past each other to produce a contraction, which is explained by the sliding filament theory. The contraction produced can be described as a twitch, summation, or tetanus, depending on the frequency of action potentials. In skeletal muscles, muscle tension is at its greatest when the muscle is stretched to an intermediate length as described by the length-tension relationship.
Unlike skeletal muscle, the contractions of smooth and cardiac muscles are myogenic (meaning that they are initiated by the smooth or heart muscle cells themselves instead of being stimulated by an outside event such as nerve stimulation), although they can be modulated by stimuli from the autonomic nervous system. The mechanisms of contraction in these muscle tissues are similar to those in skeletal muscle tissues.
Muscle contra
Document 1:::
Myology is the study of the muscular system, including the study of the structure, function and diseases of muscle. The muscular system consists of skeletal muscle, which contracts to move or position parts of the body (e.g., the bones that articulate at joints), smooth and cardiac muscle that propels, expels or controls the flow of fluids and contained substance.
See also
Myotomy
Oral myology
Document 2:::
Myomeres are blocks of skeletal muscle tissue arranged in sequence, commonly found in aquatic chordates. Myomeres are separated from adjacent myomeres by connective fascia (myosepta) and most easily seen in larval fishes or in the olm. Myomere counts are sometimes used for identifying specimens, since their number corresponds to the number of vertebrae in the adults. Location varies, with some species containing these only near the tails, while some have them located near the scapular or pelvic girdles. Depending on the species, myomeres could be arranged in an epaxial or hypaxial manner. Hypaxial refers to ventral muscles and related structures while epaxial refers to more dorsal muscles. The horizontal septum divides these two regions in vertebrates from cyclostomes to gnathostomes. In terrestrial chordates, the myomeres become fused as well as indistinct, due to the disappearance of myosepta.
Shape
The shape of myomeres varies by species. Myomeres are commonly zig-zag, "V" (lancelets), "W" (fishes), or straight (tetrapods)– shaped muscle fibers. Generally, cyclostome myomeres are arranged in vertical strips while those of jawed fishes are folded in a complex matter due to swimming capability evolution. Specifically, myomeres of elasmobranchs and eels are “W”-shaped. Contrastingly, myomeres of tetrapods run vertically and do not display complex folding. Another species with simply-lain myomeres are mudpuppies. Myomeres overlap each other in succession, meaning myomere activation also allows neighboring myomeres to activate.
Myomeres are made up of myoglobin-rich dark muscle as well as white muscle. Dark muscle, generally, functions as slow-twitch muscle fibers while white muscle is composed of fast-twitch fibers.
Function
Specifically, three types of myomeres in fish-like chordates include amphioxine (lancelet), cyclostomine (jawless fish), and gnathostomine (jawed fish). A common function shared by all of these is that they function to flex the body lateral
Document 3:::
Anatomical terminology is used to uniquely describe aspects of skeletal muscle, cardiac muscle, and smooth muscle such as their actions, structure, size, and location.
Types
There are three types of muscle tissue in the body: skeletal, smooth, and cardiac.
Skeletal muscle
Skeletal muscle, or "voluntary muscle", is a striated muscle tissue that primarily joins to bone with tendons. Skeletal muscle enables movement of bones, and maintains posture. The widest part of a muscle that pulls on the tendons is known as the belly.
Muscle slip
A muscle slip is a slip of muscle that can either be an anatomical variant, or a branching of a muscle as in rib connections of the serratus anterior muscle.
Smooth muscle
Smooth muscle is involuntary and found in parts of the body where it conveys action without conscious intent. The majority of this type of muscle tissue is found in the digestive and urinary systems where it acts by propelling forward food, chyme, and feces in the former and urine in the latter. Other places smooth muscle can be found are within the uterus, where it helps facilitate birth, and the eye, where the pupillary sphincter controls pupil size.
Cardiac muscle
Cardiac muscle is specific to the heart. It is also involuntary in its movement, and is additionally self-excitatory, contracting without outside stimuli.
Actions of skeletal muscle
As well as anatomical terms of motion, which describe the motion made by a muscle, unique terminology is used to describe the action of a set of muscles.
Agonists and antagonists
Agonist muscles and antagonist muscles are muscles that cause or inhibit a movement.
Agonist muscles are also called prime movers since they produce most of the force, and control of an action. Agonists cause a movement to occur through their own activation. For example, the triceps brachii contracts, producing a shortening (concentric) contraction, during the up phase of a push-up (elbow extension). During the down phase of a push-up, th
Document 4:::
Vertebrates
Tendon cells, or tenocytes, are elongated fibroblast type cells. The cytoplasm is stretched between the collagen fibres of the tendon. They have a central cell nucleus with a prominent nucleolus. Tendon cells have a well-developed rough endoplasmic reticulum and they are responsible for synthesis and turnover of tendon fibres and ground substance.
Invertebrates
Tendon cells form a connecting epithelial layer between the muscle and shell in molluscs. In gastropods, for example, the retractor muscles connect to the shell via tendon cells. Muscle cells are attached to the collagenous myo-tendon space via hemidesmosomes. The myo-tendon space is then attached to the base of the tendon cells via basal hemidesmosomes, while apical hemidesmosomes, which sit atop microvilli, attach the tendon cells to a thin layer of collagen. This is in turn attached to the shell via organic fibres which insert into the shell. Molluscan tendon cells appear columnar and contain a large basal cell nucleus. The cytoplasm is filled with granular endoplasmic reticulum and sparse golgi. Dense bundles of microfilaments run the length of the cell connecting the basal to the apical hemidesmosomes.
See also
List of human cell types derived from the germ layers
List of distinct cell types in the adult human body
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What term is used to describe muscle fibers that allows muscles to contract?
A. organisms
B. fluorescence
C. organelle
D. ligaments
Answer:
|
|
sciq-6646
|
multiple_choice
|
What is the term for atherosclerosis of arteries that supply the heart muscle?
|
[
"essential heart disease",
"coronary heart disease",
"rapid heart disease",
"cardiovascular disease"
] |
B
|
Relavent Documents:
Document 0:::
Atherosclerosis is a pattern of the disease arteriosclerosis, characterized by development of abnormalities called lesions in walls of arteries. These lesions may lead to narrowing of the arteries' walls due to buildup of atheromatous plaques. At onset there are usually no symptoms, but if they develop, symptoms generally begin around middle age. In severe cases, it can result in coronary artery disease, stroke, peripheral artery disease, or kidney disorders, depending on which body parts(s) the affected arteries are located in the body.
The exact cause of atherosclerosis is unknown and is proposed to be multifactorial. Risk factors include abnormal cholesterol levels, elevated levels of inflammatory biomarkers, high blood pressure, diabetes, smoking (both active and passive smoking), obesity, genetic factors, family history, lifestyle habits, and an unhealthy diet. Plaque is made up of fat, cholesterol, calcium, and other substances found in the blood. The narrowing of arteries limits the flow of oxygen-rich blood to parts of the body. Diagnosis is based upon a physical exam, electrocardiogram, and exercise stress test, among others.
Prevention is generally by eating a healthy diet, exercising, not smoking, and maintaining a normal weight. Treatment of established disease may include medications to lower cholesterol such as statins, blood pressure medication, or medications that decrease clotting, such as aspirin. A number of procedures may also be carried out such as percutaneous coronary intervention, coronary artery bypass graft, or carotid endarterectomy.
Atherosclerosis generally starts when a person is young and worsens with age. Almost all people are affected to some degree by the age of 65. It is the number one cause of death and disability in developed countries. Though it was first described in 1575, there is evidence that the condition occurred in people more than 5,000 years ago.
Signs and symptoms
Atherosclerosis is asymptomatic for decades because
Document 1:::
Lucien Campeau (June 20, 1927March 15, 2010) was a Canadian cardiologist. He was a full professor at the Université de Montréal. He is best known for performing the world's first transradial coronary angiogram. Campeau was one of the founding staff of the Montreal Heart Institute, joining in 1957. He is also well known for developing the Canadian Cardiovascular Society grading of angina pectoris.
Education
Campeau received his M.D. degree from the University of Laval in 1953 and completed a fellowship in Cardiology at Johns Hopkins Hospital from 1956 to 1957. He later became a professor at University of Montreal in 1961 and was one of the co-founders of the Montreal Heart Institute.
In his lifetime, Campeau was awarded the Research Achievement Award of the Canadian Cardiovascular Society. In 2004, he was named “Cardiologue émérite 2004” by the Association des cardiologues du Québec.
Document 2:::
Heart nanotechnology is the "Engineering of functional systems at the molecular scale" ("Nanotechnology Research").
Nanotechnology
Nanotechnology deals with structures and materials that are approximately one to one-hundred nanometers in length. At this microscopic level, quantum mechanics take place and are in effect, resulting in behaviors that would seem quite strange compared to what humans see with the naked eye (regular matter). Nanotechnology is used for a wide variety of fields of technology, ranging from energy to electronics to medicine. In the category of medicine, nanotechnology is still relatively new and has not yet been widely adopted by the field. It is possible that nanotechnology could be the new breakthrough of medicine and may eventually be the solution and cure for many of the health problems that humans encounter. Nanotechnology may lead to the cure for illnesses such as the common cold, diseases, and cancer. It is already starting to be used as a treatment for some serious health issues; more specifically it is being used to treat the heart and cancer.
Nanomedicine
Nanotechnology in the field of medicine is more commonly referred to as nanomedicine. Nanomedicine that deals with helping the heart is really starting to take off and gain in popularity compared to most of the other fields that nanomedicine currently has to offer. There are several heart problems that nanotechnology has promising evidence of being effective in the treatment of heart disease in the near future.
Examples
It should hopefully be able to treat heart valves that are defective; and detect and treat arterial plaque in the heart ("Nanotechnology Made Clear"). Nanomedicine should be able to help heal the hearts of people that have already been victims of heart disease and heart attacks. On the other hand, it will also play a key role in finding people with a high risk of having heart disease, and will be able to help prevent heart attacks from happening in the first p
Document 3:::
Cardiophysics is an interdisciplinary science that stands at the junction of cardiology and medical physics, with researchers using the methods of, and theories from, physics to study cardiovascular system at different levels of its organisation, from the molecular scale to whole organisms. Being formed historically as part of systems biology, cardiophysics designed to reveal connections between the physical mechanisms, underlying the organization of the cardiovascular system, and biological features of its functioning.
Zbigniew R. Struzik seems to be a first author who used the term in a scientific publication in 2004.
One can use interchangeably also the terms cardiovascular physics.
See also
Medical physics
Important publications in medical physics
Biomedicine
Biomedical engineering
Physiome
Nanomedicine
Document 4:::
Cardiac muscle (also called heart muscle or myocardium) is one of three types of vertebrate muscle tissues, with the other two being skeletal muscle and smooth muscle. It is an involuntary, striated muscle that constitutes the main tissue of the wall of the heart. The cardiac muscle (myocardium) forms a thick middle layer between the outer layer of the heart wall (the pericardium) and the inner layer (the endocardium), with blood supplied via the coronary circulation. It is composed of individual cardiac muscle cells joined by intercalated discs, and encased by collagen fibers and other substances that form the extracellular matrix.
Cardiac muscle contracts in a similar manner to skeletal muscle, although with some important differences. Electrical stimulation in the form of a cardiac action potential triggers the release of calcium from the cell's internal calcium store, the sarcoplasmic reticulum. The rise in calcium causes the cell's myofilaments to slide past each other in a process called excitation-contraction coupling.
Diseases of the heart muscle known as cardiomyopathies are of major importance. These include ischemic conditions caused by a restricted blood supply to the muscle such as angina, and myocardial infarction.
Structure
Gross anatomy
Cardiac muscle tissue or myocardium forms the bulk of the heart. The heart wall is a three-layered structure with a thick layer of myocardium sandwiched between the inner endocardium and the outer epicardium (also known as the visceral pericardium). The inner endocardium lines the cardiac chambers, covers the cardiac valves, and joins with the endothelium that lines the blood vessels that connect to the heart. On the outer aspect of the myocardium is the epicardium which forms part of the pericardial sac that surrounds, protects, and lubricates the heart.
Within the myocardium, there are several sheets of cardiac muscle cells or cardiomyocytes. The sheets of muscle that wrap around the left ventricle clos
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is the term for atherosclerosis of arteries that supply the heart muscle?
A. essential heart disease
B. coronary heart disease
C. rapid heart disease
D. cardiovascular disease
Answer:
|
|
sciq-10074
|
multiple_choice
|
How does the bacteria look when seen with the naked eye?
|
[
"oily smear",
"slippery smear",
"dry smear",
"slimy smear"
] |
D
|
Relavent Documents:
Document 0:::
Microbial cytology is the study of microscopic and submicroscopic details of microorganisms. Origin of "Microbial" 1880-85; < Greek mīkro- micro- small + bíos life). "Cytology" 1857; < Cyto-is derived from the Greek "kytos" meaning "hollow, as a cell or container." + -logy meaning "the study of"). Microbial cytology is analyzed under a microscope for cells which were collected from a part of the body. The main purpose of microbial cytology is to see the structure of the cells, and how they form and operate.
Document 1:::
In microbiology, colonial morphology refers to the visual appearance of bacterial or fungal colonies on an agar plate. Examining colonial morphology is the first step in the identification of an unknown microbe. The systematic assessment of the colonies' appearance, focusing on aspects like size, shape, colour, opacity, and consistency, provides clues to the identity of the organism, allowing microbiologists to select appropriate tests to provide a definitive identification.
Procedure
When a specimen arrives in the microbiology laboratory, it is inoculated into an agar plate and placed in an incubator to encourage microbial growth. Because the appearance of microbial colonies changes as they grow, colonial morphology is examined at a specific time after the plate is inoculated. Usually, the plate is read at 18–24 hours post-inoculation, but times may differ for slower-growing organisms like fungi. The microbiologist examines the appearance of the colony, noting specific features such as size, colour, shape, consistency, and opacity. A hand lens or magnifying glass may be used to view colonies in greater detail.
The opacity of a microbial colony can be described as transparent, translucent, or opaque. Staphylococci are usually opaque, while many Streptococcus species are translucent. The overall shape of the colony may be characterized as circular, irregular, or punctiform (like pinpoints). The vertical growth or elevation of the colony, another identifying characteristic, is assessed by tilting the agar plate to the side and is denoted as flat, raised, convex, pulvinate (very convex), umbilicate (having a depression in the centre) or umbonate (having a bump in the centre). The edge of the colony may be separately described using terms like smooth, rough, irregular and filamentous. Bacillus anthracis is notable for its filamentous appearance, which is sometimes described as resembling Medusa's head.
Consistency is examined by physically manipulating the colony w
Document 2:::
Thrombolites (from Ancient Greek θρόμβος thrómbos meaning "clot" and λῐ́θος líthos meaning "stone") are clotted accretionary structures formed in shallow water by the trapping, binding, and cementation of sedimentary grains by biofilms of microorganisms, especially cyanobacteria.
Structures
Thrombolites have a clotted structure without the laminae of stromatolites. Each clot within a thrombolite mound is a separate cyanobacterial colony. The clots are on the scale of millimetres to centimetres and may be interspersed with sand, mud or sparry carbonate. Clots that make up thrombolites are called thromboids to avoid confusion with other clotted textures. The larger clots make up more than 40% of a thrombolite's volume and each clot has a complex internal structure of cells and rimmed lobes resulting primarily from calcification of the cyanobacterial colony. Very little sediment is found within the clots because the main growth method is calcification rather than sediment trapping. There is active debate about the size of thromboids, with some seeing thromboids as a macrostructural feature (domical hemispheroid) and others viewing thromboids as a mesostructural feature (random polylobate and subspherical mesoclots).
Types
There are two main types of thrombolites:
Calcified microbe thrombolites
This type of thrombolites contain clots that are dominantly composed of calcified microfossil components. These clots do not have a fixed form or size and can expand vertically. Furthermore, burrows and trilobite fragments can exist in these thrombolites.
Coarse agglutinated thrombolites
This type of thrombolites is composed of small openings that trap fine-grained sediments. They are also known "thrombolitic-stromatolites" due to their close relation with the same composition of stromatolites. Because they trap sediment, their formation is linked to the rise of algal-cyanobacterial mats.
Differences from stromatolites
Thrombolites can be distinguished from microbialite
Document 3:::
Clue cells are epithelial cells of the vagina that get their distinctive stippled appearance by being covered with bacteria. The etymology behind the term "clue" cell derives from the original research article from Gardner and Dukes describing the characteristic cells. The name was chosen for its brevity in describing the sine qua non of bacterial vaginosis.
They are a medical sign of bacterial vaginosis, particularly that caused by Gardnerella vaginalis, a group of Gram-variable bacteria. This bacterial infection is characterized by a foul, fishy smelling, thin gray vaginal discharge, and an increase in vaginal pH from around 4.5 to over 5.5.
Document 4:::
The Germs is a comic strip in the UK comic The Beano. It first appeared in issue 2374, dated 16 January 1988, replacing the Rasher strip, where the characters had been introduced the previous week.
The strip was about a boy called Will who had three germs inside him (Ugly Jack Bacteria, Jeremy Germ, and Iris the Virus), and they were constantly making Will ill, requiring many visits to the doctor. Iris' name later changed to Violet Virus.
The Germs were sometimes stopped in their tracks with a grown-up looking germ called Auntie Biotic (a pun on "antibiotic").
The strip was originally drawn by David Sutherland, and was taken over by Vic Neill later on. Around this time Will's name was added onto the title, the wording changing every so often, such as including (also featuring Ill Will), (also starring Ill Will) or (with Ill Will). Due to Neill's work on Billy Whizz and Tim Traveller, the strip appeared on an increasingly irregular basis in the late 1990s, and after Neill's death in 2000, it disappeared from The Beano. In 2004 it briefly returned with a new artist, Nigel Parkinson. However, the strip only made three appearances which were spread over the year, and was dropped once again.
It returned to the comic in October 2011 as reprints of the David Sutherland strips, and later Vic Neill reprints in April 2012, along with Number 13, this time retitled as Totally Gross Germs. The following month, they were retitled again, as "The Germs: Totally Gross".
In Issue 3618 they returned to their original title.
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
How does the bacteria look when seen with the naked eye?
A. oily smear
B. slippery smear
C. dry smear
D. slimy smear
Answer:
|
|
sciq-2442
|
multiple_choice
|
The product of a wave's wavelength and its frequency is what?
|
[
"trough",
"speed",
"amplitude",
"velocity"
] |
B
|
Relavent Documents:
Document 0:::
This is a list of wave topics.
0–9
21 cm line
A
Abbe prism
Absorption spectroscopy
Absorption spectrum
Absorption wavemeter
Acoustic wave
Acoustic wave equation
Acoustics
Acousto-optic effect
Acousto-optic modulator
Acousto-optics
Airy disc
Airy wave theory
Alfvén wave
Alpha waves
Amphidromic point
Amplitude
Amplitude modulation
Animal echolocation
Antarctic Circumpolar Wave
Antiphase
Aquamarine Power
Arrayed waveguide grating
Artificial wave
Atmospheric diffraction
Atmospheric wave
Atmospheric waveguide
Atom laser
Atomic clock
Atomic mirror
Audience wave
Autowave
Averaged Lagrangian
B
Babinet's principle
Backward wave oscillator
Bandwidth-limited pulse
beat
Berry phase
Bessel beam
Beta wave
Black hole
Blazar
Bloch's theorem
Blueshift
Boussinesq approximation (water waves)
Bow wave
Bragg diffraction
Bragg's law
Breaking wave
Bremsstrahlung, Electromagnetic radiation
Brillouin scattering
Bullet bow shockwave
Burgers' equation
Business cycle
C
Capillary wave
Carrier wave
Cherenkov radiation
Chirp
Ernst Chladni
Circular polarization
Clapotis
Closed waveguide
Cnoidal wave
Coherence (physics)
Coherence length
Coherence time
Cold wave
Collimated light
Collimator
Compton effect
Comparison of analog and digital recording
Computation of radiowave attenuation in the atmosphere
Continuous phase modulation
Continuous wave
Convective heat transfer
Coriolis frequency
Coronal mass ejection
Cosmic microwave background radiation
Coulomb wave function
Cutoff frequency
Cutoff wavelength
Cymatics
D
Damped wave
Decollimation
Delta wave
Dielectric waveguide
Diffraction
Direction finding
Dispersion (optics)
Dispersion (water waves)
Dispersion relation
Dominant wavelength
Doppler effect
Doppler radar
Douglas Sea Scale
Draupner wave
Droplet-shaped wave
Duhamel's principle
E
E-skip
Earthquake
Echo (phenomenon)
Echo sounding
Echolocation (animal)
Echolocation (human)
Eddy (fluid dynamics)
Edge wave
Eikonal equation
Ekman layer
Ekman spiral
Ekman transport
El Niño–Southern Oscillation
El
Document 1:::
A wavenumber–frequency diagram is a plot displaying the relationship between the wavenumber (spatial frequency) and the frequency (temporal frequency) of certain phenomena. Usually frequencies are placed on the vertical axis, while wavenumbers are placed on the horizontal axis.
In the atmospheric sciences, these plots are a common way to visualize atmospheric waves.
In the geosciences, especially seismic data analysis, these plots also called f–k plot, in which energy density within a given time interval is contoured on a frequency-versus-wavenumber basis. They are used to examine the direction and apparent velocity of seismic waves and in velocity filter design.
Origins
In general, the relationship between wavelength , frequency , and the phase velocity of a sinusoidal wave is:
Using the wavenumber () and angular frequency () notation, the previous equation can be rewritten as
On the other hand, the group velocity is equal to the slope of the wavenumber–frequency diagram:
Analyzing such relationships in detail often yields information on the physical properties of the medium, such as density, composition, etc.
See also
Dispersion relation
Document 2:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 3:::
In the physical sciences, the wavenumber (or wave number), also known as repetency, is the spatial frequency of a wave, measured in cycles per unit distance (ordinary wavenumber) or radians per unit distance (angular wavenumber). It is analogous to temporal frequency, which is defined as the number of wave cycles per unit time (ordinary frequency) or radians per unit time (angular frequency).
In multidimensional systems, the wavenumber is the magnitude of the wave vector. The space of wave vectors is called reciprocal space. Wave numbers and wave vectors play an essential role in optics and the physics of wave scattering, such as X-ray diffraction, neutron diffraction, electron diffraction, and elementary particle physics. For quantum mechanical waves, the wavenumber multiplied by the reduced Planck's constant is the canonical momentum.
Wavenumber can be used to specify quantities other than spatial frequency. For example, in optical spectroscopy, it is often used as a unit of temporal frequency assuming a certain speed of light.
Definition
Wavenumber, as used in spectroscopy and most chemistry fields, is defined as the number of wavelengths per unit distance, typically centimeters (cm−1):
where λ is the wavelength. It is sometimes called the "spectroscopic wavenumber". It equals the spatial frequency.
For example, a wavenumber in inverse centimeters can be converted to a frequency in gigahertz by multiplying by 29.9792458 cm/ns (the speed of light, in centimeters per nanosecond); conversely, an electromagnetic wave at 29.9792458 GHz has a wavelength of 1 cm in free space.
In theoretical physics, a wave number, defined as the number of radians per unit distance, sometimes called "angular wavenumber", is more often used:
When wavenumber is represented by the symbol , a frequency is still being represented, albeit indirectly. As described in the spectroscopy section, this is done through the relationship , where s is a frequency in hertz. This is done for con
Document 4:::
In signal processing, the energy of a continuous-time signal x(t) is defined as the area under the squared magnitude of the considered signal i.e., mathematically
Unit of will be (unit of signal)2.
And the energy of a discrete-time signal x(n) is defined mathematically as
Relationship to energy in physics
Energy in this context is not, strictly speaking, the same as the conventional notion of energy in physics and the other sciences. The two concepts are, however, closely related, and it is possible to convert from one to the other:
where Z represents the magnitude, in appropriate units of measure, of the load driven by the signal.
For example, if x(t) represents the potential (in volts) of an electrical signal propagating across a transmission line, then Z would represent the characteristic impedance (in ohms) of the transmission line. The units of measure for the signal energy would appear as volt2·seconds, which is not dimensionally correct for energy in the sense of the physical sciences. After dividing by Z, however, the dimensions of E would become volt2·seconds per ohm,
which is equivalent to joules, the SI unit for energy as defined in the physical sciences.
Spectral energy density
Similarly, the spectral energy density of signal x(t) is
where X(f) is the Fourier transform of x(t).
For example, if x(t) represents the magnitude of the electric field component (in volts per meter) of an optical signal propagating through free space, then the dimensions of X(f) would become volt·seconds per meter and would represent the signal's spectral energy density (in volts2·second2 per meter2) as a function of frequency f (in hertz). Again, these units of measure are not dimensionally correct in the true sense of energy density as defined in physics. Dividing by Zo, the characteristic impedance of free space (in ohms), the dimensions become joule-seconds per meter2 or, equivalently, joules per meter2 per hertz, which is dimensionally correct in SI
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
The product of a wave's wavelength and its frequency is what?
A. trough
B. speed
C. amplitude
D. velocity
Answer:
|
|
sciq-9300
|
multiple_choice
|
A membrane has what property if it can control what molecules or ions enter or leave the cell?
|
[
"selective permeability",
"impermeability",
"indiscreet permeability",
"susceptible permeability"
] |
A
|
Relavent Documents:
Document 0:::
Semipermeable membrane is a type of biological or synthetic, polymeric membrane that will allow certain molecules or ions to pass through it by osmosis. The rate of passage depends on the pressure, concentration, and temperature of the molecules or solutes on either side, as well as the permeability of the membrane to each solute. Depending on the membrane and the solute, permeability may depend on solute size, solubility, properties, or chemistry. How the membrane is constructed to be selective in its permeability will determine the rate and the permeability. Many natural and synthetic materials which are rather thick are also semipermeable. One example of this is the thin film on the inside of the egg.
Biological membranes are selectively permeable, with the passage of molecules controlled by facilitated diffusion, passive transport or active transport regulated by proteins embedded in the membrane.
Biological membranes
An example of a biological semi-permeable membrane is the lipid bilayer, on which is based the plasma membrane that surrounds all biological cells. A group of phospholipids (consisting of a phosphate head and two fatty acid tails) arranged into a double layer, the phospholipid bilayer is a semipermeable membrane that is very specific in its permeability. The hydrophilic phosphate heads are in the outside layer and exposed to the water content outside and within the cell. The hydrophobic tails are the layer hidden in the inside of the membrane. Cholesterol molecules are also found throughout the plasma membrane and act as a buffer of membrane fluidity. The phospholipid bilayer is most permeable to small, uncharged solutes. Protein channels are embedded in or through phospholipids, and, collectively, this model is known as the fluid mosaic model. Aquaporins are protein channel pores permeable to water.
Cellular communication
Information can also pass through the plasma membrane when signaling molecules bind to receptors in the cell membrane. Th
Document 1:::
Membrane proteins are common proteins that are part of, or interact with, biological membranes. Membrane proteins fall into several broad categories depending on their location. Integral membrane proteins are a permanent part of a cell membrane and can either penetrate the membrane (transmembrane) or associate with one or the other side of a membrane (integral monotopic). Peripheral membrane proteins are transiently associated with the cell membrane.
Membrane proteins are common, and medically important—about a third of all human proteins are membrane proteins, and these are targets for more than half of all drugs. Nonetheless, compared to other classes of proteins, determining membrane protein structures remains a challenge in large part due to the difficulty in establishing experimental conditions that can preserve the correct conformation of the protein in isolation from its native environment.
Function
Membrane proteins perform a variety of functions vital to the survival of organisms:
Membrane receptor proteins relay signals between the cell's internal and external environments.
Transport proteins move molecules and ions across the membrane. They can be categorized according to the Transporter Classification database.
Membrane enzymes may have many activities, such as oxidoreductase, transferase or hydrolase.
Cell adhesion molecules allow cells to identify each other and interact. For example, proteins involved in immune response
The localization of proteins in membranes can be predicted reliably using hydrophobicity analyses of protein sequences, i.e. the localization of hydrophobic amino acid sequences.
Integral membrane proteins
Integral membrane proteins are permanently attached to the membrane. Such proteins can be separated from the biological membranes only using detergents, nonpolar solvents, or sometimes denaturing agents. They can be classified according to their relationship with the bilayer:
Integral polytopic proteins are transmembran
Document 2:::
Membrane potential (also transmembrane potential or membrane voltage) is the difference in electric potential between the interior and the exterior of a biological cell. That is, there is a difference in the energy required for electric charges to move from the internal to exterior cellular environments and vice versa, as long as there is no acquisition of kinetic energy or the production of radiation. The concentration gradients of the charges directly determine this energy requirement. For the exterior of the cell, typical values of membrane potential, normally given in units of milli volts and denoted as mV, range from –80 mV to –40 mV.
All animal cells are surrounded by a membrane composed of a lipid bilayer with proteins embedded in it. The membrane serves as both an insulator and a diffusion barrier to the movement of ions. Transmembrane proteins, also known as ion transporter or ion pump proteins, actively push ions across the membrane and establish concentration gradients across the membrane, and ion channels allow ions to move across the membrane down those concentration gradients. Ion pumps and ion channels are electrically equivalent to a set of batteries and resistors inserted in the membrane, and therefore create a voltage between the two sides of the membrane.
Almost all plasma membranes have an electrical potential across them, with the inside usually negative with respect to the outside. The membrane potential has two basic functions. First, it allows a cell to function as a battery, providing power to operate a variety of "molecular devices" embedded in the membrane. Second, in electrically excitable cells such as neurons and muscle cells, it is used for transmitting signals between different parts of a cell. Signals are generated by opening or closing of ion channels at one point in the membrane, producing a local change in the membrane potential. This change in the electric field can be quickly sensed by either adjacent or more distant ion chann
Document 3:::
Membranome is the set of biological membranes existing in a specific organism. The term was proposed by British biologist Thomas Cavalier-Smith to discuss epigenetics of biological membranes. The term was also used to define the entire set of membrane proteins in an organism or a combination of membrane proteome and lipidome.
Document 4:::
In cellular biology, membrane transport refers to the collection of mechanisms that regulate the passage of solutes such as ions and small molecules through biological membranes, which are lipid bilayers that contain proteins embedded in them. The regulation of passage through the membrane is due to selective membrane permeability – a characteristic of biological membranes which allows them to separate substances of distinct chemical nature. In other words, they can be permeable to certain substances but not to others.
The movements of most solutes through the membrane are mediated by membrane transport proteins which are specialized to varying degrees in the transport of specific molecules. As the diversity and physiology of the distinct cells is highly related to their capacities to attract different external elements, it is postulated that there is a group of specific transport proteins for each cell type and for every specific physiological stage. This differential expression is regulated through the differential transcription of the genes coding for these proteins and its translation, for instance, through genetic-molecular mechanisms, but also at the cell biology level: the production of these proteins can be activated by cellular signaling pathways, at the biochemical level, or even by being situated in cytoplasmic vesicles. The cell membrane regulates the transport of materials entering and exiting the cell.
Background
Thermodynamically the flow of substances from one compartment to another can occur in the direction of a concentration or electrochemical gradient or against it. If the exchange of substances occurs in the direction of the gradient, that is, in the direction of decreasing potential, there is no requirement for an input of energy from outside the system; if, however, the transport is against the gradient, it will require the input of energy, metabolic energy in this case.
For example, a classic chemical mechanism for separation that does
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
A membrane has what property if it can control what molecules or ions enter or leave the cell?
A. selective permeability
B. impermeability
C. indiscreet permeability
D. susceptible permeability
Answer:
|
|
sciq-7284
|
multiple_choice
|
In which period of the periodic table is nickel found?
|
[
"First Period",
"Second Period",
"fourth period",
"Third Period"
] |
C
|
Relavent Documents:
Document 0:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 1:::
In chemistry and physics, the iron group refers to elements that are in some way related to iron; mostly in period (row) 4 of the periodic table. The term has different meanings in different contexts.
In chemistry, the term is largely obsolete, but it often means iron, cobalt, and nickel, also called the iron triad; or, sometimes, other elements that resemble iron in some chemical aspects.
In astrophysics and nuclear physics, the term is still quite common, and it typically means those three plus chromium and manganese—five elements that are exceptionally abundant, both on Earth and elsewhere in the universe, compared to their neighbors in the periodic table. Titanium and vanadium are also produced in Type Ia supernovae.
General chemistry
In chemistry, "iron group" used to refer to iron and the next two elements in the periodic table, namely cobalt and nickel. These three comprised the "iron triad". They are the top elements of groups 8, 9, and 10 of the periodic table; or the top row of "group VIII" in the old (pre-1990) IUPAC system, or of "group VIIIB" in the CAS system. These three metals (and the three of the platinum group, immediately below them) were set aside from the other elements because they have obvious similarities in their chemistry, but are not obviously related to any of the other groups. The iron group and its alloys exhibit ferromagnetism.
The similarities in chemistry were noted as one of Döbereiner's triads and by Adolph Strecker in 1859. Indeed, Newlands' "octaves" (1865) were harshly criticized for separating iron from cobalt and nickel. Mendeleev stressed that groups of "chemically analogous elements" could have similar atomic weights as well as atomic weights which increase by equal increments, both in his original 1869 paper and his 1889 Faraday Lecture.
Analytical chemistry
In the traditional methods of qualitative inorganic analysis, the iron group consists of those cations which
have soluble chlorides; and
are not precipitated
Document 2:::
The periodic table is an arrangement of the chemical elements, structured by their atomic number, electron configuration and recurring chemical properties. In the basic form, elements are presented in order of increasing atomic number, in the reading sequence. Then, rows and columns are created by starting new rows and inserting blank cells, so that rows (periods) and columns (groups) show elements with recurring properties (called periodicity). For example, all elements in group (column) 18 are noble gases that are largely—though not completely—unreactive.
The history of the periodic table reflects over two centuries of growth in the understanding of the chemical and physical properties of the elements, with major contributions made by Antoine-Laurent de Lavoisier, Johann Wolfgang Döbereiner, John Newlands, Julius Lothar Meyer, Dmitri Mendeleev, Glenn T. Seaborg, and others.
Early history
Nine chemical elements – carbon, sulfur, iron, copper, silver, tin, gold, mercury, and lead, have been known since before antiquity, as they are found in their native form and are relatively simple to mine with primitive tools. Around 330 BCE, the Greek philosopher Aristotle proposed that everything is made up of a mixture of one or more roots, an idea originally suggested by the Sicilian philosopher Empedocles. The four roots, which the Athenian philosopher Plato called elements, were earth, water, air and fire. Similar ideas about these four elements existed in other ancient traditions, such as Indian philosophy.
A few extra elements were known in the age of alchemy: zinc, arsenic, antimony, and bismuth. Platinum was also known to pre-Columbian South Americans, but knowledge of it did not reach Europe until the 16th century.
First categorizations
The history of the periodic table is also a history of the discovery of the chemical elements. The first person in recorded history to discover a new element was Hennig Brand, a bankrupt German merchant. Brand tried to discover
Document 3:::
Material is a substance or mixture of substances that constitutes an object. Materials can be pure or impure, living or non-living matter. Materials can be classified on the basis of their physical and chemical properties, or on their geological origin or biological function. Materials science is the study of materials, their properties and their applications.
Raw materials can be processed in different ways to influence their properties, by purification, shaping or the introduction of other materials. New materials can be produced from raw materials by synthesis.
In industry, materials are inputs to manufacturing processes to produce products or more complex materials.
Historical elements
Materials chart the history of humanity. The system of the three prehistoric ages (Stone Age, Bronze Age, Iron Age) were succeeded by historical ages: steel age in the 19th century, polymer age in the middle of the following century (plastic age) and silicon age in the second half of the 20th century.
Classification by use
Materials can be broadly categorized in terms of their use, for example:
Building materials are used for construction
Building insulation materials are used to retain heat within buildings
Refractory materials are used for high-temperature applications
Nuclear materials are used for nuclear power and weapons
Aerospace materials are used in aircraft and other aerospace applications
Biomaterials are used for applications interacting with living systems
Material selection is a process to determine which material should be used for a given application.
Classification by structure
The relevant structure of materials has a different length scale depending on the material. The structure and composition of a material can be determined by microscopy or spectroscopy.
Microstructure
In engineering, materials can be categorised according to their microscopic structure:
Plastics: a wide range of synthetic or semi-synthetic materials that use polymers as a main ingred
Document 4:::
A pre-STEM program is a course of study at any two-year college that prepares a student to transfer to a four-year school to earn a bachelor's degree in a STEM field.
Overview
The concept of a pre-STEM program is being developed to address America's need for more college-trained professionals in science, technology, engineering, and mathematics (STEM). It is an innovation meant to fill a gap at community colleges that do not have 'major' degree paths that students identify with on their way to earning an Associates degree. Students must complete a considerable amount of STEM coursework before transferring from a two-year school to a four-year school and earn a baccalaureate degree in a STEM field. Schools with a pre-STEM program are able to identify those students and support them with STEM-specific academic and career advising, increasing the student's chances of going on to earn a STEM baccalaureate degree in a timely fashion.
With over 50% of America's college-bound students starting their college career at public or private two-year school, and with a very small proportion of students who start college at a two-year school matriculating to and earning STEM degrees from four-year schools, pre-STEM programs have great potential for broadening participation in baccalaureate STEM studies.
Example programs
The effectiveness of pre-STEM programs is being investigated by a consortium of schools in Missouri: Moberly Area Community College, St. Charles Community College, Metropolitan Community College, and Truman State University.
A larger group of schools met at the Belknap Springs Meetings in October 2009 to discuss the challenges and opportunities presented by STEM-focused partnerships between 2-year and 4-year schools. Each program represented a two-year school and a four-year school that were trying to increase the number of people who earn a baccalaureate degree in a STEM area through various means, some of which were pre-STEM programs. Other methods includes
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
In which period of the periodic table is nickel found?
A. First Period
B. Second Period
C. fourth period
D. Third Period
Answer:
|
|
sciq-8147
|
multiple_choice
|
What type of lines run next to each other?
|
[
"contour lines",
"crater lines",
"perpendicular lines",
"fault lines"
] |
A
|
Relavent Documents:
Document 0:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 1:::
In mathematical psychology and education theory, a knowledge space is a combinatorial structure used to formulate mathematical models describing the progression of a human learner. Knowledge spaces were introduced in 1985 by Jean-Paul Doignon and Jean-Claude Falmagne, and remain in extensive use in the education theory. Modern applications include two computerized tutoring systems, ALEKS and the defunct RATH.
Formally, a knowledge space assumes that a domain of knowledge is a collection of concepts or skills, each of which must be eventually mastered. Not all concepts are interchangeable; some require other concepts as prerequisites. Conversely, competency at one skill may ease the acquisition of another through similarity. A knowledge space marks out which collections of skills are feasible: they can be learned without mastering any other skills. Under reasonable assumptions, the collection of feasible competencies forms the mathematical structure known as an antimatroid.
Researchers and educators usually explore the structure of a discipline's knowledge space as a latent class model.
Motivation
Knowledge Space Theory attempts to address shortcomings of standardized testing when used in educational psychometry. Common tests, such as the SAT and ACT, compress a student's knowledge into a very small range of ordinal ranks, in the process effacing the conceptual dependencies between questions. Consequently, the tests cannot distinguish between true understanding and guesses, nor can they identify a student's particular weaknesses, only the general proportion of skills mastered. The goal of knowledge space theory is to provide a language by which exams can communicate
What the student can do and
What the student is ready to learn.
Model structure
Knowledge Space Theory-based models presume that an educational subject can be modeled as a finite set of concepts, skills, or topics. Each feasible state of knowledge about is then a subset of ; the set of
Document 2:::
Advanced Placement (AP) Physics C: Electricity and Magnetism (also known as AP Physics C: E&M or AP E&M) is an introductory physics course administered by the College Board as part of its Advanced Placement program. It is intended to proxy a second-semester calculus-based university course in electricity and magnetism. The content of Physics C: E&M overlaps with that of AP Physics 2, but Physics 2 is algebra-based and covers other topics outside of electromagnetism, while Physics C is calculus-based and only covers electromagnetism. Physics C: E&M may be combined with its mechanics counterpart to form a year-long course that prepares for both exams.
Course content
E&M is equivalent to an introductory college course in electricity and magnetism for physics or engineering majors. The course modules are:
Electrostatics
Conductors, capacitors, and dielectrics
Electric circuits
Magnetic fields
Electromagnetism.
Methods of calculus are used wherever appropriate in formulating physical principles and in applying them to physical problems. Therefore, students should have completed or be concurrently enrolled in a calculus class.
AP test
The course culminates in an optional exam for which high-performing students may receive some credit towards their college coursework, depending on the institution.
Registration
The AP examination for AP Physics C: Electricity and Magnetism is separate from the AP examination for AP Physics C: Mechanics. Before 2006, test-takers paid only once and were given the choice of taking either one or two parts of the Physics C test.
Format
The exam is typically administered on a Monday afternoon in May. The exam is configured in two categories: a 35-question multiple choice section and a 3-question free response section. Test takers are allowed to use an approved calculator during the entire exam. The test is weighted such that each section is worth half of the final score. This and AP Physics C: Mechanics are the shortest AP exams, with
Document 3:::
A pre-STEM program is a course of study at any two-year college that prepares a student to transfer to a four-year school to earn a bachelor's degree in a STEM field.
Overview
The concept of a pre-STEM program is being developed to address America's need for more college-trained professionals in science, technology, engineering, and mathematics (STEM). It is an innovation meant to fill a gap at community colleges that do not have 'major' degree paths that students identify with on their way to earning an Associates degree. Students must complete a considerable amount of STEM coursework before transferring from a two-year school to a four-year school and earn a baccalaureate degree in a STEM field. Schools with a pre-STEM program are able to identify those students and support them with STEM-specific academic and career advising, increasing the student's chances of going on to earn a STEM baccalaureate degree in a timely fashion.
With over 50% of America's college-bound students starting their college career at public or private two-year school, and with a very small proportion of students who start college at a two-year school matriculating to and earning STEM degrees from four-year schools, pre-STEM programs have great potential for broadening participation in baccalaureate STEM studies.
Example programs
The effectiveness of pre-STEM programs is being investigated by a consortium of schools in Missouri: Moberly Area Community College, St. Charles Community College, Metropolitan Community College, and Truman State University.
A larger group of schools met at the Belknap Springs Meetings in October 2009 to discuss the challenges and opportunities presented by STEM-focused partnerships between 2-year and 4-year schools. Each program represented a two-year school and a four-year school that were trying to increase the number of people who earn a baccalaureate degree in a STEM area through various means, some of which were pre-STEM programs. Other methods includes
Document 4:::
This is a list of Wikipedia articles about curves used in different fields: mathematics (including geometry, statistics, and applied mathematics), physics, engineering, economics, medicine, biology, psychology, ecology, etc.
Mathematics (Geometry)
Algebraic curves
Rational curves
Rational curves are subdivided according to the degree of the polynomial.
Degree 1
Line
Degree 2
Plane curves of degree 2 are known as conics or conic sections and include
Circle
Unit circle
Ellipse
Parabola
Hyperbola
Unit hyperbola
Degree 3
Cubic plane curves include
Cubic parabola
Folium of Descartes
Cissoid of Diocles
Conchoid of de Sluze
Right strophoid
Semicubical parabola
Serpentine curve
Trident curve
Trisectrix of Maclaurin
Tschirnhausen cubic
Witch of Agnesi
Degree 4
Quartic plane curves include
Ampersand curve
Bean curve
Bicorn
Bow curve
Bullet-nose curve
Cartesian oval
Cruciform curve
Deltoid curve
Devil's curve
Hippopede
Kampyle of Eudoxus
Kappa curve
Lemniscate
Lemniscate of Booth
Lemniscate of Gerono
Lemniscate of Bernoulli
Limaçon
Cardioid
Limaçon trisectrix
Ovals of Cassini
Squircle
Trifolium Curve
Degree 5
Degree 6
Astroid
Atriphtaloid
Nephroid
Quadrifolium
Curve families of variable degree
Epicycloid
Epispiral
Epitrochoid
Hypocycloid
Lissajous curve
Poinsot's spirals
Rational normal curve
Rose curve
Curves with genus 1
Bicuspid curve
Cassinoide
Cubic curve
Elliptic curve
Watt's curve
Curves with genus > 1
Bolza surface (genus 2)
Klein quartic (genus 3)
Bring's curve (genus 4)
Macbeath surface (genus 7)
Butterfly curve (algebraic) (genus 7)
Curve families with variable genus
Polynomial lemniscate
Fermat curve
Sinusoidal spiral
Superellipse
Hurwitz surface
Elkies trinomial curves
Hyperelliptic curve
Classical modular curve
Cassini oval
Transcendental curves
Bowditch curve
Brachistochrone
Butterfly curve (transcendental)
Catenary
Clélies
Cochleoid
Cycloid
Horopter
Isochrone
Isochrone of Huygens (Tautochrone)
Isochrone of Leibniz
Isochrone of Varignon
Lamé
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What type of lines run next to each other?
A. contour lines
B. crater lines
C. perpendicular lines
D. fault lines
Answer:
|
|
sciq-1630
|
multiple_choice
|
What does fiber help keep at normal levels in the body?
|
[
"bloodpressure",
"hydration",
"salt and cholesterol",
"sugar and lipids"
] |
D
|
Relavent Documents:
Document 0:::
Animal nutrition focuses on the dietary nutrients needs of animals, primarily those in agriculture and food production, but also in zoos, aquariums, and wildlife management.
Constituents of diet
Macronutrients (excluding fiber and water) provide structural material (amino acids from which proteins are built, and lipids from which cell membranes and some signaling molecules are built) and energy. Some of the structural material can be used to generate energy internally, though the net energy depends on such factors as absorption and digestive effort, which vary substantially from instance to instance. Vitamins, minerals, fiber, and water do not provide energy, but are required for other reasons. A third class dietary material, fiber (i.e., non-digestible material such as cellulose), seems also to be required, for both mechanical and biochemical reasons, though the exact reasons remain unclear.
Molecules of carbohydrates and fats consist of carbon, hydrogen, and oxygen atoms. Carbohydrates range from simple monosaccharides (glucose, fructose, galactose) to complex polysaccharides (starch). Fats are triglycerides, made of assorted fatty acid monomers bound to glycerol backbone. Some fatty acids, but not all, are essential in the diet: they cannot be synthesized in the body. Protein molecules contain nitrogen atoms in addition to carbon, oxygen, and hydrogen. The fundamental components of protein are nitrogen-containing amino acids. Essential amino acids cannot be made by the animal. Some of the amino acids are convertible (with the expenditure of energy) to glucose and can be used for energy production just as ordinary glucose. By breaking down existing protein, some glucose can be produced internally; the remaining amino acids are discarded, primarily as urea in urine. This occurs normally only during prolonged starvation.
Other dietary substances found in plant foods (phytochemicals, polyphenols) are not identified as essential nutrients but appear to impact healt
Document 1:::
Relatively speaking, the brain consumes an immense amount of energy in comparison to the rest of the body. The mechanisms involved in the transfer of energy from foods to neurons are likely to be fundamental to the control of brain function. Human bodily processes, including the brain, all require both macronutrients, as well as micronutrients.
Insufficient intake of selected vitamins, or certain metabolic disorders, may affect cognitive processes by disrupting the nutrient-dependent processes within the body that are associated with the management of energy in neurons, which can subsequently affect synaptic plasticity, or the ability to encode new memories.
Macronutrients
The human brain requires nutrients obtained from the diet to develop and sustain its physical structure and cognitive functions. Additionally, the brain requires caloric energy predominately derived from the primary macronutrients to operate. The three primary macronutrients include carbohydrates, proteins, and fats. Each macronutrient can impact cognition through multiple mechanisms, including glucose and insulin metabolism, neurotransmitter actions, oxidative stress and inflammation, and the gut-brain axis. Inadequate macronutrient consumption or proportion could impair optimal cognitive functioning and have long-term health implications.
Carbohydrates
Through digestion, dietary carbohydrates are broken down and converted into glucose, which is the sole energy source for the brain. Optimal brain function relies on adequate carbohydrate consumption, as carbohydrates provide the quickest source of glucose for the brain. Glucose deficiencies such as hypoglycaemia reduce available energy for the brain and impair all cognitive processes and performance. Additionally, situations with high cognitive demand, such as learning a new task, increase brain glucose utilization, depleting blood glucose stores and initiating the need for supplementation.
Complex carbohydrates, especially those with high d
Document 2:::
Human nutrition deals with the provision of essential nutrients in food that are necessary to support human life and good health. Poor nutrition is a chronic problem often linked to poverty, food security, or a poor understanding of nutritional requirements. Malnutrition and its consequences are large contributors to deaths, physical deformities, and disabilities worldwide. Good nutrition is necessary for children to grow physically and mentally, and for normal human biological development.
Overview
The human body contains chemical compounds such as water, carbohydrates, amino acids (found in proteins), fatty acids (found in lipids), and nucleic acids (DNA and RNA). These compounds are composed of elements such as carbon, hydrogen, oxygen, nitrogen, and phosphorus. Any study done to determine nutritional status must take into account the state of the body before and after experiments, as well as the chemical composition of the whole diet and of all the materials excreted and eliminated from the body (including urine and feces).
Nutrients
The seven major classes of nutrients are carbohydrates, fats, fiber, minerals, proteins, vitamins, and water. Nutrients can be grouped as either macronutrients or micronutrients (needed in small quantities). Carbohydrates, fats, and proteins are macronutrients, and provide energy. Water and fiber are macronutrients but do not provide energy. The micronutrients are minerals and vitamins.
The macronutrients (excluding fiber and water) provide structural material (amino acids from which proteins are built, and lipids from which cell membranes and some signaling molecules are built), and energy. Some of the structural material can also be used to generate energy internally, and in either case it is measured in Joules or kilocalories (often called "Calories" and written with a capital 'C' to distinguish them from little 'c' calories). Carbohydrates and proteins provide 17 kJ approximately (4 kcal) of energy per gram, while fats prov
Document 3:::
GRE Subject Biochemistry, Cell and Molecular Biology was a standardized exam provided by ETS (Educational Testing Service) that was discontinued in December 2016. It is a paper-based exam and there are no computer-based versions of it. ETS places this exam three times per year: once in April, once in October and once in November. Some graduate programs in the United States recommend taking this exam, while others require this exam score as a part of the application to their graduate programs. ETS sends a bulletin with a sample practice test to each candidate after registration for the exam. There are 180 questions within the biochemistry subject test.
Scores are scaled and then reported as a number between 200 and 990; however, in recent versions of the test, the maximum and minimum reported scores have been 760 (corresponding to the 99 percentile) and 320 (1 percentile) respectively. The mean score for all test takers from July, 2009, to July, 2012, was 526 with a standard deviation of 95.
After learning that test content from editions of the GRE® Biochemistry, Cell and Molecular Biology (BCM) Test has been compromised in Israel, ETS made the decision not to administer this test worldwide in 2016–17.
Content specification
Since many students who apply to graduate programs in biochemistry do so during the first half of their fourth year, the scope of most questions is largely that of the first three years of a standard American undergraduate biochemistry curriculum. A sampling of test item content is given below:
Biochemistry (36%)
A Chemical and Physical Foundations
Thermodynamics and kinetics
Redox states
Water, pH, acid-base reactions and buffers
Solutions and equilibria
Solute-solvent interactions
Chemical interactions and bonding
Chemical reaction mechanisms
B Structural Biology: Structure, Assembly, Organization and Dynamics
Small molecules
Macromolecules (e.g., nucleic acids, polysaccharides, proteins and complex lipids)
Supramolecular complexes (e.g.
Document 4:::
A scholar is a person who is a researcher or has expertise in an academic discipline. A scholar can also be an academic, who works as a professor, teacher, or researcher at a university. An academic usually holds an advanced degree or a terminal degree, such as a master's degree or a doctorate (PhD). Independent scholars and public intellectuals work outside of the academy yet may publish in academic journals and participate in scholarly public discussion.
Definitions
In contemporary English usage, the term scholar sometimes is equivalent to the term academic, and describes a university-educated individual who has achieved intellectual mastery of an academic discipline, as instructor and as researcher. Moreover, before the establishment of universities, the term scholar identified and described an intellectual person whose primary occupation was professional research. In 1847, minister Emanuel Vogel Gerhart spoke of the role of the scholar in society:
Gerhart argued that a scholar can not be focused on a single discipline, contending that knowledge of multiple disciplines is necessary to put each into context and to inform the development of each:
A 2011 examination outlined the following attributes commonly accorded to scholars as "described by many writers, with some slight variations in the definition":
Scholars may rely on the scholarly method or scholarship, a body of principles and practices used by scholars to make their claims about the world as valid and trustworthy as possible, and to make them known to the scholarly public. It is the methods that systemically advance the teaching, research, and practice of a given scholarly or academic field of study through rigorous inquiry. Scholarship is creative, can be documented, can be replicated or elaborated, and can be and is peer-reviewed through various methods.
Role in society
Scholars have generally been upheld as creditable figures of high social standing, who are engaged in work important to society.
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What does fiber help keep at normal levels in the body?
A. bloodpressure
B. hydration
C. salt and cholesterol
D. sugar and lipids
Answer:
|
|
sciq-1653
|
multiple_choice
|
The uranium series is a chain constituting one what, which encompass naturally occurring isotopes of the heaviest elements?
|
[
"nuclear decay family",
"fuel decay family",
"radioactive decay family",
"gland decay family"
] |
C
|
Relavent Documents:
Document 0:::
In nuclear science, the decay chain refers to a series of radioactive decays of different radioactive decay products as a sequential series of transformations. It is also known as a "radioactive cascade". The typical radioisotope does not decay directly to a stable state, but rather it decays to another radioisotope. Thus there is usually a series of decays until the atom has become a stable isotope, meaning that the nucleus of the atom has reached a stable state.
Decay stages are referred to by their relationship to previous or subsequent stages. A parent isotope is one that undergoes decay to form a daughter isotope. One example of this is uranium (atomic number 92) decaying into thorium (atomic number 90). The daughter isotope may be stable or it may decay to form a daughter isotope of its own. The daughter of a daughter isotope is sometimes called a granddaughter isotope. Note that the parent isotope becomes the daughter isotope, unlike in the case of a biological parent and daughter.
The time it takes for a single parent atom to decay to an atom of its daughter isotope can vary widely, not only between different parent-daughter pairs, but also randomly between identical pairings of parent and daughter isotopes. The decay of each single atom occurs spontaneously, and the decay of an initial population of identical atoms over time t, follows a decaying exponential distribution, e−λt, where λ is called a decay constant. One of the properties of an isotope is its half-life, the time by which half of an initial number of identical parent radioisotopes can be expected statistically to have decayed to their daughters, which is inversely related to λ. Half-lives have been determined in laboratories for many radioisotopes (or radionuclides). These can range from nearly instantaneous (less than 10−21 seconds) to more than 1019 years.
The intermediate stages each emit the same amount of radioactivity as the original radioisotope (i.e., there is a one-to-one relationsh
Document 1:::
Uranium in the environment is a global health concern, and comes from both natural and man-made sources. Mining, phosphates in agriculture, weapons manufacturing, and nuclear power are sources of uranium in the environment.
In the natural environment, radioactivity of uranium is generally low, but uranium is a toxic metal that can disrupt normal functioning of the kidney, brain, liver, heart, and numerous other systems. Chemical toxicity can cause public health issues when uranium is present in groundwater, especially if concentrations in food and water are increased by mining activity. The biological half-life (the average time it takes for the human body to eliminate half the amount in the body) for uranium is about 15 days.
Uranium's radioactivity can present health and environmental issues in the case of nuclear waste produced by nuclear power plants or weapons manufacturing.
Uranium is weakly radioactive and remains so because of its long physical half-life (4.468 billion years for uranium-238). The use of depleted uranium (DU) in munitions is controversial because of questions about potential long-term health effects.
Natural occurrence
Uranium is a naturally occurring element found in low levels within all rock, soil, and water. This is the highest-numbered element to be found naturally in significant quantities on earth. According to the United Nations Scientific Committee on the Effects of Atomic Radiation the normal concentration of uranium in soil is 300 μg/kg to 11.7 mg/kg.
It is considered to be more plentiful than antimony, beryllium, cadmium, gold, mercury, silver, or tungsten and is about as abundant as tin, arsenic or molybdenum. It is found in many minerals including uraninite (most common uranium ore), autunite, uranophane, torbernite, and coffinite. Significant concentrations of uranium occur in some substances such as phosphate rock deposits, and minerals such as lignite, and monazite sands in uranium-rich ores (it is recovered commercial
Document 2:::
Uranium tailings or uranium tails are a radioactive waste byproduct (tailings) of conventional uranium mining and uranium enrichment. They contain the radioactive decay products from the uranium decay chains, mainly the U-238 chain, and heavy metals. Long-term storage or disposal of tailings may pose a danger for public health and safety.
Production
Uranium mill tailings are primarily the sandy process waste material from a conventional uranium mill. Milling is the first step in making fuel for nuclear reactors from natural uranium ore. The uranium extract is transformed into yellowcake.
The raw uranium ore is brought to the surface and crushed into a fine sand. The valuable uranium-bearing minerals are then removed via heap leaching with the use of acids or bases, and the remaining radioactive sludge, called "uranium tailings", is stored in huge impoundments. A short ton (907 kg) of ore yields one to five pounds (0.45 to 2.3 kg) of uranium depending on the uranium content of the mineral. Uranium tailings can retain up to 85% of the ore's original radioactivity.
Composition
The tailings contain mainly decay products from the decay chain involving Uranium-238. Uranium tailings contain over a dozen radioactive nuclides, which are the primary hazard posed by the tailings. The most important of these are thorium-230, radium-226, radon-222 (radon gas) and the daughter isotopes of radon decay, including polonium-210. All of those are naturally occurring radioactive materials or "NORM".
Health risks
Tailings contain heavy metals and radioactive radium. Radium then decays over thousands of years and radioactive radon gas is produced. Tailings are kept in piles for long-term storage or disposal and need to be maintained and monitored for leaks over the long term.
If uranium tailings are stored aboveground and allowed to dry out, the radioactive sand can be carried great distances by the wind, entering the food chain and bodies of water. The danger posed by such sa
Document 3:::
A monoisotopic element is an element which has only a single stable isotope (nuclide). There are 26 such elements, as listed.
Stability is experimentally defined for chemical elements, as there are a number of stable nuclides with atomic numbers over ~40 which are theoretically unstable, but apparently have half-lives so long that they have not been observed either directly or indirectly (from measurement of products) to decay.
Monoisotopic elements are characterized, except in one case, by odd numbers of protons (odd Z), and even numbers of neutrons. Because of the energy gain from nuclear pairing, the odd number of protons imparts instability to isotopes of an odd Z, which in heavier elements requires a completely paired set of neutrons to offset this tendency into stability. (The five stable nuclides with odd Z and odd neutron numbers are hydrogen-2, lithium-6, boron-10, nitrogen-14, and tantalum-180m1.)
The single mononuclidic exception to the odd Z rule is beryllium; its single stable, primordial isotope, beryllium-9, has 4 protons and 5 neutrons. This element is prevented from having a stable isotope with equal numbers of neutrons and protons (beryllium-8, with 4 of each) by its instability toward alpha decay, which is favored due to the extremely tight binding of helium-4 nuclei. It is prevented from having a stable isotope with 4 protons and 6 neutrons by the very large mismatch in proton/neutron ratio for such a light element. (Nevertheless, beryllium-10 has a half-life of 1.36 million years, which is too short to be primordial, but still indicates unusual stability for a light isotope with such an imbalance.)
Differentiation from mononuclidic elements
The set of monoisotopic elements overlap but are not the same as the set of 21 mononuclidic elements, which are characterized as having essentially only one isotope (nuclide) found in nature. The reason for this is the occurrence of certain long-lived radioactive primordial nuclides in nature, which may
Document 4:::
In nuclear physics, a decay product (also known as a daughter product, daughter isotope, radio-daughter, or daughter nuclide) is the remaining nuclide left over from radioactive decay. Radioactive decay often proceeds via a sequence of steps (decay chain). For example, 238U decays to 234Th which decays to 234mPa which decays, and so on, to 206Pb (which is stable):
In this example:
234Th, 234mPa,...,206Pb are the decay products of 238U.
234Th is the daughter of the parent 238U.
234mPa (234 metastable) is the granddaughter of 238U.
These might also be referred to as the daughter products of 238U.
Decay products are important in understanding radioactive decay and the management of radioactive waste.
For elements above lead in atomic number, the decay chain typically ends with an isotope of lead or bismuth. Bismuth itself decays to thallium, but the decay is so slow as to be practically negligible.
In many cases, individual members of the decay chain are as radioactive as the parent, but far smaller in volume/mass. Thus, although uranium is not dangerously radioactive when pure, some pieces of naturally occurring pitchblende are quite dangerous owing to their radium-226 content, which is soluble and not a ceramic like the parent. Similarly, thorium gas mantles are very slightly radioactive when new, but become more radioactive after only a few months of storage as the daughters of 232Th build up.
Although it cannot be predicted whether any given atom of a radioactive substance will decay at any given time, the decay products of a radioactive substance are extremely predictable. Because of this, decay products are important to scientists in many fields who need to know the quantity or type of the parent product. Such studies are done to measure pollution levels (in and around nuclear facilities) and for other matters.
See also
Decay chain
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
The uranium series is a chain constituting one what, which encompass naturally occurring isotopes of the heaviest elements?
A. nuclear decay family
B. fuel decay family
C. radioactive decay family
D. gland decay family
Answer:
|
|
sciq-2641
|
multiple_choice
|
What is the type of cancer where bone marrow produces abnormal white blood cells?
|
[
"liver",
"kidney",
"anemia",
"leukemia"
] |
D
|
Relavent Documents:
Document 0:::
In haematology atypical localization of immature precursors (ALIP) refers to finding of atypically localized precursors (myeloblasts and promyelocytes) on bone marrow biopsy. In healthy humans, precursors are rare and are found localized near the endosteum, and consist of 1-2 cells. In some cases of myelodysplastic syndromes, immature precursors might be located in the intertrabecular region and occasionally aggregate as clusters of 3 ~ 5 cells. The presence of ALIPs is associated with worse prognosis of MDS . Recently, in bone marrow sections of patients with acute myeloid leukemia cells similar to ALIPs were defined as ALIP-like clusters. The presence of ALIP-like clusters in AML patients within remission was reported to be associated with early relapse of the disease.
Document 1:::
Bone marrow is a semi-solid tissue found within the spongy (also known as cancellous) portions of bones. In birds and mammals, bone marrow is the primary site of new blood cell production (or haematopoiesis). It is composed of hematopoietic cells, marrow adipose tissue, and supportive stromal cells. In adult humans, bone marrow is primarily located in the ribs, vertebrae, sternum, and bones of the pelvis. Bone marrow comprises approximately 5% of total body mass in healthy adult humans, such that a man weighing 73 kg (161 lbs) will have around 3.7 kg (8 lbs) of bone marrow.
Human marrow produces approximately 500 billion blood cells per day, which join the systemic circulation via permeable vasculature sinusoids within the medullary cavity. All types of hematopoietic cells, including both myeloid and lymphoid lineages, are created in bone marrow; however, lymphoid cells must migrate to other lymphoid organs (e.g. thymus) in order to complete maturation.
Bone marrow transplants can be conducted to treat severe diseases of the bone marrow, including certain forms of cancer such as leukemia. Several types of stem cells are related to bone marrow. Hematopoietic stem cells in the bone marrow can give rise to hematopoietic lineage cells, and mesenchymal stem cells, which can be isolated from the primary culture of bone marrow stroma, can give rise to bone, adipose, and cartilage tissue.
Structure
The composition of marrow is dynamic, as the mixture of cellular and non-cellular components (connective tissue) shifts with age and in response to systemic factors. In humans, marrow is colloquially characterized as "red" or "yellow" marrow (, , respectively) depending on the prevalence of hematopoietic cells vs fat cells. While the precise mechanisms underlying marrow regulation are not understood, compositional changes occur according to stereotypical patterns. For example, a newborn baby's bones exclusively contain hematopoietically active "red" marrow, and there is a pro
Document 2:::
Megakaryocyte–erythroid progenitor cells, among other blood cells, are generated as a result of hematopoiesis, which occurs in the bone marrow. Hematopoietic stem cells can differentiate into one of two progenitor cells: the common lymphoid progenitor and the common myeloid progenitor. MEPs derive from the common myeloid progenitor lineage. Megakaryocyte/erythrocyte progenitor cells must commit to becoming either platelet-producing megakaryocytes via megakaryopoiesis or erythrocyte-producing erythroblasts via erythropoiesis. Most of the blood cells produced in the bone marrow during hematopoiesis come from megakaryocyte/erythrocyte progenitor cells.
Document 3:::
Biphenotypic acute leukaemia (BAL) is an uncommon type of leukemia which arises in multipotent progenitor cells which have the ability to differentiate into both myeloid and lymphoid lineages. It is a subtype of "leukemia of ambiguous lineage".
The direct reasons leading to BAL are still not clear. BAL can be de novo or secondary to previous cytotoxic therapy. Many factors, such viruses, hereditary factors, and radiation, might have a relationship with BAL.
BAL is hard to treat. Usually the chemotherapy is chosen according to the morphology of the blast (ALL or AML). A blood-forming stem-cell transplantation is highly recommended.
About 5% of acute leukaemia cases are BAL. BAL can occur in all ages of people but occurs more in adults than in children.
Signs and symptoms
BAL has similar symptoms to other types of leukemia, but they are usually more serious.
Symptoms caused by bone marrow damage
Bruising, spotting: the reason is lack of platelets. it is very common in BAL patients, most of patients die due to the
A low level of red blood cells in the bloodstream: Because the decline of hematopoietic function, need blood transfusion therapy
Persistent fever, infection prolonged healing:
Diffuse hemorrhage: which is dangerous and might lead to death.
Symptoms caused by blood cancer cells infiltration into tissues:
Swollen lymph nodes
Joint pain
Swelling of the gums
Enlargement of the liver and spleen
Headache and vomiting: blood cancer infiltration into the wear performance of the central nervous system.
Skin lumps: Because look was slightly green, also known as the "Green tumor."
Pericardial or pleural effusion
Causes
The cause that directly leads to BAL is unclear. Exposure to radiation, chemical exposure, virus and genetics are the primary reasons proposed by researchers.
Mechanisms
The mechanism of BAL is related to several mutations. The most common abnormalities are t(9;22) and MLL gene rearrangement at 11q23.
T(9;22) affect the ABL gene at
Document 4:::
Prefibrotic primary myelofibrosis (Pre-PMF) is a rare blood cancer, classified by the World Health Organization as a distinct type of myeloproliferative neoplasm in 2016. The disease is progressive to overt primary myelofibrosis, though the rate of progression is variable and not all patients progress. Symptoms and presentation can mimic essential thrombocythemia, with the main differentiator for pre-PMF being the presence of fibrosis in the bone marrow.
Diagnosis
A bone marrow examination is required for diagnosis.
Major Criteria
The bone marrow histology should demonstrate the following:
A proliferation and atypia of the bone marrow cells that produce platelets (megakaryocytes)
Reticulin fibrosis which doesn't exceed grade 1. Grade 2 or 3 is a diagnostic criteria for primary myelofibrosis.
Age-adjusted cellularity
Proliferation of granulocytes, a type of white blood cell
Decreased production of red blood cells (erythropoiesis)
Presence of JAK2, CALR, MPL or other clonal marker.
Minor Criteria
According to the WHO, at least one of these minor criteria should be present:
Anemia which is not attributable to another condition
High white blood cell count (leukocytosis)
An enlarged spleen (splenomegaly)
LDH levels above the upper limit of the reference range.
Comparison with primary myleofibrosis
Reticulin or collagen fibrosis grade 2 or 3 is a diagnostic criteria for primary myelofibrosis.
Comparison with Essential Thrombocythemia
Both pre-PMF and Essential thrombocythemia can share diagnostic similarities, such as a proliferation of megakaryocytes and a presence of a mutation. The presence of Reticulin fibrosis in pre-PMF provides the clearest distinction between the two.
Treatment
Patients considered low risk for thrombosis or major bleeding should be observed only. Low-dose aspirin is recommended for patients without a history of thrombosis. For intermediate risk patients, symptom driven therapy for anaemia or constitutional symptoms.
For high
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is the type of cancer where bone marrow produces abnormal white blood cells?
A. liver
B. kidney
C. anemia
D. leukemia
Answer:
|
|
sciq-6302
|
multiple_choice
|
Breeding in birds occurs through an internal type of what process?
|
[
"segmentation",
"fertilization",
"stimulation",
"sedimentation"
] |
B
|
Relavent Documents:
Document 0:::
Passerine birds produce song through the vocal organ, the syrinx, which is composed of bilaterally symmetric halves located where the trachea separates into the two bronchi. Using endoscopic techniques, it has been observed that song is produced by air passing between a set of medial and lateral labia on each side of the syrinx. Song is produced bilaterally, in both halves, through each separate set of labia unless air is prevented from flowing through one side of the syrinx. Birds regulate the airflow through the syrinx with muscles—M. syringealis dorsalis and M. tracheobronchialis dorsalis—that control the medial and lateral labia in the syrinx, whose action may close off airflow. Song may, hence, be produced unilaterally through one side of the syrinx when the labia are closed in the opposite side.
Early experiments discover lateralization
Lateral dominance of the hypoglossal nerve conveying messages from the brain to the syrinx was first observed in the 1970s. This lateral dominance was determined in a breed of canary, the waterschlager canary, bred for its long and complex song, by lesioning the ipsilateral tracheosyringeal branch of the hypoglossal nerve, disabling either the left or right syrinx. The numbers of song elements in the birds’ repertoires were greatly attenuated when the left side was cut, but only modestly attenuated when the right side was disabled, indicating left syringeal dominance of song production in these canaries. Similar lateralized effects have been observed in other species such as the white-crowned sparrow (Zonotrichia leucophrys), the Java sparrow (Lonchura oryzivora) and the zebra finch (Taeniopygia guttata), which is right-side dominant. However, denervation in these birds does not entirely silence the affected syllables but creates qualitative changes in phonology and frequency.
Respiratory control and neurophysiology
In waterslager canaries, which produce most syllables using the left syrinx, as soon as a unilaterally produced
Document 1:::
The difficulty of defining or measuring intelligence in non-human animals makes the subject difficult to study scientifically in birds. In general, birds have relatively large brains compared to their head size. The visual and auditory senses are well developed in most species, though the tactile and olfactory senses are well realized only in a few groups. Birds communicate using visual signals as well as through the use of calls and song. The testing of intelligence in birds is therefore usually based on studying responses to sensory stimuli.
The corvids (ravens, crows, jays, magpies, etc.) and psittacines (parrots, macaws, and cockatoos) are often considered the most intelligent birds, and are among the most intelligent animals in general. Pigeons, finches, domestic fowl, and birds of prey have also been common subjects of intelligence studies.
Studies
Bird intelligence has been studied through several attributes and abilities. Many of these studies have been on birds such as quail, domestic fowl, and pigeons kept under captive conditions. It has, however, been noted that field studies have been limited, unlike those of the apes. Birds in the crow family (corvids) as well as parrots (psittacines) have been shown to live socially, have long developmental periods, and possess large forebrains, all of which have been hypothesized to allow for greater cognitive abilities.
Counting has traditionally been considered an ability that shows intelligence. Anecdotal evidence from the 1960s has suggested that crows can count up to 3. Researchers need to be cautious, however, and ensure that birds are not merely demonstrating the ability to subitize, or count a small number of items quickly. Some studies have suggested that crows may indeed have a true numerical ability. It has been shown that parrots can count up to 6.
Cormorants used by Chinese fishermen were given every eighth fish as a reward, and found to be able to keep count up to 7. E.H. Hoh wrote in Natural Histo
Document 2:::
Bird flight is the primary mode of locomotion used by most bird species in which birds take off and fly. Flight assists birds with feeding, breeding, avoiding predators, and migrating.
Bird flight is one of the most complex forms of locomotion in the animal kingdom. Each facet of this type of motion, including hovering, taking off, and landing, involves many complex movements. As different bird species adapted over millions of years through evolution for specific environments, prey, predators, and other needs, they developed specializations in their wings, and acquired different forms of flight.
Various theories exist about how bird flight evolved, including flight from falling or gliding (the trees down hypothesis), from running or leaping (the ground up hypothesis), from wing-assisted incline running or from proavis (pouncing) behavior.
Basic mechanics of bird flight
Lift, drag and thrust
The fundamentals of bird flight are similar to those of aircraft, in which the aerodynamic forces sustaining flight are lift, drag, and thrust. Lift force is produced by the action of air flow on the wing, which is an airfoil. The airfoil is shaped such that the air provides a net upward force on the wing, while the movement of air is directed downward. Additional net lift may come from airflow around the bird's body in some species, especially during intermittent flight while the wings are folded or semi-folded (cf. lifting body).
Aerodynamic drag is the force opposite to the direction of motion, and hence the source of energy loss in flight. The drag force can be separated into two portions, lift-induced drag, which is the inherent cost of the wing producing lift (this energy ends up primarily in the wingtip vortices), and parasitic drag, including skin friction drag from the friction of air and body surfaces and form drag from the bird's frontal area. The streamlining of bird's body and wings reduces these forces. Unlike aircraft, which have engines to produce thrust, bi
Document 3:::
Chickens (Gallus gallus domesticus) and their eggs have been used extensively as research models throughout the history of biology. Today they continue to serve as an important model for normal human biology as well as pathological disease processes.
History
Chicken embryos as a research model
Human fascination with the chicken and its egg are so deeply rooted in history that it is hard to say exactly when avian exploration began. As early as 1400 BCE, ancient Egyptians artificially incubated chicken eggs to propagate their food supply. The developing chicken in the egg first appears in written history after catching the attention of the famous Greek philosopher, Aristotle, around 350 BCE. As Aristotle opened chicken eggs at various time points of incubation, he noted how the organism changed over time. Through his writing of Historia Animalium, he introduced some of the earliest studies of embryology based on his observations of the chicken in the egg.
Aristotle recognized significant similarities between human and chicken development. From his studies of the developing chick, he was able to correctly decipher the role of the placenta and umbilical cord in the human.
Chick research of the 16th century significantly modernized ideas about human physiology. European scientists, including Ulisse Aldrovandi, Volcher Cotier and William Harvey, used the chick to demonstrate tissue differentiation, disproving the widely held belief of the time that organisms are "preformed" in their adult version and only grow larger during development. Distinct tissue areas were recognized that grew and gave rise to specific structures, including the blastoderm, or chick origin. Harvey also closely watched the development of the heart and blood and was the first to note the directional flow of blood between veins and arteries. The relatively large size of the chick as a model organism allowed scientists during this time to make these significant observations without the hel
Document 4:::
Empathy in chickens is the ability of a chicken to understand and share the feelings of another chicken. The Biotechnology and Biological Sciences Research Council's (BBSRC) Animal Welfare Initiative defines and recognizes that "...hens possess a fundamental capacity to empathise..." These empathetic responses in animals are well documented and are usually discussed along with issues related to cognition. The difference between animal cognition and animal emotion is recognized by ethicists. The specific emotional attribute of empathy in chickens has not been only investigated in terms of its existence but it has applications that have resulted in the designed reduction of stress in farm-raised poultry.
Definition
The difference between animal cognition and animal emotion is recognized by ethicists. Animal cognition covers all aspects related to the thought processes in animals. Though the topics related to cognition such as self-recognition, memory, other emotions and problem-solving have been investigated, the ability to share the emotional state of another has now been established in hens.
Chickens have the basic foundations of emotional empathy. Empathy is sometimes regarded as a form of emotional intelligence and is demonstrated when hens display signs of anxiety when they observed their chicks in distressful situations. The hens have been said to "feel their chicks' pain" and to "be affected by, and share, the emotional state of another."
Scientific evidence
A study funded by the BBSRC and published in 2011 was the first to demonstrate that chickens possess empathy and the first study to use both behavioral and physiological methods to measure these traits in birds. Chicks were exposed to a puff of air, which they find mildly distressing. During the exposure, their mother's behaviour and physiological responses were monitored non-invasively. The hens altered their behaviour by decreased preening, increased alertness, and an increased numbers of vocalisati
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Breeding in birds occurs through an internal type of what process?
A. segmentation
B. fertilization
C. stimulation
D. sedimentation
Answer:
|
|
sciq-7592
|
multiple_choice
|
Some species of fish carry their fertilized eggs in their mouth until they hatch; this is called what?
|
[
"spawning",
"mouth brooding",
"internal reproduction",
"schooling"
] |
B
|
Relavent Documents:
Document 0:::
A trophic egg is an egg whose function is not reproduction but nutrition; in essence, the trophic egg serves as food for offspring hatched from viable eggs. In most species that produce them, a trophic egg is usually an unfertilised egg. The production of trophic eggs has been observed in a highly diverse range of species, including fish, amphibians, spiders and insects. The function is not limited to any particular level of parental care, but occurs in some sub-social species of insects, the spider A. ferox, and a few other species like the frogs Leptodactylus fallax and Oophaga, and the catfish Bagrus meridionalis.
Parents of some species deliver trophic eggs directly to their offspring, whereas some other species simply produce the trophic eggs after laying the viable eggs; they then leave the trophic eggs where the viable offspring are likely to find them.
The mackerel sharks present the most extreme example of proximity between reproductive eggs and trophic eggs; their viable offspring feed on trophic eggs in utero.
Despite the diversity of species and life strategies in which trophic eggs occur, all trophic egg functions are similarly derived from similar ancestral functions, which once amounted to the sacrifice of potential future offspring in order to provide food for the survival of rival (usually earlier) offspring. In more derived examples the trophic eggs are not viable, being neither fertilised, nor even fully formed in some cases, so they do not represent actually potential offspring, although they still represent parental investment corresponding to the amount of food it took to produce them.
Morphology
Trophic eggs are not always morphologically distinct from normal reproductive eggs; however if there is no physical distinction there tends to be some kind of specialised behaviour in the way that trophic eggs are delivered by the parents.
In some beetles, trophic eggs are paler in colour and softer in texture than reproductive eggs, with a smooth
Document 1:::
External fertilization is a mode of reproduction in which a male organism's sperm fertilizes a female organism's egg outside of the female's body.
It is contrasted with internal fertilization, in which sperm are introduced via insemination and then combine with an egg inside the body of a female organism. External fertilization typically occurs in water or a moist area to facilitate the movement of sperm to the egg. The release of eggs and sperm into the water is known as spawning. In motile species, spawning females often travel to a suitable location to release their eggs.
However, sessile species are less able to move to spawning locations and must release gametes locally. Among vertebrates, external fertilization is most common in amphibians and fish. Invertebrates utilizing external fertilization are mostly benthic, sessile, or both, including animals such as coral, sea anemones, and tube-dwelling polychaetes. Benthic marine plants also use external fertilization to reproduce. Environmental factors and timing are key challenges to the success of external fertilization. While in the water, the male and female must both release gametes at similar times in order to fertilize the egg. Gametes spawned into the water may also be washed away, eaten, or damaged by external factors.
Sexual selection
Sexual selection may not seem to occur during external fertilization, but there are ways it actually can. The two types of external fertilizers are nest builders and broadcast spawners. For female nest builders, the main choice is the location of where to lay her eggs. A female can choose a nest close to the male she wants to fertilize her eggs, but there is no guarantee that the preferred male will fertilize any of the eggs. Broadcast spawners have a very weak selection, due to the randomness of releasing gametes. To look into the effect of female choice on external fertilization, an in vitro sperm competition experiment was performed. The results concluded that ther
Document 2:::
The anamniotes are an informal group of craniates comprising all fishes and amphibians, which lay their eggs in aquatic environments. They are distinguished from the amniotes (reptiles, birds and mammals), which can reproduce on dry land either by laying shelled eggs or by carrying fertilized eggs within the female. Older sources, particularly before the 20th century, may refer to anamniotes as "lower vertebrates" and amniotes as "higher vertebrates", based on the antiquated idea of the evolutionary great chain of being.
The name "anamniote" is a back-formation word created by adding the prefix an- to the word amniote, which in turn refers to the amnion, an extraembryonic membrane present during the amniotes' embryonic development which serves as a biochemical barrier that shields the embryo from environmental fluctuations by regulating the oxygen, carbon dioxide and metabolic waste exchanges and secreting a cushioning fluid. As the name suggests, anamniote embryos lack an amnion during embryonic development, and therefore rely on the presence of external water to provide oxygen and help dilute and excrete waste products (particularly ammonia) via diffusion in order for the embryo to complete development without being intoxicated by their own metabolites. This means anamniotes are almost always dependent on an aqueous (or at least very moist) environment for reproduction and are thus restricted to spawning in or near water bodies. They are also highly sensitive to chemical and temperature variation in the surrounding water, and are also more vulnerable to egg predation and parasitism.
During their life cycle, all anamniote classes pass through a completely aquatic egg stage, as well as an aquatic larval stage during which all hatchlings are gill-dependent and morphologically resemble tiny finless fish (known as a fry or a tadpole for fish and amphibians, respectively), before metamorphosizing into juvenile and adult forms (which might be aquatic, semiaquatic or e
Document 3:::
Fish go through various life stages between fertilization and adulthood. The life of a fish start as spawned eggs which hatch into immotile larvae. These larval hatchlings are not yet capable of feeding themselves and carry a yolk sac which provides stored nutrition. Before the yolk sac completely disappears, the young fish must mature enough to be able to forage independently. When they have developed to the point where they are capable of feeding by themselves, the fish are called fry. When, in addition, they have developed scales and working fins, the transition to a juvenile fish is complete and it is called a fingerling, so called as they are typically about the size of human fingers. The juvenile stage lasts until the fish is fully grown, sexually mature and interacting with other adult fish.
Growth stages
Ichthyoplankton (planktonic or drifting fish) are the eggs and larvae of fish. They are usually found in the sunlit zone of the water column, less than 200 metres deep, sometimes called the epipelagic or photic zone. Ichthyoplankton are planktonic, meaning they cannot swim effectively under their own power, but must drift with ocean currents. Fish eggs cannot swim at all, and are unambiguously planktonic. Early stage larvae swim poorly, but later stage larvae swim better and cease to be planktonic as they grow into juveniles. Fish larvae are part of the zooplankton that eat smaller plankton, while fish eggs carry their own food supply. Both eggs and larvae are themselves eaten by larger animals.
According to Kendall et al. 1984 there are three main developmental stages of fish:
Egg stage: From spawning to hatching. This stage is named so, instead of being called an embryonic stage, because there are aspects, such as those to do with the egg envelope, that are not just embryonic aspects.
Larval stage: From the eggs hatching till to when fin rays are present and the growth of protective scales has started (squamation). A key event is when the notochord
Document 4:::
Mouthbrooding, also known as oral incubation and buccal incubation, is the care given by some groups of animals to their offspring by holding them in the mouth of the parent for extended periods of time. Although mouthbrooding is performed by a variety of different animals, such as the Darwin's frog, fish are by far the most diverse mouthbrooders. Mouthbrooding has evolved independently in several different families of fish.
Mouthbrooding behaviour
Paternal mouthbrooders are species where the male looks after the eggs. Paternal mouthbrooders include the arowana, various mouthbrooding bettas and gouramies such as Betta pugnax, and sea catfish such as Ariopsis felis. Among cichlids, paternal mouthbrooding is relatively rare, but is found among some of the tilapiines, most notably the black-chin tilapia Sarotherodon melanotheron.
In the case of the maternal mouthbrooders, the female takes the eggs. Maternal mouthbrooders are found among both African and South American cichlids. African examples are the haplochromines, such as the mbuna, Astatotilapia burtoni, and the dwarf mouthbrooders Pseudocrenilabrus multicolor, and some of the tilapiines, such as Oreochromis mossambicus and Oreochromis niloticus. The South American maternal mouthbrooders are all members of the subfamily Geophaginae (commonly known as "eartheaters" on account of their substrate-sifting feeding mode) such as Gymnogeophagus balzanii and Geophagus steindachneri.
Biparental mouthbrooding occurs where both parents take some of the eggs. This is relatively rare, but is found among the cichlid genus Xenotilapia, and a single catfish, the spatula-barbled catfish (Phyllonemus typus).
Typically, after courtship, the male fertilises the eggs and then collects them in his mouth, holding onto them until they hatch. During this time he cannot feed. Among the maternal mouthbrooding cichlids, it is quite common (e.g., among the mbuna) for the male to fertilise the eggs only once they are in the female's mouth.
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Some species of fish carry their fertilized eggs in their mouth until they hatch; this is called what?
A. spawning
B. mouth brooding
C. internal reproduction
D. schooling
Answer:
|
|
sciq-7433
|
multiple_choice
|
Glaciers have been melting since what period?
|
[
"stone age",
"industrial age",
"pleistocene ice age",
"bronze age"
] |
C
|
Relavent Documents:
Document 0:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 1:::
A pre-STEM program is a course of study at any two-year college that prepares a student to transfer to a four-year school to earn a bachelor's degree in a STEM field.
Overview
The concept of a pre-STEM program is being developed to address America's need for more college-trained professionals in science, technology, engineering, and mathematics (STEM). It is an innovation meant to fill a gap at community colleges that do not have 'major' degree paths that students identify with on their way to earning an Associates degree. Students must complete a considerable amount of STEM coursework before transferring from a two-year school to a four-year school and earn a baccalaureate degree in a STEM field. Schools with a pre-STEM program are able to identify those students and support them with STEM-specific academic and career advising, increasing the student's chances of going on to earn a STEM baccalaureate degree in a timely fashion.
With over 50% of America's college-bound students starting their college career at public or private two-year school, and with a very small proportion of students who start college at a two-year school matriculating to and earning STEM degrees from four-year schools, pre-STEM programs have great potential for broadening participation in baccalaureate STEM studies.
Example programs
The effectiveness of pre-STEM programs is being investigated by a consortium of schools in Missouri: Moberly Area Community College, St. Charles Community College, Metropolitan Community College, and Truman State University.
A larger group of schools met at the Belknap Springs Meetings in October 2009 to discuss the challenges and opportunities presented by STEM-focused partnerships between 2-year and 4-year schools. Each program represented a two-year school and a four-year school that were trying to increase the number of people who earn a baccalaureate degree in a STEM area through various means, some of which were pre-STEM programs. Other methods includes
Document 2:::
Glacio-geological databases compile data on glacially associated sedimentary deposits and erosional activity from former and current ice-sheets, usually from published peer-reviewed sources. Their purposes are generally directed towards two ends: (Mode 1) compiling information about glacial landforms, which often inform about former ice-flow directions; and (Mode 2) compiling information which dates the absence or presence of ice.
These databases are used for a variety of purposes: (i) as bibliographic tools for researchers; (ii) as the quantitative basis of mapping of landforms or dates of ice presence/absence; and (iii) as quantitative databases which are used to constrain physically based mathematical models of ice-sheets.
Antarctic Ice Sheet: The AGGDB is a Mode 2 glacio-geological database for the Antarctic ice-sheet using information from around 150 published sources, covering glacial activity mainly from the past 30,000 years. It is available online, and aims to be comprehensive to the end of 2007.
British Ice Sheet: BRITICE is a Mode 1 database which aims to map all glacial landforms of Great Britain.
Eurasian Ice Sheet: DATED-1 is a Mode 2 database for the Eurasian ice-sheet. Its sister-project DATED-2 uses the information in DATED-1 to map the retreat of the Eurasian ice-sheet since the Last Glacial Maximum.
See also
Glacial landforms
Sediment
Geology
Ice sheet
Exposure Age Dating
Radio-carbon dating
Document 3:::
The Glaciogenic Reservoir Analogue Studies Project (GRASP) is a research group studying the subglacial to proglacial record of Pleistocene glacial events. It is based in the Delft University of Technology.
Introduction to glaciogenic reservoirs
Glaciogenic reservoirs are sedimentary rocks deposited under an ice sheet influence and that are involved into a gas or oil reservoir. The glacial earth system is complex to study. A large amount on past and ongoing scientific programs work(ed) on our cryosphere and generate a lot of debate about its dynamic, sustainability and behavior against climate changes. Past glaciations or ice ages record occurred several times (Timeline of glaciation) along the geological time scale. As they are hundreds of million years old, these ancient glaciations are even more hard to analyse and study. Earth at that time had a different atmosphere composition, the chemistry of the oceans was also different, life evolution on earth had also a great impact on the dynamic of these ice sheets, the continents were in a particular setting, etc. Geologists have a broad idea of all those parameters but glaciologists know that this is the combination of those setting that bring to our current ice-age.
A glacial system is able to produce a very large amount of sediment due to the tremendous erosive forces of ice at its base. Those sediments are particularly coarse-grained (principally sandstones and conglomerates) and produced in consequent volumes . For their good reservoir properties, ancient glacially-related sediments have been targeted by oil industries. They are currently massively exploited in North Africa, in the Arabic peninsula, South Africa, and few small fields are present in Asia, Australia and Northern Europe. The main ice ages concerned are the Late Ordovician glaciation (Hirnantian) and the Permo-Carboniferous glaciations.
Project objectives
Analogy is a usual geologist method, using the present day observations and project/adapt it
Document 4:::
The Institut de technologie agroalimentaire (ITA) is a collegial institute specialized in agricultural technology and food production in Quebec, Canada. The institution is composed of two campuses, one in Saint-Hyacinthe and the other in La Pocatière. The institution is managed by the Ministère de l'Agriculture, des Pêcheries et de l'Alimentation du Québec (MAPAQ).
History
The origins of the ITA date back to the 19th century. The first francophone school of agriculture was founded in 1859 in Sainte-Anne-de-la-Pocatière, while the dairy school in Saint-Hyacinthe was created in 1892, the first such institution in North America.
In 1962, the Ministry of Agriculture, Fisheries and Food of Quebec (known today in French as the Ministère de l'Agriculture, des Pêcheries et de l'Alimentation, and in 1962 as the Ministère de l'Agriculture et de la Colonisation) formed the Instituts de technologie agroalimentaire. While the La Pocatière campus was an extension of the Faculty of Agronomy of Université Laval, the Saint-Hyacinthe campus was originally a dairy school founded in 1892.
Training programs
The ITA offers a total of eight CEGEP-level training programs, which lead to a Quebec Diploma of College Studies. Most programs are offered at both campuses. They include:
Gestion et technologies d'entreprise agricole
Gestion et technologies d'entreprise agricole : Profils en production animale biologique
Technologie des productions animales
Paysage et commercialisation en horticulture ornementale
Technologie de la production horticole agroenvironnementale
Technologie du génie agromécanique
Technologie des procédés et de la qualité des aliments
Techniques équines
The ITA's programs listed above allow graduates to pursue university-level studies in related fields such as agronomy, agricultural economics, agricultural engineering, food engineering, biology, food science, and landscape architecture, amongst others.
The ITA also offers one training program in equine massage therapy,
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Glaciers have been melting since what period?
A. stone age
B. industrial age
C. pleistocene ice age
D. bronze age
Answer:
|
|
sciq-7748
|
multiple_choice
|
What relatively new addition to and highest level of the system of taxonomic classification includes just three divisions, the eukarya, the archaea, and the bacteria?
|
[
"content",
"domain",
"element",
"branch"
] |
B
|
Relavent Documents:
Document 0:::
Scientists trying to reconstruct evolutionary history have been challenged by the fact that genes can sometimes transfer between distant branches on the tree of life. This movement of genes can occur through horizontal gene transfer (HGT), scrambling the information on which biologists relied to reconstruct the phylogeny of organisms. Conversely, HGT can also help scientists to reconstruct and date the tree of life. Indeed, a gene transfer can be used as a phylogenetic marker, or as the proof of contemporaneity of the donor and recipient organisms, and as a trace of extinct biodiversity.
HGT happens very infrequently – at the individual organism level, it is highly improbable for any such event to take place. However, on the grander scale of evolutionary history, these events occur with some regularity. On one hand, this forces biologists to abandon the use of individual genes as good markers for the history of life. On the other hand, this provides an almost unexploited large source of information about the past.
Three domains of life
The three main early branches of the tree of life have been intensively studied by microbiologists because the first organisms were microorganisms. Microbiologists (led by Carl Woese) have introduced the term domain for the three main branches of this tree, where domain is a phylogenetic term similar in meaning to biological kingdom. To reconstruct this tree of life, the gene sequence encoding the small subunit of ribosomal RNA (SSU rRNA, 16s rRNA) has proven useful, and the tree (as shown in the picture) relies heavily on information from this single gene.
These three domains of life represent the main evolutionary lineages of early cellular life and currently include Bacteria, Archaea (single-celled organisms superficially similar to bacteria), and Eukarya. Eukarya includes only organisms having a well-defined nucleus, such as fungi, protists, and all organisms in the plant and animals kingdoms (see figure).
The gene most com
Document 1:::
A supergroup, in evolutionary biology, is a large group of organisms that share one common ancestor and have important defining characteristics. It is an informal, mostly arbitrary rank in biological taxonomy that is often greater than phylum or kingdom, although some supergroups are also treated as phyla.
Eukaryotic supergroups
Since the decade of 2000's, the eukaryotic tree of life (abbreviated as eToL) has been divided into 5–8 major groupings called 'supergroups'. These groupings were established after the idea that only monophyletic groups should be accepted as ranks, as an alternative to the use of paraphyletic kingdom Protista. In the early days of the eToL six traditional supergroups were considered: Amoebozoa, Opisthokonta, "Excavata", Archaeplastida, "Chromalveolata" and Rhizaria. Since then, the eToL has been rearranged profoundly, and most of these groups were found as paraphyletic or lacked defining morphological characteristics that unite their members, which makes the 'supergroup' label more arbitrary.
Document 2:::
Bergey's Manual of Systematic Bacteriology is the main resource for determining the identity of prokaryotic organisms, emphasizing bacterial species, using every characterizing aspect.
The manual was published subsequent to the Bergey's Manual of Determinative Bacteriology, though the latter is still published as a guide for identifying unknown bacteria. First published in 1923 by David Hendricks Bergey, it is used to classify bacteria based on their structural and functional attributes by arranging them into specific familial orders. However, this process has become more empirical in recent years.
The Taxonomic Outline of Bacteria and Archaea is a derived publication indexing taxon names from version two of the manual. It used to be available for free from the Bergey's manual trust website until September 2018. Michigan State University provides an alternative version that indexes NamesforLife records.
The five-volume BMSB is officially replaced by Bergey's Manual of Systematics of Archaea and Bacteria (BMSAB), a continuously-updated online book, since 2015.
Organization
The change in volume set to "Systematic Bacteriology" came in a new contract in 1980, whereupon the new style included "relationships between organisms" and had "expanded scope" overall. This new style was picked up for a four-volume set that first began publishing in 1984. The information in the volumes was separated as:
Volume 1 included information on all types of Gram-negative bacteria that were considered to have "medical and industrial importance." Volume 2 included information on all types of Gram-positive bacteria. Volume 3 deals with all of the remaining, slightly different Gram-negative bacteria, along with the Archaea. Volume 4 has information on filamentous actinomycetes and other, similar bacteria.
The current volumes differ drastically from previous volumes in that many higher taxa are not defined in terms of phenotype, but solely on 16S phylogeny, as is the case of the classes
Document 3:::
The two-domain system is a biological classification by which all organisms in the tree of life are classified into two big domains, Bacteria and Archaea. It emerged from development of knowledge of archaea diversity and challenges to the widely accepted three-domain system that defines life into Bacteria, Archaea, and Eukarya. It was preceded by the eocyte hypothesis of James A. Lake in the 1980s, which was largely superseded by the three-domain system, due to evidence at the time. Better understanding of archaea, especially of their roles in the origin of eukaryotes through symbiogenesis with bacteria, led to the revival of the eocyte hypothesis in the 2000s. The two-domain system became more widely accepted after the discovery of a large group (superphylum) of archaea called Asgard in 2017, which evidence suggests to be the evolutionary root of eukaryotes, implying that eukaryotes are members of the domain Archaea.
While the features of Asgard archaea do not directly rule out the three-domain system, the notion that eukaryotes originated from archaea and thus belong to Archaea has been strengthened by genetic and proteomic studies. Under the three-domain system, Eukarya is mainly distinguished by the presence of "eukaryotic signature proteins", that are not found in archaea and bacteria. However, Asgards contain genes that code for multiple such proteins, indicating that "eukaryotic signature proteins" originated in archaea.
Background
Classification of life into two main divisions is not a new concept, with the first such proposal by French biologist Édouard Chatton in 1938. Chatton distinguished organisms into:
Procaryotes (including bacteria)
Eucaryotes (including protozoans)
These were later named empires, and Chatton's classification as the two-empire system. Chatton used the name Eucaryotes only for protozoans, excluded other eukaryotes, and published in limited circulation so that his work was not recognised. His classification was rediscovered by
Document 4:::
The three-domain system is a biological classification introduced by Carl Woese, Otto Kandler, and Mark Wheelis in 1990 that divides cellular life forms into three domains, namely Archaea, Bacteria, and Eukarya. The key difference from earlier classifications such as the two-empire system and the five-kingdom classification is the splitting of Archaea from Bacteria as completely different organisms. It has been challenged by the two-domain system that divides organisms into Bacteria and Archaea only, as Eukaryotes are considered as one group of Archaea.
Background
Woese argued, on the basis of differences in 16S rRNA genes, that bacteria, archaea, and eukaryotes each arose separately from an ancestor with poorly developed genetic machinery, often called a progenote. To reflect these primary lines of descent, he treated each as a domain, divided into several different kingdoms. Originally his split of the prokaryotes was into Eubacteria (now Bacteria) and Archaebacteria (now Archaea). Woese initially used the term "kingdom" to refer to the three primary phylogenic groupings, and this nomenclature was widely used until the term "domain" was adopted in 1990.
Acceptance of the validity of Woese's phylogenetically valid classification was a slow process. Prominent biologists including Salvador Luria and Ernst Mayr objected to his division of the prokaryotes. Not all criticism of him was restricted to the scientific level. A decade of labor-intensive oligonucleotide cataloging left him with a reputation as "a crank", and Woese would go on to be dubbed "Microbiology's Scarred Revolutionary" by a news article printed in the journal Science in 1997. The growing amount of supporting data led the scientific community to accept the Archaea by the mid-1980s. Today, very few scientists still accept the concept of a unified Prokarya.
Classification
The three-domain system adds a level of classification (the domains) "above" the kingdoms present in the previously used five- or
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What relatively new addition to and highest level of the system of taxonomic classification includes just three divisions, the eukarya, the archaea, and the bacteria?
A. content
B. domain
C. element
D. branch
Answer:
|
|
sciq-966
|
multiple_choice
|
Nervous tissue consists of just two basic types of nerve cells: neurons and?
|
[
"glial cells",
"white blood cells",
"mammalian cells",
"red blood cells"
] |
A
|
Relavent Documents:
Document 0:::
Nervous tissue, also called neural tissue, is the main tissue component of the nervous system. The nervous system regulates and controls body functions and activity. It consists of two parts: the central nervous system (CNS) comprising the brain and spinal cord, and the peripheral nervous system (PNS) comprising the branching peripheral nerves. It is composed of neurons, also known as nerve cells, which receive and transmit impulses, and neuroglia, also known as glial cells or glia, which assist the propagation of the nerve impulse as well as provide nutrients to the neurons.
Nervous tissue is made up of different types of neurons, all of which have an axon. An axon is the long stem-like part of the cell that sends action potentials to the next cell. Bundles of axons make up the nerves in the PNS and tracts in the CNS.
Functions of the nervous system are sensory input, integration, control of muscles and glands, homeostasis, and mental activity.
Structure
Nervous tissue is composed of neurons, also called nerve cells, and neuroglial cells. Four types of neuroglia found in the CNS are astrocytes, microglial cells, ependymal cells, and oligodendrocytes. Two types of neuroglia found in the PNS are satellite glial cells and Schwann cells. In the central nervous system (CNS), the tissue types found are grey matter and white matter. The tissue is categorized by its neuronal and neuroglial components.
Components
Neurons are cells with specialized features that allow them to receive and facilitate nerve impulses, or action potentials, across their membrane to the next neuron. They possess a large cell body (soma), with cell projections called dendrites and an axon. Dendrites are thin, branching projections that receive electrochemical signaling (neurotransmitters) to create a change in voltage in the cell. Axons are long projections that carry the action potential away from the cell body toward the next neuron. The bulb-like end of the axon, called the axon terminal, i
Document 1:::
H2.00.04.4.01001: Lymphoid tissue
H2.00.05.0.00001: Muscle tissue
H2.00.05.1.00001: Smooth muscle tissue
H2.00.05.2.00001: Striated muscle tissue
H2.00.06.0.00001: Nerve tissue
H2.00.06.1.00001: Neuron
H2.00.06.2.00001: Synapse
H2.00.06.2.00001: Neuroglia
h3.01: Bones
h3.02: Joints
h3.03: Muscles
h3.04: Alimentary system
h3.05: Respiratory system
h3.06: Urinary system
h3.07: Genital system
h3.08:
Document 2:::
The neuron doctrine is the concept that the nervous system is made up of discrete individual cells, a discovery due to decisive neuro-anatomical work of Santiago Ramón y Cajal and later presented by, among others, H. Waldeyer-Hartz. The term neuron (spelled neurone in British English) was itself coined by Waldeyer as a way of identifying the cells in question. The neuron doctrine, as it became known, served to position neurons as special cases under the broader cell theory evolved some decades earlier. He appropriated the concept not from his own research but from the disparate observation of the histological work of Albert von Kölliker, Camillo Golgi, Franz Nissl, Santiago Ramón y Cajal, Auguste Forel and others.
Historical context
Theodor Schwann proposed in 1839 that the tissues of all organisms are composed of cells. Schwann was expanding on the proposal of his good friend Matthias Jakob Schleiden the previous year that all plant tissues were composed of cells. The nervous system stood as an exception. Although nerve cells had been described in tissue by numerous investigators including Jan Purkinje, Gabriel Valentin, and Robert Remak, the relationship between the nerve cells and other features such as dendrites and axons was not clear. The connections between the large cell bodies and smaller features could not be observed, and it was possible that neurofibrils would stand as an exception to cell theory as non-cellular components of living tissue. Technical limitations of microscopy and tissue preparation were largely responsible. Chromatic aberration, spherical aberration and the dependence on natural light all played a role in limiting microscope performance in the early 19th century. Tissue was typically lightly mashed in water and pressed between a glass slide and cover slip. There was also a limited number of dyes and fixatives available prior to the middle of the 19th century.
A landmark development came from Camillo Golgi who invented a silver
Document 3:::
These are timelines of brain development events in different animal species.
Mouse brain development timeline
Macaque brain development timeline
Human brain development timeline
See also
Encephalization quotient
Evolution of the brain
Neural development
External links
Translating Neurodevelopmental Time Across Mammalian Species
Vertebrate developmental biology
Embryology of nervous system
Developmental neuroscience
Document 4:::
The development of the nervous system in humans, or neural development or neurodevelopment involves the studies of embryology, developmental biology, and neuroscience to describe the cellular and molecular mechanisms by which the complex nervous system forms in humans, develops during prenatal development, and continues to develop postnatally.
Some landmarks of neural development in the embryo include the formation and differentiation of neurons from stem cell precursors (neurogenesis); the migration of immature neurons from their birthplaces in the embryo to their final positions; the outgrowth of axons from neurons and guidance of the motile growth cone through the embryo towards postsynaptic partners, the generation of synapses between these axons and their postsynaptic partners, the synaptic pruning that occurs in adolescence, and finally the lifelong changes in synapses which are thought to underlie learning and memory.
Typically, these neurodevelopmental processes can be broadly divided into two classes: activity-independent mechanisms and activity-dependent mechanisms. Activity-independent mechanisms are generally believed to occur as hardwired processes determined by genetic programs played out within individual neurons. These include differentiation, migration and axon guidance to their initial target areas. These processes are thought of as being independent of neural activity and sensory experience. Once axons reach their target areas, activity-dependent mechanisms come into play. Neural activity and sensory experience will mediate formation of new synapses, as well as synaptic plasticity, which will be responsible for refinement of the nascent neural circuits.
Development of the human brain
Overview
The central nervous system (CNS) is derived from the ectoderm—the outermost tissue layer of the embryo. In the third week of human embryonic development the neuroectoderm appears and forms the neural plate along the dorsal side of the embryo. The neural
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Nervous tissue consists of just two basic types of nerve cells: neurons and?
A. glial cells
B. white blood cells
C. mammalian cells
D. red blood cells
Answer:
|
|
sciq-7191
|
multiple_choice
|
The rate of heat transfer by radiation is largely determined by what?
|
[
"color",
"sound",
"environment",
"scope"
] |
A
|
Relavent Documents:
Document 0:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 1:::
A pre-STEM program is a course of study at any two-year college that prepares a student to transfer to a four-year school to earn a bachelor's degree in a STEM field.
Overview
The concept of a pre-STEM program is being developed to address America's need for more college-trained professionals in science, technology, engineering, and mathematics (STEM). It is an innovation meant to fill a gap at community colleges that do not have 'major' degree paths that students identify with on their way to earning an Associates degree. Students must complete a considerable amount of STEM coursework before transferring from a two-year school to a four-year school and earn a baccalaureate degree in a STEM field. Schools with a pre-STEM program are able to identify those students and support them with STEM-specific academic and career advising, increasing the student's chances of going on to earn a STEM baccalaureate degree in a timely fashion.
With over 50% of America's college-bound students starting their college career at public or private two-year school, and with a very small proportion of students who start college at a two-year school matriculating to and earning STEM degrees from four-year schools, pre-STEM programs have great potential for broadening participation in baccalaureate STEM studies.
Example programs
The effectiveness of pre-STEM programs is being investigated by a consortium of schools in Missouri: Moberly Area Community College, St. Charles Community College, Metropolitan Community College, and Truman State University.
A larger group of schools met at the Belknap Springs Meetings in October 2009 to discuss the challenges and opportunities presented by STEM-focused partnerships between 2-year and 4-year schools. Each program represented a two-year school and a four-year school that were trying to increase the number of people who earn a baccalaureate degree in a STEM area through various means, some of which were pre-STEM programs. Other methods includes
Document 2:::
Thermofluids is a branch of science and engineering encompassing four intersecting fields:
Heat transfer
Thermodynamics
Fluid mechanics
Combustion
The term is a combination of "thermo", referring to heat, and "fluids", which refers to liquids, gases and vapors. Temperature, pressure, equations of state, and transport laws all play an important role in thermofluid problems. Phase transition and chemical reactions may also be important in a thermofluid context. The subject is sometimes also referred to as "thermal fluids".
Heat transfer
Heat transfer is a discipline of thermal engineering that concerns the transfer of thermal energy from one physical system to another. Heat transfer is classified into various mechanisms, such as heat conduction, convection, thermal radiation, and phase-change transfer. Engineers also consider the transfer of mass of differing chemical species, either cold or hot, to achieve heat transfer.
Sections include :
Energy transfer by heat, work and mass
Laws of thermodynamics
Entropy
Refrigeration Techniques
Properties and nature of pure substances
Applications
Engineering : Predicting and analysing the performance of machines
Thermodynamics
Thermodynamics is the science of energy conversion involving heat and other forms of energy, most notably mechanical work. It studies and interrelates the macroscopic variables, such as temperature, volume and pressure, which describe physical, thermodynamic systems.
Fluid mechanics
Fluid Mechanics the study of the physical forces at work during fluid flow. Fluid mechanics can be divided into fluid kinematics, the study of fluid motion, and fluid kinetics, the study of the effect of forces on fluid motion. Fluid mechanics can further be divided into fluid statics, the study of fluids at rest, and fluid dynamics, the study of fluids in motion. Some of its more interesting concepts include momentum and reactive forces in fluid flow and fluid machinery theory and performance.
Sections include:
Flu
Document 3:::
The transmission curve or transmission characteristic is the mathematical function or graph that describes the transmission fraction of an optical or electronic filter as a function of frequency or wavelength. It is an instance of a transfer function but, unlike the case of, for example, an amplifier, output never exceeds input (maximum transmission is 100%). The term is often used in commerce, science, and technology to characterise filters.
The term has also long been used in fields such as geophysics and astronomy to characterise the properties of regions through which radiation passes, such as the ionosphere.
See also
Electronic filter — examples of transmission characteristics of electronic filters
Document 4:::
In the study of heat transfer, absorptance of the surface of a material is its effectiveness in absorbing radiant energy. It is the ratio of the absorbed to the incident radiant power.
Mathematical definitions
Hemispherical absorptance
Hemispherical absorptance of a surface, denoted is defined as
where
is the radiant flux absorbed by that surface;
is the radiant flux received by that surface.
Spectral hemispherical absorptance
Spectral hemispherical absorptance in frequency and spectral hemispherical absorptance in wavelength of a surface, denoted and respectively, are defined as
where
is the spectral radiant flux in frequency absorbed by that surface;
is the spectral radiant flux in frequency received by that surface;
is the spectral radiant flux in wavelength absorbed by that surface;
is the spectral radiant flux in wavelength received by that surface.
Directional absorptance
Directional absorptance of a surface, denoted , is defined as
where
is the radiance absorbed by that surface;
is the radiance received by that surface.
Spectral directional absorptance
Spectral directional absorptance in frequency and spectral directional absorptance in wavelength of a surface, denoted and respectively, are defined as
where
is the spectral radiance in frequency absorbed by that surface;
is the spectral radiance received by that surface;
is the spectral radiance in wavelength absorbed by that surface;
is the spectral radiance in wavelength received by that surface.
Other radiometric coefficients
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
The rate of heat transfer by radiation is largely determined by what?
A. color
B. sound
C. environment
D. scope
Answer:
|
|
ai2_arc-709
|
multiple_choice
|
Which condition is associated with most warm fronts?
|
[
"tornado formations",
"low temperatures",
"violent air mass collisions",
"cloud formations bringing precipitation"
] |
D
|
Relavent Documents:
Document 0:::
This is a list of meteorology topics. The terms relate to meteorology, the interdisciplinary scientific study of the atmosphere that focuses on weather processes and forecasting. (see also: List of meteorological phenomena)
A
advection
aeroacoustics
aerobiology
aerography (meteorology)
aerology
air parcel (in meteorology)
air quality index (AQI)
airshed (in meteorology)
American Geophysical Union (AGU)
American Meteorological Society (AMS)
anabatic wind
anemometer
annular hurricane
anticyclone (in meteorology)
apparent wind
Atlantic Oceanographic and Meteorological Laboratory (AOML)
Atlantic hurricane season
atmometer
atmosphere
Atmospheric Model Intercomparison Project (AMIP)
Atmospheric Radiation Measurement (ARM)
(atmospheric boundary layer [ABL]) planetary boundary layer (PBL)
atmospheric chemistry
atmospheric circulation
atmospheric convection
atmospheric dispersion modeling
atmospheric electricity
atmospheric icing
atmospheric physics
atmospheric pressure
atmospheric sciences
atmospheric stratification
atmospheric thermodynamics
atmospheric window (see under Threats)
B
ball lightning
balloon (aircraft)
baroclinity
barotropity
barometer ("to measure atmospheric pressure")
berg wind
biometeorology
blizzard
bomb (meteorology)
buoyancy
Bureau of Meteorology (in Australia)
C
Canada Weather Extremes
Canadian Hurricane Centre (CHC)
Cape Verde-type hurricane
capping inversion (in meteorology) (see "severe thunderstorms" in paragraph 5)
carbon cycle
carbon fixation
carbon flux
carbon monoxide (see under Atmospheric presence)
ceiling balloon ("to determine the height of the base of clouds above ground level")
ceilometer ("to determine the height of a cloud base")
celestial coordinate system
celestial equator
celestial horizon (rational horizon)
celestial navigation (astronavigation)
celestial pole
Celsius
Center for Analysis and Prediction of Storms (CAPS) (in Oklahoma in the US)
Center for the Study o
Document 1:::
The following outline is provided as an overview of and topical guide to the field of Meteorology.
Meteorology The interdisciplinary, scientific study of the Earth's atmosphere with the primary focus being to understand, explain, and forecast weather events. Meteorology, is applied to and employed by a wide variety of diverse fields, including the military, energy production, transport, agriculture, and construction.
Essence of meteorology
Meteorology
Climate – the average and variations of weather in a region over long periods of time.
Meteorology – the interdisciplinary scientific study of the atmosphere that focuses on weather processes and forecasting (in contrast with climatology).
Weather – the set of all the phenomena in a given atmosphere at a given time.
Branches of meteorology
Microscale meteorology – the study of atmospheric phenomena about 1 km or less, smaller than mesoscale, including small and generally fleeting cloud "puffs" and other small cloud features
Mesoscale meteorology – the study of weather systems about 5 kilometers to several hundred kilometers, smaller than synoptic scale systems but larger than microscale and storm-scale cumulus systems, skjjoch as sea breezes, squall lines, and mesoscale convective complexes
Synoptic scale meteorology – is a horizontal length scale of the order of 1000 kilometres (about 620 miles) or more
Methods in meteorology
Surface weather analysis – a special type of weather map that provides a view of weather elements over a geographical area at a specified time based on information from ground-based weather stations
Weather forecasting
Weather forecasting – the application of science and technology to predict the state of the atmosphere for a future time and a given location
Data collection
Pilot Reports
Weather maps
Weather map
Surface weather analysis
Forecasts and reporting of
Atmospheric pressure
Dew point
High-pressure area
Ice
Black ice
Frost
Low-pressure area
Precipitation
Document 2:::
A Blue Norther, also known as a Texas Norther, is a fast moving cold front marked by a rapid drop in temperature, strong winds, and dark blue or "black" skies. The cold front originates from the north, hence the "norther", and can send temperatures plummeting by 20 or 30 degrees in merely minutes.
Effects
The Midwestern United States lacks natural geographic barriers to protect itself from the frigid winter air masses that originate in Canada and the arctic. Multiple times per year conditions will become favorable to push severe cold fronts as far south as Texas, bringing sleet and snow and causing the windchill to plunge into the teens. Depending on the time of year, high temperatures that immediately precede a Texas Norther can reach 85 °F (29°C) or even 90 °F (32°C) under bright sunlight in nearly-calm conditions before the cold front approaches.
However, most Blue Northers don't advance as far south as Mexico, and even the most severe examples typically reach their apex midway through Texas. For example, cities in North Texas, like Dallas, experience drastically more Blue Northers than cities along the Gulf of Mexico, like Houston. As a city is struck by a Blue Norther, its temperatures can be 30 to 50 degrees colder than neighboring cities that are only a few miles away that have not yet been struck. Blue Northers can be dangerous due to their volatile temperature swings which catch some people unprepared.
Frequency
Blue Northers occur multiple times per year. They are usually recorded between the months of November and March, although they have been recorded less frequently in October and April as well. The Blue Norther phenomenon is especially common in November, when the last vestiges of autumn are still clinging to life. One of the most famous Blue Northers was the Great Blue Norther of November 11, 1911, which spawned multiple tornadoes and dropped temperatures 40 degrees in only 15 minutes and 67 degrees in 10 hours, a world record.
See also
Weath
Document 3:::
The Todd Weather Folios are a collection of continental Australian synoptic charts that were published from 1879 to 1909.
The charts were created by Sir Charles Todd's office at the Adelaide Observatory. In addition to the charts, the folios include clippings of newspaper articles and telegraphic and handwritten information about the weather. The area covered is mainly the east and south-east of Australia, with occasional reference to other parts of Australasia and the world.
The maps are bound into approximately six-month folios, 63 of which cover the entire period. There are approximately 10,000 continental weather maps along with 750 rainfall maps for South Australia, 10 million printed words of news text, and innumerable handwritten observations and correspondences about the weather.
The folios are an earlier part of the National Archives of Australia listed collection series number D1384.
The History of the Folios
With the advent of the telegraph it was possible to simultaneously collect data, such as surface temperature and sea-level pressure, to draw synoptic weather charts. With Charles Todd's appointment as Postmaster General to the Colony, he trained not only his telegraph operators, but also his postmasters as weather observers. These observers provided valuable data points that, in combination with telegraphed observations from the other colonies (including New Zealand), showed the development and progress of weather activity across a large part of the Southern Hemisphere. Todd's best known feat was his construction management of the Overland Telegraph from Adelaide to Port Darwin. This line of communication was critical to his capacity to create continent-wide synoptic charts as the telegraphic observations from the Outback enabled the connection of data points on the east coast of Australia with similar data points on the west and southern coasts. These continent-scale isobaric lines allowed Todd and his staff to draw synoptic charts that in the
Document 4:::
A temperature gradient is a physical quantity that describes in which direction and at what rate the temperature changes the most rapidly around a particular location. The temperature gradient is a dimensional quantity expressed in units of degrees (on a particular temperature scale) per unit length. The SI unit is kelvin per meter (K/m).
Temperature gradients in the atmosphere are important in the atmospheric sciences (meteorology, climatology and related fields).
Mathematical description
Assuming that the temperature T is an intensive quantity, i.e., a single-valued, continuous and differentiable function of three-dimensional space (often called a scalar field), i.e., that
where x, y and z are the coordinates of the location of interest, then the temperature gradient is the vector quantity defined as
Physical processes
Climatology
On a global and annual basis, the dynamics of the atmosphere (and the oceans) can be understood as attempting to reduce the large difference of temperature between the poles and the equator by redistributing warm and cold air and water, known as Earth's heat engine.
Meteorology
Differences in air temperature between different locations are critical in weather forecasting and climate. The absorption of solar light at or near the planetary surface increases the temperature gradient and may result in convection (a major process of cloud formation, often associated with precipitation).
Meteorological fronts are regions where the horizontal temperature gradient may reach relatively high values, as these are boundaries between air masses with rather distinct properties.
Clearly, the temperature gradient may change substantially in time, as a result of diurnal or seasonal heating and cooling for instance. This most likely happens during an inversion. For instance, during the day the temperature at ground level may be cold while it's warmer up in the atmosphere. As the day shifts over to night the temperature might drop rapidly while
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Which condition is associated with most warm fronts?
A. tornado formations
B. low temperatures
C. violent air mass collisions
D. cloud formations bringing precipitation
Answer:
|
|
sciq-6294
|
multiple_choice
|
Sounds, chemicals, and visual cues are examples of what between animals?
|
[
"migration",
"communication",
"procreation",
"circulation"
] |
B
|
Relavent Documents:
Document 0:::
Communication occurs when an animal produces a signal and uses it to influences the behaviour of another animal. A signal can be any behavioural, structural or physiological trait that has evolved specifically to carry information about the sender and/or the external environment and to stimulate the sensory system of the receiver to change their behaviour. A signal is different from a cue in that cues are informational traits that have not been selected for communication purposes. For example, if an alerted bird gives a warning call to a predator and causes the predator to give up the hunt, the bird is using the sound as a signal to communicate its awareness to the predator. On the other hand, if a rat forages in the leaves and makes a sound that attracts a predator, the sound itself is a cue and the interaction is not considered a communication attempt.
Air and water have different physical properties which lead to different velocity and clarity of the signal transmission process during communication. This means that common understanding of communication mechanisms and structures of terrestrial animals cannot be applied to aquatic animals. For example, a horse can sniff the air to detect pheromones but a fish which is surrounded by water will need a different method to detect chemicals.
Aquatic animals can communicate through various signal modalities including visual, auditory, tactile, chemical and electrical signals. Communication using any of these forms requires specialised signal producing and detecting organs. Thus, the structure, distribution and mechanism of these sensory systems vary amongst different classes and species of aquatic animals and they also differ greatly to those of terrestrial animals.
The basic functions of communication in aquatic animals are similar to those of terrestrial animals. In general, communication can be used to facilitate social recognition and aggregation, to locate, attract and evaluate mating partners and to engage in te
Document 1:::
Animal communication is the transfer of information from one or a group of animals (sender or senders) to one or more other animals (receiver or receivers) that affects the current or future behavior of the receivers. Information may be sent intentionally, as in a courtship display, or unintentionally, as in the transfer of scent from predator to prey with kairomones. Information may be transferred to an "audience" of several receivers. Animal communication is a rapidly growing area of study in disciplines including animal behavior, sociology, neurology and animal cognition. Many aspects of animal behavior, such as symbolic name use, emotional expression, learning and sexual behavior, are being understood in new ways.
When the information from the sender changes the behavior of a receiver, the information is referred to as a "signal". Signalling theory predicts that for a signal to be maintained in the population, both the sender and receiver should usually receive some benefit from the interaction. Signal production by senders and the perception and subsequent response of receivers are thought to coevolve. Signals often involve multiple mechanisms, e.g. both visual and auditory, and for a signal to be understood the coordinated behaviour of both sender and receiver require careful study.
Animal languages
The sounds animals make are important because they communicate the animals' state. Some animals species have been taught simple versions of human languages. Animals can use, for example, electrolocation and echolocation to communicate about prey and location. Keski-Korsu suggests a challenge of human/animal communication is that humans don't recognize animals as self aware and deliberately communicating.
Modes
Visual
Gestures: Most animals understand communication through a visual display of distinctive body parts or bodily movements. Animals will reveal or accentuate a body part to relay certain information. The parent herring gull displays its bright yell
Document 2:::
Eric Michael Johnson (20 March 2014). The Gap: The Science of What Separates Us From Other Animals, by Thomas Suddendorf. The Times Higher Education.
Document 3:::
Magnetoreception is a sense which allows an organism to detect the Earth's magnetic field. Animals with this sense include some arthropods, molluscs, and vertebrates (fish, amphibians, reptiles, birds, and mammals). The sense is mainly used for orientation and navigation, but it may help some animals to form regional maps. Experiments on migratory birds provide evidence that they make use of a cryptochrome protein in the eye, relying on the quantum radical pair mechanism to perceive magnetic fields. This effect is extremely sensitive to weak magnetic fields, and readily disturbed by radio-frequency interference, unlike a conventional iron compass.
Birds have iron-containing materials in their upper beaks. There is some evidence that this provides a magnetic sense, mediated by the trigeminal nerve, but the mechanism is unknown.
Cartilaginous fish including sharks and stingrays can detect small variations in electric potential with their electroreceptive organs, the ampullae of Lorenzini. These appear to be able to detect magnetic fields by induction. There is some evidence that these fish use magnetic fields in navigation.
History
Biologists have long wondered whether migrating animals such as birds and sea turtles have an inbuilt magnetic compass, enabling them to navigate using the Earth's magnetic field. Until late in the 20th century, evidence for this was essentially only behavioural: many experiments demonstrated that animals could indeed derive information from the magnetic field around them, but gave no indication of the mechanism. In 1972, Roswitha and Wolfgang Wiltschko showed that migratory birds responded to the direction and inclination (dip) of the magnetic field. In 1977, M. M. Walker and colleagues identified iron-based (magnetite) magnetoreceptors in the snouts of rainbow trout. In 2003, G. Fleissner and colleagues found iron-based receptors in the upper beaks of homing pigeons, both seemingly connected to the animal's trigeminal nerve. Resear
Document 4:::
Structures built by non-human animals, often called animal architecture, are common in many species. Examples of animal structures include termite mounds, ant hills, wasp and beehives, burrow complexes, beaver dams, elaborate nests of birds, and webs of spiders.
Often, these structures incorporate sophisticated features such as temperature regulation, traps, bait, ventilation, special-purpose chambers and many other features. They may be created by individuals or complex societies of social animals with different forms carrying out specialized roles. These constructions may arise from complex building behaviour of animals such as in the case of night-time nests for chimpanzees, from inbuilt neural responses, which feature prominently in the construction of bird songs, or triggered by hormone release as in the case of domestic sows, or as emergent properties from simple instinctive responses and interactions, as exhibited by termites, or combinations of these. The process of building such structures may involve learning and communication, and in some cases, even aesthetics. Tool use may also be involved in building structures by animals.
Building behaviour is common in many non-human mammals, birds, insects and arachnids. It is also seen in a few species of fish, reptiles, amphibians, molluscs, urochordates, crustaceans, annelids and some other arthropods. It is virtually absent from all the other animal phyla.
Functions
Animals create structures primarily for three reasons:
to create protected habitats, i.e. homes.
to catch prey and for foraging, i.e. traps.
for communication between members of the species (intra-specific communication), i.e. display.
Animals primarily build habitat for protection from extreme temperatures and from predation. Constructed structures raise physical problems which need to be resolved, such as humidity control or ventilation, which increases the complexity of the structure. Over time, through evolution, animals use shelters for ot
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Sounds, chemicals, and visual cues are examples of what between animals?
A. migration
B. communication
C. procreation
D. circulation
Answer:
|
|
sciq-1147
|
multiple_choice
|
Ice caps are found only in greenland and which other place?
|
[
"antarctica",
"Siberia",
"the tundra",
"rainforests"
] |
A
|
Relavent Documents:
Document 0:::
Several universities have designed interdisciplinary courses with a focus on human biology at the undergraduate level. There is a wide variation in emphasis ranging from business, social studies, public policy, healthcare and pharmaceutical research.
Americas
Human Biology major at Stanford University, Palo Alto (since 1970)
Stanford's Human Biology Program is an undergraduate major; it integrates the natural and social sciences in the study of human beings. It is interdisciplinary and policy-oriented and was founded in 1970 by a group of Stanford faculty (Professors Dornbusch, Ehrlich, Hamburg, Hastorf, Kennedy, Kretchmer, Lederberg, and Pittendrigh). It is a very popular major and alumni have gone to post-graduate education, medical school, law, business and government.
Human and Social Biology (Caribbean)
Human and Social Biology is a Level 4 & 5 subject in the secondary and post-secondary schools in the Caribbean and is optional for the Caribbean Secondary Education Certification (CSEC) which is equivalent to Ordinary Level (O-Level) under the British school system. The syllabus centers on structure and functioning (anatomy, physiology, biochemistry) of human body and the relevance to human health with Caribbean-specific experience. The syllabus is organized under five main sections: Living organisms and the environment, life processes, heredity and variation, disease and its impact on humans, the impact of human activities on the environment.
Human Biology Program at University of Toronto
The University of Toronto offers an undergraduate program in Human Biology that is jointly offered by the Faculty of Arts & Science and the Faculty of Medicine. The program offers several major and specialist options in: human biology, neuroscience, health & disease, global health, and fundamental genetics and its applications.
Asia
BSc (Honours) Human Biology at All India Institute of Medical Sciences, New Delhi (1980–2002)
BSc (honours) Human Biology at AIIMS (New
Document 1:::
The Institut de technologie agroalimentaire (ITA) is a collegial institute specialized in agricultural technology and food production in Quebec, Canada. The institution is composed of two campuses, one in Saint-Hyacinthe and the other in La Pocatière. The institution is managed by the Ministère de l'Agriculture, des Pêcheries et de l'Alimentation du Québec (MAPAQ).
History
The origins of the ITA date back to the 19th century. The first francophone school of agriculture was founded in 1859 in Sainte-Anne-de-la-Pocatière, while the dairy school in Saint-Hyacinthe was created in 1892, the first such institution in North America.
In 1962, the Ministry of Agriculture, Fisheries and Food of Quebec (known today in French as the Ministère de l'Agriculture, des Pêcheries et de l'Alimentation, and in 1962 as the Ministère de l'Agriculture et de la Colonisation) formed the Instituts de technologie agroalimentaire. While the La Pocatière campus was an extension of the Faculty of Agronomy of Université Laval, the Saint-Hyacinthe campus was originally a dairy school founded in 1892.
Training programs
The ITA offers a total of eight CEGEP-level training programs, which lead to a Quebec Diploma of College Studies. Most programs are offered at both campuses. They include:
Gestion et technologies d'entreprise agricole
Gestion et technologies d'entreprise agricole : Profils en production animale biologique
Technologie des productions animales
Paysage et commercialisation en horticulture ornementale
Technologie de la production horticole agroenvironnementale
Technologie du génie agromécanique
Technologie des procédés et de la qualité des aliments
Techniques équines
The ITA's programs listed above allow graduates to pursue university-level studies in related fields such as agronomy, agricultural economics, agricultural engineering, food engineering, biology, food science, and landscape architecture, amongst others.
The ITA also offers one training program in equine massage therapy,
Document 2:::
Tech City College (Formerly STEM Academy) is a free school sixth form located in the Islington area of the London Borough of Islington, England.
It originally opened in September 2013, as STEM Academy Tech City and specialised in Science, Technology, Engineering and Maths (STEM) and the Creative Application of Maths and Science. In September 2015, STEM Academy joined the Aspirations Academy Trust was renamed Tech City College. Tech City College offers A-levels and BTECs as programmes of study for students.
Document 3:::
Female education in STEM refers to child and adult female representation in the educational fields of science, technology, engineering, and mathematics (STEM). In 2017, 33% of students in STEM fields were women.
The organization UNESCO has stated that this gender disparity is due to discrimination, biases, social norms and expectations that influence the quality of education women receive and the subjects they study. UNESCO also believes that having more women in STEM fields is desirable because it would help bring about sustainable development.
Current status of girls and women in STEM education
Overall trends in STEM education
Gender differences in STEM education participation are already visible in early childhood care and education in science- and math-related play, and become more pronounced at higher levels of education. Girls appear to lose interest in STEM subjects with age, particularly between early and late adolescence. This decreased interest affects participation in advanced studies at the secondary level and in higher education. Female students represent 35% of all students enrolled in STEM-related fields of study at this level globally. Differences are also observed by disciplines, with female enrollment lowest in engineering, manufacturing and construction, natural science, mathematics and statistics and ICT fields. Significant regional and country differences in female representation in STEM studies can be observed, though, suggesting the presence of contextual factors affecting girls’ and women's engagement in these fields. Women leave STEM disciplines in disproportionate numbers during their higher education studies, in their transition to the world of work and even in their career cycle.
Learning achievement in STEM education
Data on gender differences in learning achievement present a complex picture, depending on what is measured (subject, knowledge acquisition against knowledge application), the level of education/age of students, and
Document 4:::
Math in Moscow (MiM) is a one-semester study abroad program for North American and European undergraduates held at the Independent University of Moscow (IUM) in Moscow, Russia. The program consists mainly of math courses that are taught in English. The program was first offered in 2001, and since 2008 has been run jointly by the Independent University of Moscow, Moscow Center for Continuous Mathematical Education, and the Higher School of Economics (HSE).
The program has hosted over 200 participants, including students from Harvard, Princeton, MIT, Harvey Mudd, Berkeley, Cornell, Yale, Wesleyan, McGill, Toronto, and Montreal.
Features
The MiM semester lasts fifteen weeks with fourteen weeks of teaching and one week of exams. Math courses are lectured by professors of the Independent University of Moscow and the Math Department of National Research University Higher School of Economics. The cultural elements of the program include organized trips to Saint Petersburg and to the Golden Ring towns of Vladimir and Suzdal. Students live in the dormitory of the Higher School of Economics.
Each semester the American Mathematical Society offers up to five "Math in Moscow" scholarships provided by the National Science Foundation to US undergraduates, and the Canadian Mathematical Society offers one or two NSERC scholarships to Canadian students.
The program is often reviewed favorably by North American students and their departments.
Curriculum
The primary curriculum is entirely mathematical, drawing from every major field of mathematics. All courses are taught jointly with the Higher School of Economics, and are often attended by students from the HSE master's program. Likewise, Math in Moscow participants may attend open lectures and seminars at the Higher School of Economics. The Math in Moscow courses are formally divided into three groups according to the expected prerequisites, however admitted students may choose to attend whichever and as many courses as they
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Ice caps are found only in greenland and which other place?
A. antarctica
B. Siberia
C. the tundra
D. rainforests
Answer:
|
|
sciq-8545
|
multiple_choice
|
Which muscle is broad, triangular and located on the inferior part of the back where it inserts into the aponeurosis?
|
[
"Synergistic dorsi",
"bronchus dorsi",
"dorsolateral dorsi",
"latissimus dorsi"
] |
D
|
Relavent Documents:
Document 0:::
The latissimus dorsi () is a large, flat muscle on the back that stretches to the sides, behind the arm, and is partly covered by the trapezius on the back near the midline. The word latissimus dorsi (plural: latissimi dorsi) comes from Latin and means "broadest [muscle] of the back", from "latissimus" ()' and "dorsum" (). The pair of muscles are commonly known as "lats", especially among bodybuilders. The latissimus dorsi is the largest muscle in the upper body.
The latissimus dorsi is responsible for extension, adduction, transverse extension also known as horizontal abduction (or horizontal extension), flexion from an extended position, and (medial) internal rotation of the shoulder joint. It also has a synergistic role in extension and lateral flexion of the lumbar spine.
Due to bypassing the scapulothoracic joints and attaching directly to the spine, the actions the latissimi dorsi have on moving the arms can also influence the movement of the scapulae, such as their downward rotation during a pull up.
Structure
Variations
The number of dorsal vertebrae to which it is attached varies from four to eight; the number of costal attachments varies; muscle fibers may or may not reach the crest of the ilium.
A muscle slip, the axillary arch, varying from 7 to 10 cm in length, and from 5 to 15 mm in breadth, occasionally springs from the upper edge of the latissimus dorsi about the middle of the posterior fold of the axilla, and crosses the axilla in front of the axillary vessels and nerves, to join the under surface of the tendon of the pectoralis major, the coracobrachialis, or the fascia over the biceps brachii. This axillary arch crosses the axillary artery, just above the spot usually selected for the application of a ligature, and may mislead a surgeon. It is present in about 7% of the population and may be easily recognized by the transverse direction of its fibers. Guy et al. extensively described this muscular variant using MRI data and positively corre
Document 1:::
The eminentia conchae is crossed by a vertical ridge, the ponticulus, which gives attachment to the Auricularis posterior muscle.
Document 2:::
Pectoral muscles (colloquially referred to as "pecs") are the muscles that connect the front of the human chest with the bones of the upper arm and shoulder. This region contains four muscles that provide movements to the upper limbs or ribs.
Pectoralis major is a thick, fan-shaped or triangular convergent muscle, which makes up the bulk of the chest muscle. It lies under the breast. It serves to flex, extend, and rotate the humerus, the long bone of the upper arm.
Pectoralis minor is a thin, triangular muscle located beneath the pectoralis major. It attaches to the ribs, and serves to stabilize the scapula, the large bone of the shoulder.
The pectoral fascia is a thin layer of tissue over the pectoralis major, extending toward the latissimus dorsi muscle on the back.
Along with the pectoralis major and pectoralis minor, the subclavius muscle forms the axilla or armpit. The subclavius moves the shoulder downward and forward.
Serratus anterior is another muscle on the front of the chest. It moves the scapula forward around the torso, as when throwing a punch.
Between the ribs are various groups of intercostal muscles, which help with breathing.
Document 3:::
The posterior triangle (or lateral cervical region) is a region of the neck.
Boundaries
The posterior triangle has the following boundaries:
Apex: Union of the sternocleidomastoid and the trapezius muscles at the superior nuchal line of the occipital bone
Anteriorly: Posterior border of the sternocleidomastoideus
Posteriorly: Anterior border of the trapezius
Inferiorly: Middle one third of the clavicle
Roof: Investing layer of the deep cervical fascia
Floor: (From superior to inferior)
1) M. semispinalis capitis
2) M. splenius capitis
3) M. levator scapulae
4) M. scalenus posterior
5) M. scalenus medius
Divisions
The posterior triangle is crossed, about 2.5 cm above the clavicle, by the inferior belly of the omohyoid muscle, which divides the space into two triangles:
an upper or occipital triangle
a lower or subclavian triangle (or supraclavicular triangle)
Contents
A) Nerves and plexuses:
Spinal accessory nerve (Cranial Nerve XI)
Branches of cervical plexus
Roots and trunks of brachial plexus
Phrenic nerve (C3,4,5)
B) Vessels:
Subclavian artery (Third part)
Transverse cervical artery
Suprascapular artery
Terminal part of external jugular vein
C) Lymph nodes:
Occipital
Supraclavicular
D) Muscles:
Inferior belly of omohyoid muscle
Anterior Scalene
Middle Scalene
Posterior Scalene
Levator Scapulae Muscle
Splenius
Clinical significance
The accessory nerve (CN XI) is particularly vulnerable to damage during lymph node biopsy. Damage results in an inability to shrug the shoulders or raise the arm above the head, particularly due to compromised trapezius muscle innervation.
The external jugular vein's superficial location within the posterior triangle also makes it vulnerable to injury.
See also
Anterior triangle of the neck
Document 4:::
The obliquus capitis inferior muscle () is a muscle in the upper back of the neck. It is one of the suboccipital muscles. Its inferior attachment is at the spinous process of the axis; its superior attachment is at the transverse process of the atlas. It is innervated by the suboccipital nerve (the posterior ramus of first cervical spinal nerve). The muscle rotates the head to its side.
Despite what its name suggest, it is the only capitis (Latin: "head") muscle that does not actually attach to the skull.
Anatomy
The obliquus capitis inferior is one of the suboccipital muscles (and the only one of these to have no attachment to the skull). It is larger than the obliquus capitis superior muscle. It forms the inferolateral boundary of the suboccipital triangle.
The muscle extends laterally and somewht superiorly from its inferior attachment to its superior attachment.
Attachments
its inferior attachment is at the lateral external aspect of the bifid spinous process of the axis (cervical vertebra C2) (inferior to the attachment of the rectus capitis posterior major muscle) and the lamina of the axis.
Its superior attachment is at (the inferoposterior aspect of) the transverse process of the atlas (cervical vertebra C1).
Innervation
The muscle receives motor innervation from the suboccipital nerve (the posterior ramus of cervical spinal nerve C1).
Relations
It lies deep to the semispinalis capitis and trapezius muscles.
Actions/movements
The muscle acts to rotate the atlas (and thus the head) ipsilaterally. It acts together with the rectus capitis posterior major muscle.
Function
The muscle is responsible for rotation of the head and first cervical vertebra (atlanto-axial joint).
The obliquus capitis inferior muscle, like the other suboccipital muscles, has an important role in proprioception. This muscle has a very high density of Golgi organs and muscle spindles which accounts for this. It is believed that proprioception may be the primary role of the
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Which muscle is broad, triangular and located on the inferior part of the back where it inserts into the aponeurosis?
A. Synergistic dorsi
B. bronchus dorsi
C. dorsolateral dorsi
D. latissimus dorsi
Answer:
|
|
sciq-5396
|
multiple_choice
|
Which property changes when a glass breaks?
|
[
"State",
"Chemical",
"density",
"its physical property"
] |
D
|
Relavent Documents:
Document 0:::
Glass typically has a tensile strength of . However, the theoretical upper bound on its strength is orders of magnitude higher: . This high value is due to the strong chemical Si–O bonds of silicon dioxide. Imperfections of the glass, such as bubbles, and in particular surface flaws, such as scratches, have a great effect on the strength of glass and decrease it even more than for other brittle materials. The chemical composition of the glass also impacts its tensile strength. The processes of thermal and chemical toughening can increase the tensile strength of glass.
Glass has a compressive strength of .
Strength of glass fiber
Glass fibers have a much higher tensile strength than regular glass (200-500 times stronger than regular glass). This is due to the reduction of flaws in glass fibers and that glass fibers have a small cross sectional area, constraining maximum defect size (Size effect on structural strength).
Strength of fiberglass
Fiberglass's strength depends on the type. S-glass has a strength of while E-glass and C-glass have a strength of .
Hardness
Glass has a hardness of 6.5 on the Mohs scale of mineral hardness.
Document 1:::
This is a list of some physical properties of common glasses. Unless otherwise stated, the technical glass compositions and many experimentally determined properties are taken from one large study. Unless stated otherwise, the properties of fused silica (quartz glass) and germania glass are derived from the SciGlass glass database by forming the arithmetic mean of all the experimental values from different authors (in general more than 10 independent sources for quartz glass and Tg of germanium oxide glass).
The list is not exhaustive.
Document 2:::
Tempered or toughened glass is a type of safety glass processed by controlled thermal or chemical treatments to increase its strength compared with normal glass. Tempering puts the outer surfaces into compression and the interior into tension. Such stresses cause the glass, when broken, to shatter into small granular chunks instead of splintering into jagged shards as ordinary annealed glass does. The granular chunks are less likely to cause injury.
Tempered glass is used for its safety and strength in a variety of applications, including passenger vehicle windows (apart from windshield), shower doors, aquariums, architectural glass doors and tables, refrigerator trays, mobile phone screen protectors, bulletproof glass components, diving masks, and plates and cookware.
Properties
Tempered glass is about four times stronger than annealed glass. The more rapid contraction of the outer layer during manufacturing induces compressive stresses in the surface of the glass balanced by tensile stresses in the body of the glass. Fully tempered 6-mm thick glass must have either a minimum surface compression of 69 MPa (10 000 psi) or an edge compression of not less than 67 MPa (9 700 psi). For it to be considered safety glass, the surface compressive stress should exceed . As a result of the increased surface stress, when broken the glass breaks into small rounded chunks as opposed to sharp jagged shards.
Compressive surface stresses give tempered glass increased strength. Annealed glass has almost no internal stress and usually forms microscopic cracks on its surface. Tension applied to the glass can drive crack propagation which, once begun, concentrates tension at the tip of the crack driving crack propagation at the speed of sound through the glass. Consequently, annealed glass is fragile and breaks into irregular and sharp pieces. The compressive stresses on the surface of tempered glass contain flaws, preventing their propagation or expansion.
Any cutting or grindi
Document 3:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 4:::
A characteristic property is a chemical or physical property that helps identify and classify substances. The characteristic properties of a substance are always the same whether the sample being observed is large or small. Thus, conversely, if the property of a substance changes as the sample size changes, that property is not a characteristic property. Examples of physical properties that are not characteristic properties are mass and volume. Examples of characteristic properties include melting points, boiling points, density, viscosity, solubility, crystal shape, and color. Substances with characteristic properties can be separated. For example, in fractional distillation, liquids are separated using the boiling point. The water Boiling point is 212 degrees Fahrenheit.
Identifying a substance
Every characteristic property is unique to one given substance. Scientists use characteristic properties to identify unknown substances. However, characteristic properties are most useful for distinguishing between two or more substances, not identifying a single substance. For example, isopropanol and water can be distinguished by the characteristic property of odor. Characteristic properties are used because the sample size and the shape of the substance does not matter. For example, 1 gram of lead is the same color as 100 tons of lead.
See also
Intensive and extensive properties
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Which property changes when a glass breaks?
A. State
B. Chemical
C. density
D. its physical property
Answer:
|
|
ai2_arc-1032
|
multiple_choice
|
One of the principal causes of acid rain is
|
[
"waste acid from chemical factories being pumped into rivers",
"acid from chemical laboratories evaporating into the air",
"gases from burning coal and oil dissolving in water in the atmosphere",
"gases from air conditioners and refrigerators escaping into the atmosphere"
] |
C
|
Relavent Documents:
Document 0:::
Trashing the Planet: How Science Can Help Us Deal With Acid Rain, Depletion of the Ozone, and Nuclear Waste (Among Other Things) is a 1990 book by zoologist and Governor of Washington Dixy Lee Ray. The book talks about the seriousness about acid rain, the problems with the ozone layer and other environmental issues. Ray co-wrote the book with journalist Lou Guzzo.
Document 1:::
Cary Institute of Ecosystem Studies (Cary Institute), formerly known as the Institute of Ecosystem Studies, is an independent, not-for-profit environmental research organization dedicated to the scientific study of the world's ecosystems and the natural and human factors that influence them. The organization is headquartered in Millbrook, NY on a research campus. Areas of expertise include disease ecology, urban ecology, freshwater ecology and provisioning, and forest health.
Details
Cary Institute's research is collaborative and multidisciplinary. Its scientists lead two of the National Science Foundation's Long Term Ecological Research Network sites: the Baltimore Ecosystem Study (Baltimore, MD; focus: urban ecology) and the Hubbard Brook Ecosystem Study (Woodstock, NH; focus: forest and freshwater health). They also play a leadership role in the Global Lake Ecological Observatory Network, an international effort that shares and interprets high resolution sensor data to understand, predict, and communicate the role and response of lakes in a changing global environment.
While working at Hubbard Brook Experimental Forest in the 1960s, Cary Institute founder Gene E. Likens co-discovered acid rain in North America. His longterm studies on precipitation and stream water chemistry were instrumental in shaping the 1990 Clean Air Act amendments. Today, Cary Institute continues to steward the longest continuous data set on acid rain and deposition through its direction of the Hubbard Brook Ecosystem Study.
Cary Institute's grounds have been home to long-term studies on the ecology of tick-borne disease for more than 20 years. Findings underpin The Tick Project, a 5-year study testing interventions with the potential to reduce Lyme disease and protect public health. A three-decade research program on the Hudson River informs sustainable shoreline management, and a synthesis of imported forest pests and pathogens is the basis for Tree-SMART Trade, a national policy init
Document 2:::
Bioremediation broadly refers to any process wherein a biological system (typically bacteria, microalgae, fungi in mycoremediation, and plants in phytoremediation), living or dead, is employed for removing environmental pollutants from air, water, soil, flue gasses, industrial effluents etc., in natural or artificial settings. The natural ability of organisms to adsorb, accumulate, and degrade common and emerging pollutants has attracted the use of biological resources in treatment of contaminated environment. In comparison to conventional physicochemical treatment methods bioremediation may offer considerable advantages as it aims to be sustainable, eco-friendly, cheap, and scalable.
Most bioremediation is inadvertent, involving native organisms. Research on bioremediation is heavily focused on stimulating the process by inoculation of a polluted site with organisms or supplying nutrients to promote the growth. In principle, bioremediation could be used to reduce the impact of byproducts created from anthropogenic activities, such as industrialization and agricultural processes. Bioremediation could prove less expensive and more sustainable than other remediation alternatives.
UNICEF, power producers, bulk water suppliers and local governments are early adopters of low cost bioremediation, such as aerobic bacteria tablets which are simply dropped into water.
While organic pollutants are susceptible to biodegradation, heavy metals are not degraded, but rather oxidized or reduced. Typical bioremediations involves oxidations. Oxidations enhance the water-solubility of organic compounds and their susceptibility to further degradation by further oxidation and hydrolysis. Ultimately biodegradation converts hydrocarbons to carbon dioxide and water. For heavy metals, bioremediation offers few solutions. Metal-containing pollutant can be removed or reduced with varying bioremediation techniques. The main challenge to bioremediations is rate: the processes are slow.
B
Document 3:::
Nutrient cycling in the Columbia River Basin involves the transport of nutrients through the system, as well as transformations from among dissolved, solid, and gaseous phases, depending on the element. The elements that constitute important nutrient cycles include macronutrients such as nitrogen (as ammonium, nitrite, and nitrate), silicate, phosphorus, and micronutrients, which are found in trace amounts, such as iron. Their cycling within a system is controlled by many biological, chemical, and physical processes.
The Columbia River Basin is the largest freshwater system of the Pacific Northwest, and due to its complexity, size, and modification by humans, nutrient cycling within the system is affected by many different components. Both natural and anthropogenic processes are involved in the cycling of nutrients. Natural processes in the system include estuarine mixing of fresh and ocean waters, and climate variability patterns such as the Pacific Decadal Oscillation and the El Nino Southern Oscillation (both climatic cycles that affect the amount of regional snowpack and river discharge). Natural sources of nutrients in the Columbia River include weathering, leaf litter, salmon carcasses, runoff from its tributaries, and ocean estuary exchange. Major anthropogenic impacts to nutrients in the basin are due to fertilizers from agriculture, sewage systems, logging, and the construction of dams.
Nutrients dynamics vary in the river basin from the headwaters to the main river and dams, to finally reaching the Columbia River estuary and ocean. Upstream in the headwaters, salmon runs are the main source of nutrients. Dams along the river impact nutrient cycling by increasing residence time of nutrients, and reducing the transport of silicate to the estuary, which directly impacts diatoms, a type of phytoplankton. The dams are also a barrier to salmon migration, and can increase the amount of methane locally produced. The Columbia River estuary exports high rates of n
Document 4:::
A Bjerrum plot (named after Niels Bjerrum), sometimes also known as a Sillén diagram (after Lars Gunnar Sillén), or a Hägg diagram (after Gunnar Hägg) is a graph of the concentrations of the different species of a polyprotic acid in a solution, as a function of pH, when the solution is at equilibrium. Due to the many orders of magnitude spanned by the concentrations, they are commonly plotted on a logarithmic scale. Sometimes the ratios of the concentrations are plotted rather than the actual concentrations. Occasionally H+ and OH− are also plotted.
Most often, the carbonate system is plotted, where the polyprotic acid is carbonic acid (a diprotic acid), and the different species are dissolved carbon dioxide, carbonic acid, bicarbonate, and carbonate. In acidic conditions, the dominant form is ; in basic (alkaline) conditions, the dominant form is ; and in between, the dominant form is . At every pH, the concentration of carbonic acid is assumed to be negligible compared to the concentration of dissolved , and so is often omitted from Bjerrum plots. These plots are very helpful in solution chemistry and natural water chemistry. In the example given here, it illustrates the response of seawater pH and carbonate speciation due to the input of man-made emission by the fossil fuel combustion.
The Bjerrum plots for other polyprotic acids, including silicic, boric, sulfuric and phosphoric acids, are other commonly used examples.
Bjerrum plot equations for carbonate system
If carbon dioxide, carbonic acid, hydrogen ions, bicarbonate and carbonate are all dissolved in water, and at chemical equilibrium, their equilibrium concentrations are often assumed to be given by:
where the subscript 'eq' denotes that these are equilibrium concentrations, K1 is the equilibrium constant for the reaction + H+ + (i.e. the first acid dissociation constant for carbonic acid), K2 is the equilibrium constant for the reaction H+ + (i.e. the second acid dissociation constant for
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
One of the principal causes of acid rain is
A. waste acid from chemical factories being pumped into rivers
B. acid from chemical laboratories evaporating into the air
C. gases from burning coal and oil dissolving in water in the atmosphere
D. gases from air conditioners and refrigerators escaping into the atmosphere
Answer:
|
|
sciq-10434
|
multiple_choice
|
Insects can have specialized mouthparts, such as a proboscis, to siphon the nectar from what?
|
[
"flowers",
"stems",
"fruits",
"shrubs"
] |
A
|
Relavent Documents:
Document 0:::
Insect cognition describes the mental capacities and study of those capacities in insects. The field developed from comparative psychology where early studies focused more on animal behavior. Researchers have examined insect cognition in bees, fruit flies, and wasps.
Research questions consist of experiments aimed to evaluate insects abilities such as perception, emotions attention, memory (wasp multiple nest), spatial cognition, tools use, problem solving, and concepts. Unlike in animal behavior the concept of group cognition plays a big part in insect studies. It is hypothesized some insect classes like ants and bees think with a group cognition to function within their societies; more recent studies show that individual cognition exists and plays a role in overall group cognitive task.
Insect cognition experiments have been more prevalent in the past decade than prior. It is logical for the understanding of cognitive capacities as adaptations to differing ecological niches under the Cognitive faculty by species when analyzing behaviors, this means viewing behaviors as adaptations to an individual's environment and not weighing them more advanced when compared to other different individuals.
Insect foraging cognition
Insects inhabit many diverse and complex environments within which they must find food. Cognition shapes how an insect comes to find its food. The particular cognitive abilities used by insects in finding food has been the focus of much scientific inquiry. The social insects are often study subjects and much has been discovered about the intelligence of insects by investigating the abilities of bee species. Fruit flies are also common study subjects.
Learning and memory
Learning biases
Through learning, insects can increase their foraging efficiency, decreasing the time spent searching for food which allows for more time and energy to invest in other fitness related activities, such as searching for mates. Depending on the ecology of the
Document 1:::
This is a list of honeydew sources. Honeydew is a sugary excretion from plant sap sucking insects such as aphids or scales. There are many trees that are hosts to aphids and scale insects that produce honeydew
Honeydew sources
Document 2:::
A leaf litter sieve is a piece of equipment used by entomologists, in particular by coleopterists (beetle collectors) (Cooter 1991, page 7) as an aid to finding invertebrates in leaf litter.
A typical leaf litter sieve consists of a gauze with holes of approximately 5 to 10 mm width. The entomologist places handfuls of leaf litter into the sieve, which is placed above a white sheet or tray. The sieve is shaken, and insects are separated from the leaf litter and fall out for inspection. Charles Valentine Riley details use of a simple sieve with a cloth bag.
A more complex combination sieve is described by Hongfu.
See also
Tullgren funnel
Document 3:::
This glossary of entomology describes terms used in the formal study of insect species by entomologists.
A–C
A synthetic chlorinated hydrocarbon insecticide, toxic to vertebrates. Though its phytotoxicity is low, solvents in some formulations may damage certain crops. cf. the related Dieldrin, Endrin, Isodrin
D–F
A synthetic chlorinated hydrocarbon insecticide, toxic to vertebrates. cf. the related Aldrin, Endrin, Isodrin
A synthetic chlorinated hydrocarbon insecticide, toxic to vertebrates. Though its phytotoxicity is low, solvents in some formulations may damage certain crops. cf. the related Dieldrin, Aldrin, Isodrin
G–L
A synthetic chlorinated hydrocarbon insecticide, toxic to vertebrates. Though its phytotoxicity is low, solvents in some formulations may damage certain crops. cf. the related Dieldrin, Aldrin, Endrin
M–O
P–R
S–Z
Figures
See also
Anatomical terms of location
Butterfly
Caterpillar
Comstock–Needham system
External morphology of Lepidoptera
Glossary of ant terms
Glossary of spider terms
Glossary of scientific names
Insect wing
Pupa
Document 4:::
Entomology, the scientific study of insects and closely related terrestrial arthropods, has been impelled by the necessity of societies to protect themselves from insect-borne diseases, crop losses to pest insects, and insect-related discomfort, as well as by people's natural curiosity. This timeline article traces the history of entomology.
Timelines of entomology
Timeline of entomology – prior to 1800
Timeline of entomology – 1800–1850
Timeline of entomology – 1850–1900
Timeline of entomology – post 1900
History of classification
Many different classifications were proposed by early entomologists. It is important to realise that whilst many early names survive, they may be at different levels in the phylogenetic hierarchy. For instance, many families were first published as genera, as for example the genus Mymar, proposed by Alexander Henry Haliday in 1829, is now represented by the family Mymaridae.
History of forensic entomology
See also
European and American voyages of scientific exploration
List of natural history dealers
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Insects can have specialized mouthparts, such as a proboscis, to siphon the nectar from what?
A. flowers
B. stems
C. fruits
D. shrubs
Answer:
|
|
sciq-8073
|
multiple_choice
|
Name a carbonyl in which the carbon atom is bonded to one carbon atom and one hydrogen atom (or two hydrogen atoms).
|
[
"aldehyde",
"ester",
"acetylcholine",
"ketone"
] |
A
|
Relavent Documents:
Document 0:::
In chemistry, the carbon-hydrogen bond ( bond) is a chemical bond between carbon and hydrogen atoms that can be found in many organic compounds. This bond is a covalent, single bond, meaning that carbon shares its outer valence electrons with up to four hydrogens. This completes both of their outer shells, making them stable.
Carbon–hydrogen bonds have a bond length of about 1.09 Å (1.09 × 10−10 m) and a bond energy of about 413 kJ/mol (see table below). Using Pauling's scale—C (2.55) and H (2.2)—the electronegativity difference between these two atoms is 0.35. Because of this small difference in electronegativities, the bond is generally regarded as being non-polar. In structural formulas of molecules, the hydrogen atoms are often omitted. Compound classes consisting solely of bonds and bonds are alkanes, alkenes, alkynes, and aromatic hydrocarbons. Collectively they are known as hydrocarbons.
In October 2016, astronomers reported that the very basic chemical ingredients of life—the carbon-hydrogen molecule (CH, or methylidyne radical), the carbon-hydrogen positive ion () and the carbon ion ()—are the result, in large part, of ultraviolet light from stars, rather than in other ways, such as the result of turbulent events related to supernovae and young stars, as thought earlier.
Bond length
The length of the carbon-hydrogen bond varies slightly with the hybridisation of the carbon atom. A bond between a hydrogen atom and an sp2 hybridised carbon atom is about 0.6% shorter than between hydrogen and sp3 hybridised carbon. A bond between hydrogen and sp hybridised carbon is shorter still, about 3% shorter than sp3 C-H. This trend is illustrated by the molecular geometry of ethane, ethylene and acetylene.
Reactions
The C−H bond in general is very strong, so it is relatively unreactive. In several compound classes, collectively called carbon acids, the C−H bond can be sufficiently acidic for proton removal. Unactivated C−H bonds are found in alkanes and are no
Document 1:::
A carbon–carbon bond is a covalent bond between two carbon atoms. The most common form is the single bond: a bond composed of two electrons, one from each of the two atoms. The carbon–carbon single bond is a sigma bond and is formed between one hybridized orbital from each of the carbon atoms. In ethane, the orbitals are sp3-hybridized orbitals, but single bonds formed between carbon atoms with other hybridizations do occur (e.g. sp2 to sp2). In fact, the carbon atoms in the single bond need not be of the same hybridization. Carbon atoms can also form double bonds in compounds called alkenes or triple bonds in compounds called alkynes. A double bond is formed with an sp2-hybridized orbital and a p-orbital that is not involved in the hybridization. A triple bond is formed with an sp-hybridized orbital and two p-orbitals from each atom. The use of the p-orbitals forms a pi bond.
Chains and branching
Carbon is one of the few elements that can form long chains of its own atoms, a property called catenation. This coupled with the strength of the carbon–carbon bond gives rise to an enormous number of molecular forms, many of which are important structural elements of life, so carbon compounds have their own field of study: organic chemistry.
Branching is also common in C−C skeletons. Carbon atoms in a molecule are categorized by the number of carbon neighbors they have:
A primary carbon has one carbon neighbor.
A secondary carbon has two carbon neighbors.
A tertiary carbon has three carbon neighbors.
A quaternary carbon has four carbon neighbors.
In "structurally complex organic molecules", it is the three-dimensional orientation of the carbon–carbon bonds at quaternary loci which dictates the shape of the molecule. Further, quaternary loci are found in many biologically active small molecules, such as cortisone and morphine.
Synthesis
Carbon–carbon bond-forming reactions are organic reactions in which a new carbon–carbon bond is formed. They are important in th
Document 2:::
A carbon–nitrogen bond is a covalent bond between carbon and nitrogen and is one of the most abundant bonds in organic chemistry and biochemistry.
Nitrogen has five valence electrons and in simple amines it is trivalent, with the two remaining electrons forming a lone pair. Through that pair, nitrogen can form an additional bond to hydrogen making it tetravalent and with a positive charge in ammonium salts. Many nitrogen compounds can thus be potentially basic but its degree depends on the configuration: the nitrogen atom in amides is not basic due to delocalization of the lone pair into a double bond and in pyrrole the lone pair is part of an aromatic sextet.
Similar to carbon–carbon bonds, these bonds can form stable double bonds, as in imines; and triple bonds, such as nitriles. Bond lengths range from 147.9 pm for simple amines to 147.5 pm for C-N= compounds such as nitromethane to 135.2 pm for partial double bonds in pyridine to 115.8 pm for triple bonds as in nitriles.
A CN bond is strongly polarized towards nitrogen (the electronegativities of C and N are 2.55 and 3.04, respectively) and subsequently molecular dipole moments can be high: cyanamide 4.27 D, diazomethane 1.5 D, methyl azide 2.17, pyridine 2.19. For this reason many compounds containing CN bonds are water-soluble. N-philes are group of radical molecules which are specifically attracted to the C=N bonds.
Carbon-nitrogen bond can be analyzed by X-ray photoelectron spectroscopy (XPS). Depending on the bonding states the peak positions differ in N1s XPS spectra.
Nitrogen functional groups
See also
Cyanide
Other carbon bonds with group 15 elements: carbon–nitrogen bonds, carbon–phosphorus bonds
Other carbon bonds with period 2 elements: carbon–lithium bonds, carbon–beryllium bonds, carbon–boron bonds, carbon–carbon bonds, carbon–nitrogen bonds, carbon–oxygen bonds, carbon–fluorine bonds
Carbon–hydrogen bond
Document 3:::
In organic chemistry, a Platonic hydrocarbon is a hydrocarbon (molecule) whose structure matches one of the five Platonic solids, with carbon atoms replacing its vertices, carbon–carbon bonds replacing its edges, and hydrogen atoms as needed.
Not all Platonic solids have molecular hydrocarbon counterparts; those that do are the tetrahedron (tetrahedrane), the cube (cubane), and the dodecahedron (dodecahedrane).
Tetrahedrane
Tetrahedrane (C4H4) is a hypothetical compound. It has not yet been synthesized without substituents, but it is predicted to be kinetically stable in spite of its angle strain. Some stable derivatives, including tetra(tert-butyl)tetrahedrane (a hydrocarbon) and tetra(trimethylsilyl)tetrahedrane, have been produced.
Cubane
Cubane (C8H8) has been synthesized. Although it has high angle strain, cubane is kinetically stable, due to a lack of readily available decomposition paths.
Octahedrane
Angle strain would make an octahedron highly unstable due to inverted tetrahedral geometry at each vertex. There would also be no hydrogen atoms because four edges meet at each corner; thus, the hypothetical octahedrane molecule would be an allotrope of elemental carbon, C6, and not a hydrocarbon. The existence of octahedrane cannot be ruled out completely, although calculations have shown that it is unlikely.
Dodecahedrane
Dodecahedrane (C20H20) was first synthesized in 1982, and has minimal angle strain; the tetrahedral angle is 109.5° and the dodecahedral angle is 108°, only a slight discrepancy.
Icosahedrane
The tetravalency (4-connectedness) of carbon excludes an icosahedron because 5 edges meet at each vertex. True pentavalent carbon is unlikely; methanium, nominally , usually exists as . The hypothetical icosahedral lacks hydrogen so it is not a hydrocarbon; it is also an ion.
Both icosahedral and octahedral structures have been observed in boron compounds such as the dodecaborate ion and some of the carbon-containing carboranes.
Other polyhedr
Document 4:::
In chemical nomenclature, the IUPAC nomenclature of organic chemistry is a method of naming organic chemical compounds as recommended by the International Union of Pure and Applied Chemistry (IUPAC). It is published in the Nomenclature of Organic Chemistry (informally called the Blue Book). Ideally, every possible organic compound should have a name from which an unambiguous structural formula can be created. There is also an IUPAC nomenclature of inorganic chemistry.
To avoid long and tedious names in normal communication, the official IUPAC naming recommendations are not always followed in practice, except when it is necessary to give an unambiguous and absolute definition to a compound. IUPAC names can sometimes be simpler than older names, as with ethanol, instead of ethyl alcohol. For relatively simple molecules they can be more easily understood than non-systematic names, which must be learnt or looked over. However, the common or trivial name is often substantially shorter and clearer, and so preferred. These non-systematic names are often derived from an original source of the compound. Also, very long names may be less clear than structural formulas.
Basic principles
In chemistry, a number of prefixes, suffixes and infixes are used to describe the type and position of the functional groups in the compound.
The steps for naming an organic compound are:
Identification of the parent hydride parent hydrocarbon chain. This chain must obey the following rules, in order of precedence:
It should have the maximum number of substituents of the suffix functional group. By suffix, it is meant that the parent functional group should have a suffix, unlike halogen substituents. If more than one functional group is present, the one with highest group precedence should be used.
It should have the maximum number of multiple bonds.
It should have the maximum length.
It should have the maximum number of substituents or branches cited as prefixes
It should have the ma
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Name a carbonyl in which the carbon atom is bonded to one carbon atom and one hydrogen atom (or two hydrogen atoms).
A. aldehyde
B. ester
C. acetylcholine
D. ketone
Answer:
|
|
sciq-6706
|
multiple_choice
|
What term describes waves that travel through matter?
|
[
"replaced waves",
"mechanical waves",
"heat waves",
"water waves"
] |
B
|
Relavent Documents:
Document 0:::
In physics, a mechanical wave is a wave that is an oscillation of matter, and therefore transfers energy through a medium. While waves can move over long distances, the movement of the medium of transmission—the material—is limited. Therefore, the oscillating material does not move far from its initial equilibrium position. Mechanical waves can be produced only in media which possess elasticity and inertia. There are three types of mechanical waves: transverse waves, longitudinal waves, and surface waves. Some of the most common examples of mechanical waves are water waves, sound waves, and seismic waves.
Like all waves, mechanical waves transport energy. This energy propagates in the same direction as the wave. A wave requires an initial energy input; once this initial energy is added, the wave travels through the medium until all its energy is transferred. In contrast, electromagnetic waves require no medium, but can still travel through one.
Transverse wave
A transverse wave is the form of a wave in which particles of medium vibrate about their mean position perpendicular to the direction of the motion of the wave.
To see an example, move an end of a Slinky (whose other end is fixed) to the left-and-right of the Slinky, as opposed to to-and-fro. Light also has properties of a transverse wave, although it is an electromagnetic wave.
Longitudinal wave
Longitudinal waves cause the medium to vibrate parallel to the direction of the wave. It consists of multiple compressions and rarefactions. The rarefaction is the farthest distance apart in the longitudinal wave and the compression is the closest distance together. The speed of the longitudinal wave is increased in higher index of refraction, due to the closer proximity of the atoms in the medium that is being compressed. Sound is a longitudinal wave.
Surface waves
This type of wave travels along the surface or interface between two media. An example of a surface wave would be waves in a pool, or in an ocean
Document 1:::
In physics, a transverse wave is a wave that oscillates perpendicularly to the direction of the wave's advance. In contrast, a longitudinal wave travels in the direction of its oscillations. All waves move energy from place to place without transporting the matter in the transmission medium if there is one. Electromagnetic waves are transverse without requiring a medium. The designation “transverse” indicates the direction of the wave is perpendicular to the displacement of the particles of the medium through which it passes, or in the case of EM waves, the oscillation is perpendicular to the direction of the wave.
A simple example is given by the waves that can be created on a horizontal length of string by anchoring one end and moving the other end up and down. Another example is the waves that are created on the membrane of a drum. The waves propagate in directions that are parallel to the membrane plane, but each point in the membrane itself gets displaced up and down, perpendicular to that plane. Light is another example of a transverse wave, where the oscillations are the electric and magnetic fields, which point at right angles to the ideal light rays that describe the direction of propagation.
Transverse waves commonly occur in elastic solids due to the shear stress generated; the oscillations in this case are the displacement of the solid particles away from their relaxed position, in directions perpendicular to the propagation of the wave. These displacements correspond to a local shear deformation of the material. Hence a transverse wave of this nature is called a shear wave. Since fluids cannot resist shear forces while at rest, propagation of transverse waves inside the bulk of fluids is not possible. In seismology, shear waves are also called secondary waves or S-waves.
Transverse waves are contrasted with longitudinal waves, where the oscillations occur in the direction of the wave. The standard example of a longitudinal wave is a sound wave or "
Document 2:::
This is a list of wave topics.
0–9
21 cm line
A
Abbe prism
Absorption spectroscopy
Absorption spectrum
Absorption wavemeter
Acoustic wave
Acoustic wave equation
Acoustics
Acousto-optic effect
Acousto-optic modulator
Acousto-optics
Airy disc
Airy wave theory
Alfvén wave
Alpha waves
Amphidromic point
Amplitude
Amplitude modulation
Animal echolocation
Antarctic Circumpolar Wave
Antiphase
Aquamarine Power
Arrayed waveguide grating
Artificial wave
Atmospheric diffraction
Atmospheric wave
Atmospheric waveguide
Atom laser
Atomic clock
Atomic mirror
Audience wave
Autowave
Averaged Lagrangian
B
Babinet's principle
Backward wave oscillator
Bandwidth-limited pulse
beat
Berry phase
Bessel beam
Beta wave
Black hole
Blazar
Bloch's theorem
Blueshift
Boussinesq approximation (water waves)
Bow wave
Bragg diffraction
Bragg's law
Breaking wave
Bremsstrahlung, Electromagnetic radiation
Brillouin scattering
Bullet bow shockwave
Burgers' equation
Business cycle
C
Capillary wave
Carrier wave
Cherenkov radiation
Chirp
Ernst Chladni
Circular polarization
Clapotis
Closed waveguide
Cnoidal wave
Coherence (physics)
Coherence length
Coherence time
Cold wave
Collimated light
Collimator
Compton effect
Comparison of analog and digital recording
Computation of radiowave attenuation in the atmosphere
Continuous phase modulation
Continuous wave
Convective heat transfer
Coriolis frequency
Coronal mass ejection
Cosmic microwave background radiation
Coulomb wave function
Cutoff frequency
Cutoff wavelength
Cymatics
D
Damped wave
Decollimation
Delta wave
Dielectric waveguide
Diffraction
Direction finding
Dispersion (optics)
Dispersion (water waves)
Dispersion relation
Dominant wavelength
Doppler effect
Doppler radar
Douglas Sea Scale
Draupner wave
Droplet-shaped wave
Duhamel's principle
E
E-skip
Earthquake
Echo (phenomenon)
Echo sounding
Echolocation (animal)
Echolocation (human)
Eddy (fluid dynamics)
Edge wave
Eikonal equation
Ekman layer
Ekman spiral
Ekman transport
El Niño–Southern Oscillation
El
Document 3:::
In physics, a surface wave is a mechanical wave that propagates along the interface between differing media. A common example is gravity waves along the surface of liquids, such as ocean waves. Gravity waves can also occur within liquids, at the interface between two fluids with different densities. Elastic surface waves can travel along the surface of solids, such as Rayleigh or Love waves. Electromagnetic waves can also propagate as "surface waves" in that they can be guided along with a refractive index gradient or along an interface between two media having different dielectric constants. In radio transmission, a ground wave is a guided wave that propagates close to the surface of the Earth.
Mechanical waves
In seismology, several types of surface waves are encountered. Surface waves, in this mechanical sense, are commonly known as either Love waves (L waves) or Rayleigh waves. A seismic wave is a wave that travels through the Earth, often as the result of an earthquake or explosion. Love waves have transverse motion (movement is perpendicular to the direction of travel, like light waves), whereas Rayleigh waves have both longitudinal (movement parallel to the direction of travel, like sound waves) and transverse motion. Seismic waves are studied by seismologists and measured by a seismograph or seismometer. Surface waves span a wide frequency range, and the period of waves that are most damaging is usually 10 seconds or longer. Surface waves can travel around the globe many times from the largest earthquakes. Surface waves are caused when P waves and S waves come to the surface.
Examples are the waves at the surface of water and air (ocean surface waves). Another example is internal waves, which can be transmitted along the interface of two water masses of different densities.
In theory of hearing physiology, the traveling wave (TW) of Von Bekesy, resulted from an acoustic surface wave of the basilar membrane into the cochlear duct. His theory purporte
Document 4:::
Acoustic waves are a type of energy propagation through a medium by means of adiabatic loading and unloading. Important quantities for describing acoustic waves are acoustic pressure, particle velocity, particle displacement and acoustic intensity. Acoustic waves travel with a characteristic acoustic velocity that depends on the medium they're passing through. Some examples of acoustic waves are audible sound from a speaker (waves traveling through air at the speed of sound), seismic waves (ground vibrations traveling through the earth), or ultrasound used for medical imaging (waves traveling through the body).
Wave properties
Acoustic wave is a mechanical wave that transmits energy through the movements of atoms and molecules. Acoustic wave transmits through liquids in longitudinal manner (movement of particles are parallel to the direction of propagation of the wave); in contrast to electromagnetic wave that transmits in transverse manner (movement of particles at a right angle to the direction of propagation of the wave). However, in solids, acoustic wave transmits in both longitudinal and transverse manners due to presence of shear moduli in such a state of matter.
Acoustic wave equation
The acoustic wave equation describes the propagation of sound waves. The acoustic wave equation for sound pressure in one dimension is given by
where
is sound pressure in Pa
is position in the direction of propagation of the wave, in m
is speed of sound in m/s
is time in s
The wave equation for particle velocity has the same shape and is given by
where
is particle velocity in m/s
For lossy media, more intricate models need to be applied in order to take into account frequency-dependent attenuation and phase speed. Such models include acoustic wave equations that incorporate fractional derivative terms, see also the acoustic attenuation article.
D'Alembert gave the general solution for the lossless wave equation. For sound pressure, a solution would be
where
is angu
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What term describes waves that travel through matter?
A. replaced waves
B. mechanical waves
C. heat waves
D. water waves
Answer:
|
|
sciq-3857
|
multiple_choice
|
Where is the hypothalamus–pituitary complex located in the body?
|
[
"tribulus of the brain",
"subthalamic of the brain",
"stem of the brain",
"diencephalon of the brain"
] |
D
|
Relavent Documents:
Document 0:::
The following diagram is provided as an overview of and topical guide to the human nervous system:
Human nervous system – the part of the human body that coordinates a person's voluntary and involuntary actions and transmits signals between different parts of the body. The human nervous system consists of two main parts: the central nervous system (CNS) and the peripheral nervous system (PNS). The CNS contains the brain and spinal cord. The PNS consists mainly of nerves, which are long fibers that connect the CNS to every other part of the body. The PNS includes motor neurons, mediating voluntary movement; the autonomic nervous system, comprising the sympathetic nervous system and the parasympathetic nervous system and regulating involuntary functions; and the enteric nervous system, a semi-independent part of the nervous system whose function is to control the gastrointestinal system.
Evolution of the human nervous system
Evolution of nervous systems
Evolution of human intelligence
Evolution of the human brain
Paleoneurology
Some branches of science that study the human nervous system
Neuroscience
Neurology
Paleoneurology
Central nervous system
The central nervous system (CNS) is the largest part of the nervous system and includes the brain and spinal cord.
Spinal cord
Brain
Brain – center of the nervous system.
Outline of the human brain
List of regions of the human brain
Principal regions of the vertebrate brain:
Peripheral nervous system
Peripheral nervous system (PNS) – nervous system structures that do not lie within the CNS.
Sensory system
A sensory system is a part of the nervous system responsible for processing sensory information. A sensory system consists of sensory receptors, neural pathways, and parts of the brain involved in sensory perception.
List of sensory systems
Sensory neuron
Perception
Visual system
Auditory system
Somatosensory system
Vestibular system
Olfactory system
Taste
Pain
Components of the nervous system
Neuron
I
Document 1:::
The human brain anatomical regions are ordered following standard neuroanatomy hierarchies. Functional, connective, and developmental regions are listed in parentheses where appropriate.
Hindbrain (rhombencephalon)
Myelencephalon
Medulla oblongata
Medullary pyramids
Arcuate nucleus
Olivary body
Inferior olivary nucleus
Rostral ventrolateral medulla
Caudal ventrolateral medulla
Solitary nucleus (Nucleus of the solitary tract)
Respiratory center-Respiratory groups
Dorsal respiratory group
Ventral respiratory group or Apneustic centre
Pre-Bötzinger complex
Botzinger complex
Retrotrapezoid nucleus
Nucleus retrofacialis
Nucleus retroambiguus
Nucleus para-ambiguus
Paramedian reticular nucleus
Gigantocellular reticular nucleus
Parafacial zone
Cuneate nucleus
Gracile nucleus
Perihypoglossal nuclei
Intercalated nucleus
Prepositus nucleus
Sublingual nucleus
Area postrema
Medullary cranial nerve nuclei
Inferior salivatory nucleus
Nucleus ambiguus
Dorsal nucleus of vagus nerve
Hypoglossal nucleus
Chemoreceptor trigger zone
Metencephalon
Pons
Pontine nuclei
Pontine cranial nerve nuclei
Chief or pontine nucleus of the trigeminal nerve sensory nucleus (V)
Motor nucleus for the trigeminal nerve (V)
Abducens nucleus (VI)
Facial nerve nucleus (VII)
Vestibulocochlear nuclei (vestibular nuclei and cochlear nuclei) (VIII)
Superior salivatory nucleus
Pontine tegmentum
Pontine micturition center (Barrington's nucleus)
Locus coeruleus
Pedunculopontine nucleus
Laterodorsal tegmental nucleus
Tegmental pontine reticular nucleus
Nucleus incertus
Parabrachial area
Medial parabrachial nucleus
Lateral parabrachial nucleus
Subparabrachial nucleus (Kölliker-Fuse nucleus)
Pontine respiratory group
Superior olivary complex
Medial superior olive
Lateral superior olive
Medial nucleus of the trapezoid body
Paramedian pontine reticular formation
Parvocellular reticular nucleus
Caudal pontine reticular nucleus
Cerebellar peduncles
Superior cerebellar peduncle
Middle cerebellar peduncle
Inferior
Document 2:::
The chemoreceptor trigger zone (CTZ) is an area of the medulla oblongata that receives inputs from blood-borne drugs or hormones, and communicates with other structures in the vomiting center to initiate vomiting. The CTZ is located within the area postrema, which is on the floor of the fourth ventricle and is outside of the blood–brain barrier. It is also part of the vomiting center itself. The neurotransmitters implicated in the control of nausea and vomiting include acetylcholine, dopamine, histamine (H1 receptor), substance P (NK-1 receptor), and serotonin (5-HT3 receptor). There are also opioid receptors present, which may be involved in the mechanism by which opiates cause nausea and vomiting. The blood–brain barrier is not as developed here; therefore, drugs such as dopamine which cannot normally enter the CNS may still stimulate the CTZ.
Evolutionary significance
The CTZ is in the medulla oblongata, which is phylogenetically the oldest part of the central nervous system. Early lifeforms developed a brainstem, or inner brain, and nothing more. This part of the brain is responsible for basic survival instincts and reactions, for example to make an organism turn its head and look where an auditory stimulus was heard. The brainstem is where the medulla is located, and therefore also the area postrema and the CTZ. Then later lifeforms developed another segment of the brain, which includes the limbic system. This area of the brain is responsible for producing emotion and emotional responses to external stimuli, and also is significantly involved in memory and reward systems. Evolutionarily, the cerebral cortex is the most recent development. This area of the brain is responsible for critical thinking and reasoning, and is actively involved in decision making. It has been discovered that a major cause of increased intelligence in species including humans is the increase in cortical neurons in the brain. The emetic response was selected for protective purposes, and
Document 3:::
The ovarian cortex is the outer portion of the ovary. The ovarian follicles are located within the ovarian cortex. The ovarian cortex is made up of connective tissue. Ovarian cortex tissue transplant has been performed to treat infertility.
Document 4:::
This article describes anatomical terminology that is used to describe the central and peripheral nervous systems - including the brain, brainstem, spinal cord, and nerves.
Anatomical terminology in neuroanatomy
Neuroanatomy, like other aspects of anatomy, uses specific terminology to describe anatomical structures. This terminology helps ensure that a structure is described accurately, with minimal ambiguity. Terms also help ensure that structures are described consistently, depending on their structure or function. Terms are often derived from Latin and Greek, and like other areas of anatomy are generally standardised based on internationally accepted lexicons such as Terminologia Anatomica.
To help with consistency, humans and other species are assumed when described to be in standard anatomical position, with the body standing erect and facing observer, arms at sides, palms forward.
Location
Anatomical terms of location depend on the location and species that is being described.
To understand the terms used for anatomical localisation, consider an animal with a straight CNS, such as a fish or lizard. In such animals the terms "rostral", "caudal", "ventral" and "dorsal" mean respectively towards the rostrum, towards the tail, towards the belly and towards the back. For a full discussion of those terms, see anatomical terms of location.
For many purposes of anatomical description, positions and directions are relative to the standard anatomical planes and axes. Such reference to the anatomical planes and axes is called the stereotactic approach.
Standard terms used throughout anatomy include anterior / posterior for the front and back of a structure, superior / inferior for above and below, medial / lateral for structures close to and away from the midline respectively, and proximal / distal for structures close to and far away from a set point.
Some terms are used more commonly in neuroanatomy, particularly:
Rostral and caudal: In animals with linear ne
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Where is the hypothalamus–pituitary complex located in the body?
A. tribulus of the brain
B. subthalamic of the brain
C. stem of the brain
D. diencephalon of the brain
Answer:
|
|
ai2_arc-99
|
multiple_choice
|
Why are alternative fuels being used in some automobiles?
|
[
"Alternative fuels are at every gas station.",
"Gasoline comes from a limited resource.",
"Alternative fuels cause pollution.",
"Gasoline engines are too expensive to make."
] |
B
|
Relavent Documents:
Document 0:::
Corn ethanol is ethanol produced from corn biomass and is the main source of ethanol fuel in the United States, mandated to be blended with gasoline in the Renewable Fuel Standard. Corn ethanol is produced by ethanol fermentation and distillation. It is debatable whether the production and use of corn ethanol results in lower greenhouse gas emissions than gasoline. Approximately 45% of U.S. corn croplands are used for ethanol production.
Uses
Since 2001, corn ethanol production has increased by more than several times. Out of 9.50 billions of bushels of corn produced in 2001, 0.71 billions of bushels were used to produce corn ethanol. Compared to 2018, out of 14.62 billions of bushels of corn produced, 5.60 billion bushels were used to produce corn ethanol, reported by the United States Department of Energy. Overall, 94% of ethanol in the United States is produced from corn.
Currently, corn ethanol is mainly used in blends with gasoline to create mixtures such as E10, E15, and E85. Ethanol is mixed into more than 98% of United States gasoline to reduce air pollution. Corn ethanol is used as an oxygenate when mixed with gasoline. E10 and E15 can be used in all engines without modification. However, blends like E85, with a much greater ethanol content, require significant modifications to be made before an engine can run on the mixture without damaging the engine. Some vehicles that currently use E85 fuel, also called flex fuel, include, the Ford Focus, Dodge Durango, and Toyota Tundra, among others.
The future use of corn ethanol as a main gasoline replacement is unknown. Corn ethanol has yet to be proven to be as cost effective as gasoline due to corn ethanol being much more expensive to create compared to gasoline. Corn ethanol has to go through an extensive milling process before it can be used as a fuel source. One major drawback with corn ethanol, is the energy returned on energy invested (EROI), meaning the energy outputted in comparison to the energy requ
Document 1:::
Biodesulfurization is the process of removing sulfur from crude oil through the use of microorganisms or their enzymes.
Background
Crude oil contains sulfur in its composition, with the latter being the most abundant element after carbon and hydrogen. Depending on its source, the amount of sulfur present in crude oil can range from 0.05 to 10%. Accordingly, the oil can be classified as sweet or sour if the sulfur concentration is below or above 0.5%, respectively.
The combustion of crude oil releases sulfur oxides (SOx) to the atmosphere, which are harmful to public health and contribute to serious environmental effects such as air pollution and acid rains. In addition, the sulfur content in crude oil is a major problem for refineries, as it promotes the corrosion of the equipment and the poisoning of the noble metal catalysts. The levels of sulfur in any oil field are too high for the fossil fuels derived from it (such as gasoline, diesel, or jet fuel ) to be used in combustion engines without pre-treatment to remove organosulfur compounds.
The reduction of the concentration of sulfur in crude oil becomes necessary to mitigate one of the leading sources of the harmful health and environmental effects caused by its combustion. In this sense, the European union has taken steps to decrease the sulfur content in diesel below 10 ppm, while the US has made efforts to restrict the sulfur content in diesel and gasoline to a maximum of 15 ppm. The reduction of sulfur compounds in oil fuels can be achieved by a process named desulfurization. Methods used for desulfurization include, among others, hydrodesulfurization, oxidative desulfurization, extractive desulfurization, and extraction by ionic liquids.
Despite their efficiency at reducing sulfur content, the conventional desulfurization methods are still accountable for a significant amount of the CO2 emissions associated with the crude oil refining process, releasing up to 9000 metric tons per year. Furthermore, the
Document 2:::
The Certification for Sustainable Transportation is a national program housed at the University of Vermont Extension that seeks to promote the practice of using energy efficient modes of transportation. The CST work centers on its eRating vehicle certification program, which is an eco-label for passenger transportation vehicles. The eRating uses a sustainability index which includes factors such as green house gas emissions per passenger mile, emission levels of criteria pollutants, and in certain circumstances factors such as training for drivers and use of endorsed carbon offsets. Once a certain threshold is met, vehicles may qualify for e1, e2, e3, or e4 levels in the certification program.
Other key components of the CST's work are online and in person training programs. The CST offers training programs geared to help drivers and organizations eliminate all unnecessary idling and on eco-driving. These training programs are focused on helping reduce environmental impacts, save fuel, and save money.
The CST is now actively working with companies in 48 states and three Canadian provinces to prevent unnecessary emissions, reduce environmental impact, and decrease consumption of fossil fuels.
This program is not to be confused with the "E-Mark" vehicle equipment safety certification promulgated by the European Union since 2002 under EU Directive 72/245/EEC and amendments to the requirements of Directive 95/54/EC.
History
In 2007, the University of Vermont began the Green Coach Certification research project, which sought to investigate what efficiency standards would be best applied to motor coaches to promote greater energy sustainability. Research was conducted on actual motor coach companies. It also researched whether a certification program could help reduce environmental impacts from the motor coach industry by educating operators and executives about the benefits, both financial and environmental, of adopting fuel saving strategies and switching to alterna
Document 3:::
Cellulosic ethanol is ethanol (ethyl alcohol) produced from cellulose (the stringy fiber of a plant) rather than from the plant's seeds or fruit. It can be produced from grasses, wood, algae, or other plants. It is generally discussed for use as a biofuel. The carbon dioxide that plants absorb as they grow offsets some of the carbon dioxide emitted when ethanol made from them is burned, so cellulosic ethanol fuel has the potential to have a lower carbon footprint than fossil fuels.
Interest in cellulosic ethanol is driven by its potential to replace ethanol made from corn or sugarcane. Since these plants are also used for food products, diverting them for ethanol production can cause food prices to rise; cellulose-based sources, on the other hand, generally do not compete with food, since the fibrous parts of plants are mostly inedible to humans. Another potential advantage is the high diversity and abundance of cellulose sources; grasses, trees and algae are found in almost every environment on Earth. Even municipal solid waste components like paper could conceivably be made into ethanol. The main current disadvantage of cellulosic ethanol is its high cost of production, which is more complex and requires more steps than corn-based or sugarcane-based ethanol.
Cellulosic ethanol received significant attention in the 2000s and early 2010s. The United States government in particular funded research into its commercialization and set targets for the proportion of cellulosic ethanol added to vehicle fuel. A large number of new companies specializing in cellulosic ethanol, in addition to many existing companies, invested in pilot-scale production plants. However, the much cheaper manufacturing of grain-based ethanol, along with the low price of oil in the 2010s, meant that cellulosic ethanol was not competitive with these established fuels. As a result, most of the new refineries were closed by the mid-2010s and many of the newly founded companies became insolvent. A f
Document 4:::
The energy content of biofuel is the chemical energy contained in a given biofuel, measured per unit mass of that fuel, as specific energy, or per unit of volume of the fuel, as energy density.
A biofuel is a fuel produced from recently living organisms. Biofuels include bioethanol, an alcohol made by fermentation—often used as a gasoline additive, and biodiesel, which is usually used as a diesel additive. Specific energy is energy per unit mass, which is used to describe the chemical energy content of a fuel, expressed in SI units as joule per kilogram (J/kg) or equivalent units. Energy density is the amount of chemical energy per unit volume of the fuel, expressed in SI units as joule per litre (J/L) or equivalent units.
Energy and CO2 output of common biofuels
The table below includes entries for popular substances already used for their energy, or being discussed for such use.
The second column shows specific energy, the energy content in megajoules per unit of mass in kilograms, useful in understanding the energy that can be extracted from the fuel.
The third column in the table lists energy density, the energy content per liter of volume, which is useful for understanding the space needed for storing the fuel.
The final two columns deal with the carbon footprint of the fuel. The fourth column contains the proportion of CO2 released when the fuel is converted for energy, with respect to its starting mass, and the fifth column lists the energy produced per kilogram of CO2 produced. As a guideline, a higher number in this column is better for the environment. But these numbers do not account for other green house gases released during burning, production, storage, or shipping. For example, methane may have hidden environmental costs that are not reflected in the table.
Notes
Yields of common crops associated with biofuels production
Notes
See also
Eichhornia crassipes#Bioenergy
Syngas
Conversion of units
Energy density
Heat of combustion
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Why are alternative fuels being used in some automobiles?
A. Alternative fuels are at every gas station.
B. Gasoline comes from a limited resource.
C. Alternative fuels cause pollution.
D. Gasoline engines are too expensive to make.
Answer:
|
|
sciq-1950
|
multiple_choice
|
The reaction rate usually increases as the concentration of what increases?
|
[
"mutations",
"complexes",
"reactants",
"generators"
] |
C
|
Relavent Documents:
Document 0:::
An elementary reaction is a chemical reaction in which one or more chemical species react directly to form products in a single reaction step and with a single transition state. In practice, a reaction is assumed to be elementary if no reaction intermediates have been detected or need to be postulated to describe the reaction on a molecular scale. An apparently elementary reaction may be in fact a stepwise reaction, i.e. a complicated sequence of chemical reactions, with reaction intermediates of variable lifetimes.
In a unimolecular elementary reaction, a molecule dissociates or isomerises to form the products(s)
At constant temperature, the rate of such a reaction is proportional to the concentration of the species
In a bimolecular elementary reaction, two atoms, molecules, ions or radicals, and , react together to form the product(s)
The rate of such a reaction, at constant temperature, is proportional to the product of the concentrations of the species and
The rate expression for an elementary bimolecular reaction is sometimes referred to as the Law of Mass Action as it was first proposed by Guldberg and Waage in 1864. An example of this type of reaction is a cycloaddition reaction.
This rate expression can be derived from first principles by using collision theory for ideal gases. For the case of dilute fluids equivalent results have been obtained from simple probabilistic arguments.
According to collision theory the probability of three chemical species reacting simultaneously with each other in a termolecular elementary reaction is negligible. Hence such termolecular reactions are commonly referred as non-elementary reactions and can be broken down into a more fundamental set of bimolecular reactions, in agreement with the law of mass action. It is not always possible to derive overall reaction schemes, but solutions based on rate equations are often possible in terms of steady-state or Michaelis-Menten approximations.
Notes
Chemical kinetics
Phy
Document 1:::
The limiting reagent (or limiting reactant or limiting agent) in a chemical reaction is a reactant that is totally consumed when the chemical reaction is completed. The amount of product formed is limited by this reagent, since the reaction cannot continue without it. If one or more other reagents are present in excess of the quantities required to react with the limiting reagent, they are described as excess reagents or excess reactants (sometimes abbreviated as "xs"), or to be in abundance.
The limiting reagent must be identified in order to calculate the percentage yield of a reaction since the theoretical yield is defined as the amount of product obtained when the limiting reagent reacts completely. Given the balanced chemical equation, which describes the reaction, there are several equivalent ways to identify the limiting reagent and evaluate the excess quantities of other reagents.
Method 1: Comparison of reactant amounts
This method is most useful when there are only two reactants. One reactant (A) is chosen, and the balanced chemical equation is used to determine the amount of the other reactant (B) necessary to react with A. If the amount of B actually present exceeds the amount required, then B is in excess and A is the limiting reagent. If the amount of B present is less than required, then B is the limiting reagent.
Example for two reactants
Consider the combustion of benzene, represented by the following chemical equation:
2 C6H6(l) + 15 O2(g) -> 12 CO2(g) + 6 H2O(l)
This means that 15 moles of molecular oxygen (O2) is required to react with 2 moles of benzene (C6H6)
The amount of oxygen required for other quantities of benzene can be calculated using cross-multiplication (the rule of three). For example,
if 1.5 mol C6H6 is present, 11.25 mol O2 is required:
If in fact 18 mol O2 are present, there will be an excess of (18 - 11.25) = 6.75 mol of unreacted oxygen when all the benzene is consumed. Benzene is then the limiting reagent.
This concl
Document 2:::
In biochemistry, a rate-limiting step is a step that controls the rate of a series of biochemical reactions. The statement is, however, a misunderstanding of how a sequence of enzyme catalyzed reaction steps operate. Rather than a single step controlling the rate, it has been discovered that multiple steps control the rate. Moreover, each controlling step controls the rate to varying degrees.
Blackman (1905) stated as an axiom: "when a process is conditioned as to its rapidity by a number of separate factors, the rate of the process is limited by the pace of the slowest factor." This implies that it should be possible, by studying the behavior of a complicated system such as a metabolic pathway, to characterize a single factor or reaction (namely the slowest), which plays the role of a master or rate-limiting step. In other words, the study of flux control can be simplified to the study of a single enzyme since, by definition, there can only be one 'rate-limiting' step. Since its conception, the 'rate-limiting' step has played a significant role in suggesting how metabolic pathways are controlled. Unfortunately, the notion of a 'rate-limiting' step is erroneous, at least under steady-state conditions. Modern biochemistry textbooks have begun to play down the concept. For example, the seventh edition of Lehninger Principles of Biochemistry explicitly states: "It has now become clear that, in most pathways, the control of flux is distributed among several enzymes, and the extent to which each contributes to the control varies with metabolic circumstances". However, the concept is still incorrectly used in research articles.
Historical perspective
From the 1920s to the 1950s, there were a number of authors who discussed the concept of rate-limiting steps, also known as master reactions. Several authors have stated that the concept of the 'rate-limiting' step is incorrect. Burton (1936) was one of the first to point out that: "In the steady state of reaction chain
Document 3:::
The Hatta number (Ha) was developed by Shirôji Hatta, who taught at Tohoku University. It is a dimensionless parameter that compares the rate of reaction in a liquid film to the rate of diffusion through the film. For a second order reaction (), the maximum rate of reaction assumes that the liquid film is saturated with gas at the interfacial concentration ; thus, the maximum rate of reaction is .
For a reaction order in and order in :
It is an important parameter used in Chemical Reaction Engineering.
Document 4:::
Conversion and its related terms yield and selectivity are important terms in chemical reaction engineering. They are described as ratios of how much of a reactant has reacted (X — conversion, normally between zero and one), how much of a desired product was formed (Y — yield, normally also between zero and one) and how much desired product was formed in ratio to the undesired product(s) (S — selectivity).
There are conflicting definitions in the literature for selectivity and yield, so each author's intended definition should be verified.
Conversion can be defined for (semi-)batch and continuous reactors and as instantaneous and overall conversion.
Assumptions
The following assumptions are made:
The following chemical reaction takes place:
,
where and are the stoichiometric coefficients. For multiple parallel reactions, the definitions can also be applied, either per reaction or using the limiting reaction.
Batch reaction assumes all reactants are added at the beginning.
Semi-Batch reaction assumes some reactants are added at the beginning and the rest fed during the batch.
Continuous reaction assumes reactants are fed and products leave the reactor continuously and in steady state.
Conversion
Conversion can be separated into instantaneous conversion and overall conversion. For continuous processes the two are the same, for batch and semi-batch there are important differences. Furthermore, for multiple reactants, conversion can be defined overall or per reactant.
Instantaneous conversion
Semi-batch
In this setting there are different definitions. One definition regards the instantaneous conversion as the ratio of the instantaneously converted amount to
the amount fed at any point in time:
.
with as the change of moles with time of species i.
This ratio can become larger than 1. It can be used to indicate whether reservoirs are built
up and it is ideally close to 1. When the feed stops, its value is not defined.
In semi-batch polymerisation,
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
The reaction rate usually increases as the concentration of what increases?
A. mutations
B. complexes
C. reactants
D. generators
Answer:
|
|
sciq-3784
|
multiple_choice
|
Spindle fibers form between the centrioles during prophase i of what process?
|
[
"meiosis",
"hydrolysis",
"mitosis",
"photosynthesis"
] |
A
|
Relavent Documents:
Document 0:::
Helical growth is when cells or organs expand, resulting in helical shaped cells or organs and typically including the breakage of symmetry. This is seen in fungi, algae, and other higher plant cells or organs. Helical growth can occur naturally, such as in tendrils or in twining plants. Asymmetry can be caused by changes in pectin or through mutation and result in left- or right-handed helices. Tendril perversion, during which a tendril curves in opposite directions at each end, is seen in many cases. The helical growth of twining plants is based on the circumnutational movement, or circular growth, of stems. Most twining plans have right-handed helices regardless of the plant's growth hemisphere.
Helical growth in single cells, such as the fungi genus Phycomyces or the algae genus Nitella, is thought to be caused by a helical arrangement of structural biological material in the cell wall. In mutant thale cress, helical growth is seen at the organ level. Analysis strongly suggests that cortical microtubules have an important role in controlling the direction of organ expansion. It is unclear how helical growth mutations affect thale cress cell wall assembly.
When seen in spiral3, a conserved GRIP1 gene, a missense mutation causes a left-handed helical organization of cortical microtubules and a severe right-hand helical growth. This mutation compromises interactions between proteins GCP2 and GCP3 in yeast. While the efficiency of microtubule dynamics and nucleation were not noticeably affected, cortical microtubule angles were less narrow and more widely distributed.
Document 1:::
In cell biology, the spindle apparatus is the cytoskeletal structure of eukaryotic cells that forms during cell division to separate sister chromatids between daughter cells. It is referred to as the mitotic spindle during mitosis, a process that produces genetically identical daughter cells, or the meiotic spindle during meiosis, a process that produces gametes with half the number of chromosomes of the parent cell.
Besides chromosomes, the spindle apparatus is composed of hundreds of proteins. Microtubules comprise the most abundant components of the machinery.
Spindle structure
Attachment of microtubules to chromosomes is mediated by kinetochores, which actively monitor spindle formation and prevent premature anaphase onset. Microtubule polymerization and depolymerization dynamic drive chromosome congression. Depolymerization of microtubules generates tension at kinetochores; bipolar attachment of sister kinetochores to microtubules emanating from opposite cell poles couples opposing tension forces, aligning chromosomes at the cell equator and poising them for segregation to daughter cells. Once every chromosome is bi-oriented, anaphase commences and cohesin, which couples sister chromatids, is severed, permitting the transit of the sister chromatids to opposite poles.
The cellular spindle apparatus includes the spindle microtubules, associated proteins, which include kinesin and dynein molecular motors, condensed chromosomes, and any centrosomes or asters that may be present at the spindle poles depending on the cell type. The spindle apparatus is vaguely ellipsoid in cross section and tapers at the ends. In the wide middle portion, known as the spindle midzone, antiparallel microtubules are bundled by kinesins. At the pointed ends, known as spindle poles, microtubules are nucleated by the centrosomes in most animal cells. Acentrosomal or anastral spindles lack centrosomes or asters at the spindle poles, respectively, and occur for example during female meio
Document 2:::
The quiescent centre is a group of cells, up to 1,000 in number, in the form of a hemisphere, with the flat face toward the root tip of vascular plants. It is a region in the apical meristem of a root where cell division proceeds very slowly or not at all, but the cells are capable of resuming meristematic activity when the tissue surrounding them is damaged.
Cells of root apical meristems do not all divide at the same rate. Determinations of relative rates of DNA synthesis show that primary roots of Zea, Vicia and Allium have quiescent centres to the meristems, in which the cells divide rarely or never in the course of normal root growth (Clowes, 1958). Such a quiescent centre includes the cells at the apices of the histogens of both stele and cortex. Its presence can be deduced from the anatomy of the apex in Zea (Clowes, 1958), but not in the other species which lack discrete histogens.
History
In 1953, during the course of analysing the organization and function of the root apices, Frederick Albert Lionel Clowes (born 10 September 1921), at the School of Botany (now Department of Plant Sciences), University of Oxford, proposed the term ‘cytogenerative centre’ to denote ‘the region of an apical meristem from which all future cells are derived’. This term had been suggested to him by Mr Harold K. Pusey, a lecturer in embryology at the Department of Zoology and Comparative Anatomy at the same university. The 1953 paper of Clowes reported results of his experiments on Fagus sylvatica and Vicia faba, in which small oblique and wedge-shaped excisions were made at the tip of the primary root, at the most distal level of the root body, near the boundary with the root cap. The results of these experiments were striking and showed that: the root which grew on following the excision was normal at the undamaged meristem side; the nonexcised meristem portion contributed to the regeneration of the excised portion; the regenerated part of the root had abnormal patterning and
Document 3:::
Ciliogenesis is defined as the building of the cell's antenna (primary cilia) or extracellular fluid mediation mechanism (motile cilium). It includes the assembly and disassembly of the cilia during the cell cycle. Cilia are important organelles of cells and are involved in numerous activities such as cell signaling, processing developmental signals, and directing the flow of fluids such as mucus over and around cells. Due to the importance of these cell processes, defects in ciliogenesis can lead to numerous human diseases related to non-functioning cilia. Ciliogenesis may also play a role in the development of left/right handedness in humans.
Cilia formation
Ciliogenesis occurs through an ordered set of steps. First, the basal bodies from centrioles must migrate to the surface of the cell and attach to the cortex. Along the way, the basal bodies attach to membrane vesicles and the basal body/membrane vesicle complex fuses with the plasma membrane of the cell. Fusion with the plasma membrane is likely what forms the membrane of the cilia. The alignment of the forming cilia is determined by the original positioning and orientation of the basal bodies. Once the alignment is determined, axonemal microtubules extend from the basal body and go beneath the developing ciliary membrane, forming the cilia.
Proteins must be synthesized in the cytoplasm of the cell and cannot be synthesized within cilia. For the cilium to elongate, proteins must be selectively imported from the cytoplasm into the cilium and transported to the tip of the cilium by intraflagellar transport (IFT). Once the cilium is completely formed, it continues to incorporate new tubulin at the tip of the cilia. However, the cilium does not elongate further, because older tubulin is simultaneously degraded. This requires an active mechanism that maintains ciliary length. Impairments in these mechanisms can affect the motility of the cell and cell signaling between cells.
Ciliogenesis types
Mo
Document 4:::
Interkinesis or interphase II is a period of rest that cells of some species enter during meiosis between meiosis I and meiosis II. No DNA replication occurs during interkinesis; however, replication does occur during the interphase I stage of meiosis (See meiosis I). During interkinesis, the spindles of the first meiotic division disassembles and the microtubules reassemble into two new spindles for the second meiotic division. Interkinesis follows telophase I; however, many plants skip telophase I and interkinesis, going immediately into prophase II. Each chromosome still consists of two chromatids. In this stage other organelle number may also increase.
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Spindle fibers form between the centrioles during prophase i of what process?
A. meiosis
B. hydrolysis
C. mitosis
D. photosynthesis
Answer:
|
|
sciq-2492
|
multiple_choice
|
Water seeping into the ground is known as?
|
[
"accumulation",
"invasion",
"infiltration",
"precipitation"
] |
C
|
Relavent Documents:
Document 0:::
Surface runoff (also known as overland flow or terrestrial runoff) is the unconfined flow of water over the ground surface, in contrast to channel runoff (or stream flow). It occurs when excess rainwater, stormwater, meltwater, or other sources, can no longer sufficiently rapidly infiltrate in the soil. This can occur when the soil is saturated by water to its full capacity, and the rain arrives more quickly than the soil can absorb it. Surface runoff often occurs because impervious areas (such as roofs and pavement) do not allow water to soak into the ground. Furthermore, runoff can occur either through natural or human-made processes.
Surface runoff is a major component of the water cycle. It is the primary agent of soil erosion by water. The land area producing runoff that drains to a common point is called a drainage basin.
Runoff that occurs on the ground surface before reaching a channel can be a nonpoint source of pollution, as it can carry human-made contaminants or natural forms of pollution (such as rotting leaves). Human-made contaminants in runoff include petroleum, pesticides, fertilizers and others. Much agricultural pollution is exacerbated by surface runoff, leading to a number of down stream impacts, including nutrient pollution that causes eutrophication.
In addition to causing water erosion and pollution, surface runoff in urban areas is a primary cause of urban flooding, which can result in property damage, damp and mold in basements, and street flooding.
Generation
Surface runoff is defined as precipitation (rain, snow, sleet, or hail) that reaches a surface stream without ever passing below the soil surface. It is distinct from direct runoff, which is runoff that reaches surface streams immediately after rainfall or melting snowfall and excludes runoff generated by the melting of snowpack or glaciers.
Snow and glacier melt occur only in areas cold enough for these to form permanently. Typically snowmelt will peak in the spring and glacie
Document 1:::
In hydrology, throughfall is the process which describes how wet leaves shed excess water onto the ground surface. These drops have greater erosive power because they are heavier than rain drops. Furthermore, where there is a high canopy, falling drops may reach terminal velocity, about , thus maximizing the drop's erosive potential.
Rates of throughfall are higher in areas of forest where the leaves are broad-leaved. This is because the flat leaves allow water to collect. Drip-tips also facilitate throughfall. Rates of throughfall are lower in coniferous forests as conifers can only hold individual droplets of water on their needles.
Throughfall is a crucial process when designing pesticides for foliar application since it will condition their washing and the fate of potential pollutants in the environment.
See also
Stemflow
Canopy interception
Forest floor interception
Tree shape
Notes
Hydrology
Forest ecology
Document 2:::
Evapotranspiration (ET) is the combined processes which move water from the Earth's surface into the atmosphere. It covers both water evaporation (movement of water to the air directly from soil, canopies, and water bodies) and transpiration (evaporation that occurs through the stomata, or openings, in plant leaves). Evapotranspiration is an important part of the local water cycle and climate, and measurement of it plays a key role in agricultural irrigation and water resource management.
Definition of evapotranspiration
Evapotranspiration is a combination of evaporation and transpiration, measured in order to better understand crop water requirements, irrigation scheduling, and watershed management. The two key components of evapotranspiration are:
Evaporation: the movement of water directly to the air from sources such as the soil and water bodies. It can be affected by factors including heat, humidity, solar radiation and wind speed.
Transpiration: the movement of water from root systems, through a plant, and exit into the air as water vapor. This exit occurs through stomata in the plant. Rate of transpiration can be influenced by factors including plant type, soil type, weather conditions and water content, and also cultivation practices.
Evapotranspiration is typically measured in millimeters of water (i.e. volume of water moved per unit area of the Earth's surface) in a set unit of time. Globally, it is estimated that on average between three-fifths and three-quarters of land precipitation is returned to the atmosphere via evapotranspiration.
Evapotranspiration does not, in general, account for other mechanisms which are involved in returning water to the atmosphere, though some of these, such as snow and ice sublimation in regions of high elevation or high latitude, can make a large contribution to atmospheric moisture even under standard conditions.
Factors that impact evapotranspiration levels
Primary factors
Because evaporation and transpiration
Document 3:::
Groundwater is the water present beneath Earth's surface in rock and soil pore spaces and in the fractures of rock formations. About 30 percent of all readily available freshwater in the world is groundwater. A unit of rock or an unconsolidated deposit is called an aquifer when it can yield a usable quantity of water. The depth at which soil pore spaces or fractures and voids in rock become completely saturated with water is called the water table. Groundwater is recharged from the surface; it may discharge from the surface naturally at springs and seeps, and can form oases or wetlands. Groundwater is also often withdrawn for agricultural, municipal, and industrial use by constructing and operating extraction wells. The study of the distribution and movement of groundwater is hydrogeology, also called groundwater hydrology.
Typically, groundwater is thought of as water flowing through shallow aquifers, but, in the technical sense, it can also contain soil moisture, permafrost (frozen soil), immobile water in very low permeability bedrock, and deep geothermal or oil formation water. Groundwater is hypothesized to provide lubrication that can possibly influence the movement of faults. It is likely that much of Earth's subsurface contains some water, which may be mixed with other fluids in some instances.
Groundwater is often cheaper, more convenient and less vulnerable to pollution than surface water. Therefore, it is commonly used for public water supplies. For example, groundwater provides the largest source of usable water storage in the United States, and California annually withdraws the largest amount of groundwater of all the states. Underground reservoirs contain far more water than the capacity of all surface reservoirs and lakes in the US, including the Great Lakes. Many municipal water supplies are derived solely from groundwater. Over 2 billion people rely on it as their primary water source worldwide.
Human use of groundwater causes environmental prob
Document 4:::
In hydrology, pipeflow is a type of subterranean water flow where water travels along cracks in the soil or old root systems found in above ground vegetation.
In such soils which have a high vegetation content water is able to travel along the 'pipes', allowing water to travel faster than throughflow. Here, water can move at speeds between 50 and 500 m/h.
Hydrology
Aquatic ecology
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Water seeping into the ground is known as?
A. accumulation
B. invasion
C. infiltration
D. precipitation
Answer:
|
|
sciq-11631
|
multiple_choice
|
Different regions of the cerebral cortex can be associated with particular functions, a concept known as what?
|
[
"localization of function",
"reversal of function",
"expressiveness of function",
"cytoplasm of function"
] |
A
|
Relavent Documents:
Document 0:::
The following outline is provided as an overview of and topical guide to brain mapping:
Brain mapping – set of neuroscience techniques predicated on the mapping of (biological) quantities or properties onto spatial representations of the (human or non-human) brain resulting in maps. Brain mapping is further defined as the study of the anatomy and function of the brain and spinal cord through the use of imaging (including intra-operative, microscopic, endoscopic and multi-modality imaging), immunohistochemistry, molecular and optogenetics, stem cell and cellular biology, engineering (material, electrical and biomedical), neurophysiology and nanotechnology.
Broad scope
History of neuroscience
History of neurology
Brain mapping
Human brain
Neuroscience
Nervous system.
The neuron doctrine
Neuron doctrine – A set of carefully constructed elementary set of observations regarding neurons. For more granularity, more current, and more advanced topics, see the cellular level section
Asserts that neurons fall under the broader cell theory, which postulates:
All living organisms are composed of one or more cells.
The cell is the basic unit of structure, function, and organization in all organisms.
All cells come from preexisting, living cells.
The Neuron doctrine postulates several elementary aspects of neurons:
The brain is made up of individual cells (neurons) that contain specialized features such as dendrites, a cell body, and an axon.
Neurons are cells differentiable from other tissues in the body.
Neurons differ in size, shape, and structure according to their location or functional specialization.
Every neuron has a nucleus, which is the trophic center of the cell (The part which must have access to nutrition). If the cell is divided, only the portion containing the nucleus will survive.
Nerve fibers are the result of cell processes and the outgrowths of nerve cells. (Several axons are bound together to form one nerve fibril. See also: Neurofilament.
Document 1:::
Cognitive skills, also called cognitive functions, cognitive abilities or cognitive capacities, are brain-based skills which are needed in acquisition of knowledge, manipulation of information and reasoning. They have more to do with the mechanisms of how people learn, remember, solve problems and pay attention, rather than with actual knowledge. Cognitive skills or functions encompass the domains of perception, attention, memory, learning, decision making, and language abilities.
Specialisation of functions
Cognitive science has provided theories of how the brain works, and these have been of great interest to researchers who work in the empirical fields of brain science. A fundamental question is whether cognitive functions, for example visual processing and language, are autonomous modules, or to what extent the functions depend on each other. Research evidence points towards a middle position, and it is now generally accepted that there is a degree of modularity in aspects of brain organisation. In other words, cognitive skills or functions are specialised, but they also overlap or interact with each other. Deductive reasoning, on the other hand, has been shown to be related to either visual or linguistic processing, depending on the task; although there are also aspects that differ from them. All in all, research evidence does not provide strong support for classical models of cognitive psychology.
Cognitive functioning
Cognitive functioning refers to a person's ability to process thoughts. It is defined as "the ability of an individual to perform the various mental activities most closely associated with learning and problem-solving. Examples include the verbal, spatial, psychomotor, and processing-speed ability." Cognition mainly refers to things like memory, speech, and the ability to learn new information. The brain is usually capable of learning new skills in the aforementioned areas, typically in early childhood, and of developing personal thoughts an
Document 2:::
The following outline is provided as an overview of and topical guide to neuroscience:
Neuroscience is the scientific study of the structure and function of the nervous system. It encompasses the branch of biology that deals with the anatomy, biochemistry, molecular biology, and physiology of neurons and neural circuits. It also encompasses cognition, and human behavior. Neuroscience has multiple concepts that each relate to learning abilities and memory functions. Additionally, the brain is able to transmit signals that cause conscious/unconscious behaviors that are responses verbal or non-verbal. This allows people to communicate with one another.
Branches of neuroscience
Neurophysiology
Neurophysiology is the study of the function (as opposed to structure) of the nervous system.
Brain mapping
Electrophysiology
Extracellular recording
Intracellular recording
Brain stimulation
Electroencephalography
Intermittent rhythmic delta activity
:Category: Neurophysiology
:Category: Neuroendocrinology
:Neuroendocrinology
Neuroanatomy
Neuroanatomy is the study of the anatomy of nervous tissue and neural structures of the nervous system.
Immunostaining
:Category: Neuroanatomy
Neuropharmacology
Neuropharmacology is the study of how drugs affect cellular function in the nervous system.
Drug
Psychoactive drug
Anaesthetic
Narcotic
Behavioral neuroscience
Behavioral neuroscience, also known as biological psychology, biopsychology, or psychobiology, is the application of the principles of biology to the study of mental processes and behavior in human and non-human animals.
Neuroethology
Developmental neuroscience
Developmental neuroscience aims to describe the cellular basis of brain development and to address the underlying mechanisms. The field draws on both neuroscience and developmental biology to provide insight into the cellular and molecular mechanisms by which complex nervous systems develop.
Aging and memory
Cognitive neuroscience
Cognitive ne
Document 3:::
The temporal lobe is one of the four major lobes of the cerebral cortex in the brain of mammals. The temporal lobe is located beneath the lateral fissure on both cerebral hemispheres of the mammalian brain.
The temporal lobe is involved in processing sensory input into derived meanings for the appropriate retention of visual memory, language comprehension, and emotion association.
Temporal refers to the head's temples.
Structure
The temporal lobe consists of structures that are vital for declarative or long-term memory. Declarative (denotative) or explicit memory is conscious memory divided into semantic memory (facts) and episodic memory (events). Medial temporal lobe structures that are critical for long-term memory include the hippocampus, along with the surrounding hippocampal region consisting of the perirhinal, parahippocampal, and entorhinal neocortical regions. The hippocampus is critical for memory formation, and the surrounding medial temporal cortex is currently theorized to be critical for memory storage. The prefrontal and visual cortices are also involved in explicit memory.
Research has shown that lesions in the hippocampus of monkeys results in limited impairment of function, whereas extensive lesions that include the hippocampus and the medial temporal cortex result in severe impairment.
Function
Visual memories
The temporal lobe communicates with the hippocampus and plays a key role in the formation of explicit long-term memory modulated by the amygdala.
Processing sensory input
Auditory Adjacent areas in the superior, posterior, and lateral parts of the temporal lobes are involved in high-level auditory processing. The temporal lobe is involved in primary auditory perception, such as hearing, and holds the primary auditory cortex. The primary auditory cortex receives sensory information from the ears and secondary areas process the information into meaningful units such as speech and words. The superior temporal gyrus includes an area (wit
Document 4:::
Computational neuroscience (also known as theoretical neuroscience or mathematical neuroscience) is a branch of neuroscience which employs mathematics, computer science, theoretical analysis and abstractions of the brain to understand the principles that govern the development, structure, physiology and cognitive abilities of the nervous system.
Computational neuroscience employs computational simulations to validate and solve mathematical models, and so can be seen as a sub-field of theoretical neuroscience; however, the two fields are often synonymous. The term mathematical neuroscience is also used sometimes, to stress the quantitative nature of the field.
Computational neuroscience focuses on the description of biologically plausible neurons (and neural systems) and their physiology and dynamics, and it is therefore not directly concerned with biologically unrealistic models used in connectionism, control theory, cybernetics, quantitative psychology, machine learning, artificial neural networks, artificial intelligence and computational learning theory;
although mutual inspiration exists and sometimes there is no strict limit between fields, with model abstraction in computational neuroscience depending on research scope and the granularity at which biological entities are analyzed.
Models in theoretical neuroscience are aimed at capturing the essential features of the biological system at multiple spatial-temporal scales, from membrane currents, and chemical coupling via network oscillations, columnar and topographic architecture, nuclei, all the way up to psychological faculties like memory, learning and behavior. These computational models frame hypotheses that can be directly tested by biological or psychological experiments.
History
The term 'computational neuroscience' was introduced by Eric L. Schwartz, who organized a conference, held in 1985 in Carmel, California, at the request of the Systems Development Foundation to provide a summary of the curr
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Different regions of the cerebral cortex can be associated with particular functions, a concept known as what?
A. localization of function
B. reversal of function
C. expressiveness of function
D. cytoplasm of function
Answer:
|
|
sciq-4305
|
multiple_choice
|
Like other bryophytes, moss plants spend most of their life cycle as what?
|
[
"protozoa",
"copepods",
"zygotes",
"gametophytes"
] |
D
|
Relavent Documents:
Document 0:::
Occurrence
40% of mosses are monoicious.
Bryophyte sexuality
Bryophytes have life cycles that are gametophyte dominated. The longer lived, more prominent autotrophic plant is the gametophyte. The sporophyte in mosses and liverworts consists of an unbranched stalk (a seta) bearing a single
Document 1:::
Bryophytes () are a group of land plants, sometimes treated as a taxonomic division, that contains three groups of non-vascular land plants (embryophytes): the liverworts, hornworts and mosses. In the strict sense, Bryophyta consists of the mosses only. Bryophytes are characteristically limited in size and prefer moist habitats although they can survive in drier environments. The bryophytes consist of about 20,000 plant species. Bryophytes produce enclosed reproductive structures (gametangia and sporangia), but they do not produce flowers or seeds. They reproduce sexually by spores and asexually by fragmentation or the production of gemmae. Though bryophytes were considered a paraphyletic group in recent years, almost all of the most recent phylogenetic evidence supports the monophyly of this group, as originally classified by Wilhelm Schimper in 1879. The term bryophyte comes .
Features
The defining features of bryophytes are:
Their life cycles are dominated by a multicellular gametophyte stage
Their sporophytes are unbranched
They do not have a true vascular tissue containing lignin (although some have specialized tissues for the transport of water)
Habitat
Bryophytes exist in a wide variety of habitats. They can be found growing in a range of temperatures (cold arctics and in hot deserts), elevations (sea-level to alpine), and moisture (dry deserts to wet rain forests). Bryophytes can grow where vascularized plants cannot because they do not depend on roots for uptake of nutrients from soil. Bryophytes can survive on rocks and bare soil.
Life cycle
Like all land plants (embryophytes), bryophytes have life cycles with alternation of generations. In each cycle, a haploid gametophyte, each of whose cells contains a fixed number of unpaired chromosomes, alternates with a diploid sporophyte, whose cells contain two sets of paired chromosomes. Gametophytes produce haploid sperm and eggs which fuse to form diploid zygotes that grow into sporophytes. Sporophytes
Document 2:::
The International Moss Stock Center (IMSC) is a biorepository which is specialized in collecting, preserving and distributing moss plants of a high value of scientific research. The IMSC is located at the Faculty of Biology, Department of Plant Biotechnology, at the Albert-Ludwigs-University of Freiburg, Germany.
Moss collection
The moss collection of the IMSC currently includes various ecotypes of Physcomitrella patens, Physcomitrium and Funaria as well as several transgenic and mutant lines of Physcomitrella patens, including knockout mosses.
Storage conditions
The long-term storage of moss samples in the IMSC is carried out via cryopreservation in the gas phase of liquid nitrogen at temperatures below −135 °C in special freezer containers.
It has been shown for Physcomitrella patens that the regeneration rate after cryopreservation is 100%.
Trackable accession numbers which may be used for citation purposes in
publications are automatically assigned to all samples.
Financial support
The IMSC is supported financially by the Chair Plant Biotechnology of Prof. Ralf Reski and the Centre for Biological Signalling Studies (bioss).
Document 3:::
Tree moss is a common name for several organisms and may refer to:
Climacium, a genus of mosses which resemble miniature trees
Climacium dendroides, a common species of Climacium
Evernia, a genus of lichens which grow on trees
Usnea, a genus of lichens which grow on trees
Document 4:::
Non-vascular plants are plants without a vascular system consisting of xylem and phloem. Instead, they may possess simpler tissues that have specialized functions for the internal transport of water.
Non-vascular plants include two distantly related groups:
Bryophytes, an informal group that taxonomists treat as three separate land-plant divisions, namely: Bryophyta (mosses), Marchantiophyta (liverworts), and Anthocerotophyta (hornworts). In all bryophytes, the primary plants are the haploid gametophytes, with the only diploid portion being the attached sporophyte, consisting of a stalk and sporangium. Because these plants lack lignified water-conducting tissues, they cannot become as tall as most vascular plants.
Algae, especially green algae. The algae consist of several unrelated groups. Only the groups included in the Viridiplantae are still considered relatives of land plants.
These groups are sometimes called "lower plants", referring to their status as the earliest plant groups to evolve, but the usage is imprecise since both groups are polyphyletic and may be used to include vascular cryptogams, such as the ferns and fern allies that reproduce using spores. Non-vascular plants are often among the first species to move into new and inhospitable territories, along with prokaryotes and protists, and thus function as pioneer species.
Non-vascular plants do not have a wide variety of specialized tissue types. Mosses and leafy liverworts have structures called phyllids that resemble leaves, but only consist of single sheets of cells with no internal air spaces, no cuticle or stomata, and no xylem or phloem. Consequently, phyllids are unable to control the rate of water loss from their tissues and are said to be poikilohydric. Some liverworts, such as Marchantia, have a cuticle, and the sporophytes of mosses have both cuticles and stomata, which were important in the evolution of land plants.
All land plants have a life cycle with an alternation of generatio
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Like other bryophytes, moss plants spend most of their life cycle as what?
A. protozoa
B. copepods
C. zygotes
D. gametophytes
Answer:
|
|
sciq-316
|
multiple_choice
|
What is likely to happen to a parasite if it kills its host?
|
[
"it dies",
"it mutates",
"it adapts",
"it thrives"
] |
A
|
Relavent Documents:
Document 0:::
The SAT Subject Test in Biology was the name of a one-hour multiple choice test given on biology by the College Board. A student chose whether to take the test depending upon college entrance requirements for the schools in which the student is planning to apply. Until 1994, the SAT Subject Tests were known as Achievement Tests; and from 1995 until January 2005, they were known as SAT IIs. Of all SAT subject tests, the Biology E/M test was the only SAT II that allowed the test taker a choice between the ecological or molecular tests. A set of 60 questions was taken by all test takers for Biology and a choice of 20 questions was allowed between either the E or M tests. This test was graded on a scale between 200 and 800. The average for Molecular is 630 while Ecological is 591.
On January 19 2021, the College Board discontinued all SAT Subject tests, including the SAT Subject Test in Biology E/M. This was effective immediately in the United States, and the tests were to be phased out by the following summer for international students. This was done as a response to changes in college admissions due to the impact of the COVID-19 pandemic on education.
Format
This test had 80 multiple-choice questions that were to be answered in one hour. All questions had five answer choices. Students received one point for each correct answer, lost ¼ of a point for each incorrect answer, and received 0 points for questions left blank. The student's score was based entirely on his or her performance in answering the multiple-choice questions.
The questions covered a broad range of topics in general biology. There were more specific questions related respectively on ecological concepts (such as population studies and general Ecology) on the E test and molecular concepts such as DNA structure, translation, and biochemistry on the M test.
Preparation
The College Board suggested a year-long course in biology at the college preparatory level, as well as a one-year course in algebra, a
Document 1:::
Advanced Placement (AP) Biology (also known as AP Bio) is an Advanced Placement biology course and exam offered by the College Board in the United States. For the 2012–2013 school year, the College Board unveiled a new curriculum with a greater focus on "scientific practices".
This course is designed for students who wish to pursue an interest in the life sciences. The College Board recommends successful completion of high school biology and high school chemistry before commencing AP Biology, although the actual prerequisites vary from school to school and from state to state. This course, nevertheless, is considered very challenging and one of the most difficult AP classes, as shown with AP Finals grade distributions.
Topic outline
The exam covers the following 8 units. The percentage indicates the portion of the multiple-choice section of the exam focused on each content area:
The course is based on and tests six skills, called scientific practices which include:
In addition to the topics above, students are required to be familiar with general lab procedure. Students should know how to collect data, analyze data to form conclusions, and apply those conclusions.
Exam
Students are allowed to use a four-function, scientific, or graphing calculator.
The exam has two sections: a 90 minute multiple choice section and a 90 minute free response section. There are 60 multiple choice questions and six free responses, two long and four short. Both sections are worth 50% of the score.
Score distribution
Commonly used textbooks
Biology, AP Edition by Sylvia Mader (2012, hardcover )
Life: The Science of Biology (Sadava, Heller, Orians, Purves, and Hillis, )
Campbell Biology AP Ninth Edition (Reece, Urry, Cain, Wasserman, Minorsky, and Andrew Jackson )
See also
Glossary of biology
A.P Bio (TV Show)
Document 2:::
MicrobeLibrary is a permanent collection of over 1400 original peer-reviewed resources for teaching undergraduate microbiology. It is provided by the American Society for Microbiology, Washington DC, United States.
Contents include curriculum activities; images and animations; reviews of books, websites and other resources; and articles from Focus on Microbiology Education, Microbiology Education and Microbe. Around 40% of the materials are free to educators and students, the remainder require a subscription. the service is suspended with the message to:
"Please check back with us in 2017".
External links
MicrobeLibrary
Microbiology
Document 3:::
Progress tests are longitudinal, feedback oriented educational assessment tools for the evaluation of development and sustainability of cognitive knowledge during a learning process. A progress test is a written knowledge exam (usually involving multiple choice questions) that is usually administered to all students in the "A" program at the same time and at regular intervals (usually twice to four times yearly) throughout the entire academic program. The test samples the complete knowledge domain expected of new graduates upon completion of their courses, regardless of the year level of the student). The differences between students’ knowledge levels show in the test scores; the further a student has progressed in the curriculum the higher the scores. As a result, these resultant scores provide a longitudinal, repeated measures, curriculum-independent assessment of the objectives (in knowledge) of the entire programme.
History
Since its inception in the late 1970s at both Maastricht University and the University of Missouri–Kansas City independently, the progress test of applied knowledge has been increasingly used in medical and health sciences programs across the globe. They are well established and increasingly used in medical education in both undergraduate and postgraduate medical education. They are used formatively and summatively.
Use in academic programs
The progress test is currently used by national progress test consortia in the United Kingdom, Italy, The Netherlands, in Germany (including Austria), and in individual schools in Africa, Saudi Arabia, South East Asia, the Caribbean, Australia, New Zealand, Sweden, Finland, UK, and the USA. The National Board of Medical Examiners in the USA also provides progress testing in various countries The feasibility of an international approach to progress testing has been recently acknowledged and was first demonstrated by Albano et al. in 1996, who compared test scores across German, Dutch and Italian medi
Document 4:::
GRE Subject Biochemistry, Cell and Molecular Biology was a standardized exam provided by ETS (Educational Testing Service) that was discontinued in December 2016. It is a paper-based exam and there are no computer-based versions of it. ETS places this exam three times per year: once in April, once in October and once in November. Some graduate programs in the United States recommend taking this exam, while others require this exam score as a part of the application to their graduate programs. ETS sends a bulletin with a sample practice test to each candidate after registration for the exam. There are 180 questions within the biochemistry subject test.
Scores are scaled and then reported as a number between 200 and 990; however, in recent versions of the test, the maximum and minimum reported scores have been 760 (corresponding to the 99 percentile) and 320 (1 percentile) respectively. The mean score for all test takers from July, 2009, to July, 2012, was 526 with a standard deviation of 95.
After learning that test content from editions of the GRE® Biochemistry, Cell and Molecular Biology (BCM) Test has been compromised in Israel, ETS made the decision not to administer this test worldwide in 2016–17.
Content specification
Since many students who apply to graduate programs in biochemistry do so during the first half of their fourth year, the scope of most questions is largely that of the first three years of a standard American undergraduate biochemistry curriculum. A sampling of test item content is given below:
Biochemistry (36%)
A Chemical and Physical Foundations
Thermodynamics and kinetics
Redox states
Water, pH, acid-base reactions and buffers
Solutions and equilibria
Solute-solvent interactions
Chemical interactions and bonding
Chemical reaction mechanisms
B Structural Biology: Structure, Assembly, Organization and Dynamics
Small molecules
Macromolecules (e.g., nucleic acids, polysaccharides, proteins and complex lipids)
Supramolecular complexes (e.g.
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is likely to happen to a parasite if it kills its host?
A. it dies
B. it mutates
C. it adapts
D. it thrives
Answer:
|
|
sciq-6783
|
multiple_choice
|
Part of the scientific process, these are statistical probabilities rather than certainties?
|
[
"results",
"assumptions",
"predictions",
"Hypothesis"
] |
C
|
Relavent Documents:
Document 0:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 1:::
Probability and statistics are two closely related fields in mathematics, sometimes combined for academic purposes. They are covered in several articles:
Probability
Statistics
Glossary of probability and statistics
Notation in probability and statistics
Timeline of probability and statistics
Document 2:::
Advanced Placement (AP) Statistics (also known as AP Stats) is a college-level high school statistics course offered in the United States through the College Board's Advanced Placement program. This course is equivalent to a one semester, non-calculus-based introductory college statistics course and is normally offered to sophomores, juniors and seniors in high school.
One of the College Board's more recent additions, the AP Statistics exam was first administered in May 1996 to supplement the AP program's math offerings, which had previously consisted of only AP Calculus AB and BC. In the United States, enrollment in AP Statistics classes has increased at a higher rate than in any other AP class.
Students may receive college credit or upper-level college course placement upon passing the three-hour exam ordinarily administered in May. The exam consists of a multiple-choice section and a free-response section that are both 90 minutes long. Each section is weighted equally in determining the students' composite scores.
History
The Advanced Placement program has offered students the opportunity to pursue college-level courses while in high school. Along with the Educational Testing Service, the College Board administered the first AP Statistics exam in May 1997. The course was first taught to students in the 1996-1997 academic year. Prior to that, the only mathematics courses offered in the AP program included AP Calculus AB and BC. Students who didn't have a strong background in college-level math, however, found the AP Calculus program inaccessible and sometimes declined to take a math course in their senior year. Since the number of students required to take statistics in college is almost as large as the number of students required to take calculus, the College Board decided to add an introductory statistics course to the AP program. Since the prerequisites for such a program doesn't require mathematical concepts beyond those typically taught in a second-year al
Document 3:::
In mathematical psychology and education theory, a knowledge space is a combinatorial structure used to formulate mathematical models describing the progression of a human learner. Knowledge spaces were introduced in 1985 by Jean-Paul Doignon and Jean-Claude Falmagne, and remain in extensive use in the education theory. Modern applications include two computerized tutoring systems, ALEKS and the defunct RATH.
Formally, a knowledge space assumes that a domain of knowledge is a collection of concepts or skills, each of which must be eventually mastered. Not all concepts are interchangeable; some require other concepts as prerequisites. Conversely, competency at one skill may ease the acquisition of another through similarity. A knowledge space marks out which collections of skills are feasible: they can be learned without mastering any other skills. Under reasonable assumptions, the collection of feasible competencies forms the mathematical structure known as an antimatroid.
Researchers and educators usually explore the structure of a discipline's knowledge space as a latent class model.
Motivation
Knowledge Space Theory attempts to address shortcomings of standardized testing when used in educational psychometry. Common tests, such as the SAT and ACT, compress a student's knowledge into a very small range of ordinal ranks, in the process effacing the conceptual dependencies between questions. Consequently, the tests cannot distinguish between true understanding and guesses, nor can they identify a student's particular weaknesses, only the general proportion of skills mastered. The goal of knowledge space theory is to provide a language by which exams can communicate
What the student can do and
What the student is ready to learn.
Model structure
Knowledge Space Theory-based models presume that an educational subject can be modeled as a finite set of concepts, skills, or topics. Each feasible state of knowledge about is then a subset of ; the set of
Document 4:::
The mathematical sciences are a group of areas of study that includes, in addition to mathematics, those academic disciplines that are primarily mathematical in nature but may not be universally considered subfields of mathematics proper.
Statistics, for example, is mathematical in its methods but grew out of bureaucratic and scientific observations, which merged with inverse probability and then grew through applications in some areas of physics, biometrics, and the social sciences to become its own separate, though closely allied, field. Theoretical astronomy, theoretical physics, theoretical and applied mechanics, continuum mechanics, mathematical chemistry, actuarial science, computer science, computational science, data science, operations research, quantitative biology, control theory, econometrics, geophysics and mathematical geosciences are likewise other fields often considered part of the mathematical sciences.
Some institutions offer degrees in mathematical sciences (e.g. the United States Military Academy, Stanford University, and University of Khartoum) or applied mathematical sciences (for example, the University of Rhode Island).
See also
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Part of the scientific process, these are statistical probabilities rather than certainties?
A. results
B. assumptions
C. predictions
D. Hypothesis
Answer:
|
|
sciq-3880
|
multiple_choice
|
What type energy sources emit fewer or no greenhouse gases?
|
[
"renewable",
"animal",
"gas",
"fossil"
] |
A
|
Relavent Documents:
Document 0:::
The indirect land use change impacts of biofuels, also known as ILUC or iLUC (pronounced as i-luck), relates to the unintended consequence of releasing more carbon emissions due to land-use changes around the world induced by the expansion of croplands for ethanol or biodiesel production in response to the increased global demand for biofuels.
As farmers worldwide respond to higher crop prices in order to maintain the global food supply-and-demand balance, pristine lands are cleared to replace the food crops that were diverted elsewhere to biofuels' production. Because natural lands, such as rainforests and grasslands, store carbon in their soil and biomass as plants grow each year, clearance of wilderness for new farms translates to a net increase in greenhouse gas emissions. Due to this off-site change in the carbon stock of the soil and the biomass, indirect land use change has consequences in the greenhouse gas (GHG) balance of a biofuel.
Other authors have also argued that indirect land use changes produce other significant social and environmental impacts, affecting biodiversity, water quality, food prices and supply, land tenure, worker migration, and community and cultural stability.
History
The estimates of carbon intensity for a given biofuel depend on the assumptions regarding several variables. As of 2008, multiple full life cycle studies had found that corn ethanol, cellulosic ethanol and Brazilian sugarcane ethanol produce lower greenhouse gas emissions than gasoline. None of these studies, however, considered the effects of indirect land-use changes, and though land use impacts were acknowledged, estimation was considered too complex and difficult to model. A controversial paper published in February 2008 in Sciencexpress by a team led by Searchinger from Princeton University concluded that such effects offset the (positive) direct effects of both corn and cellulosic ethanol and that Brazilian sugarcane performed better, but still resulted in a sma
Document 1:::
Twisted: The Distorted Mathematics of Greenhouse Denial is a 2007 book by Ian G. Enting, who is the Professorial Research Fellow in the ARC Centre of Excellence for Mathematics and Statistics of Complex Systems (MASCOS) based at the University of Melbourne. The book analyses the arguments of climate change deniers and the use and presentation of statistics. Enting contends there are contradictions in their various arguments. The author also presents calculations of the actual emission levels that would be required to stabilise CO2 concentrations. This is an update of calculations that he contributed to the pre-Kyoto IPCC report on Radiative Forcing of Climate.
See also
Climate change
Greenhouse effect
Radiative forcing
Document 2:::
Energy modeling or energy system modeling is the process of building computer models of energy systems in order to analyze them. Such models often employ scenario analysis to investigate different assumptions about the technical and economic conditions at play. Outputs may include the system feasibility, greenhouse gas emissions, cumulative financial costs, natural resource use, and energy efficiency of the system under investigation. A wide range of techniques are employed, ranging from broadly economic to broadly engineering. Mathematical optimization is often used to determine the least-cost in some sense. Models can be international, regional, national, municipal, or stand-alone in scope. Governments maintain national energy models for energy policy development.
Energy models are usually intended to contribute variously to system operations, engineering design, or energy policy development. This page concentrates on policy models. Individual building energy simulations are explicitly excluded, although they too are sometimes called energy models. IPCC-style integrated assessment models, which also contain a representation of the world energy system and are used to examine global transformation pathways through to 2050 or 2100 are not considered here in detail.
Energy modeling has increased in importance as the need for climate change mitigation has grown in importance. The energy supply sector is the largest contributor to global greenhouse gas emissions. The IPCC reports that climate change mitigation will require a fundamental transformation of the energy supply system, including the substitution of unabated (not captured by CCS) fossil fuel conversion technologies by low-GHG alternatives.
Model types
A wide variety of model types are in use. This section attempts to categorize the key types and their usage. The divisions provided are not hard and fast and mixed-paradigm models exist. In addition, the results from more general models can be
Document 3:::
Roland Geyer is professor of industrial ecology at the Bren School of Environmental Science and Management, University of California at Santa Barbara. He is a specialist in the ecological impact of plastics.
In March 2021, Geyer wrote in The Guardian that humanity should ban fossil fuels, just at it had earlier banned tetraethyllead (TEL) and chlorofluorocarbons (CFC).
Document 4:::
Energy quality is a measure of the ease with which a form of energy can be converted to useful work or to another form of energy: i.e. its content of thermodynamic free energy. A high quality form of energy has a high content of thermodynamic free energy, and therefore a high proportion of it can be converted to work; whereas with low quality forms of energy, only a small proportion can be converted to work, and the remainder is dissipated as heat. The concept of energy quality is also used in ecology, where it is used to track the flow of energy between different trophic levels in a food chain and in thermoeconomics, where it is used as a measure of economic output per unit of energy. Methods of evaluating energy quality often involve developing a ranking of energy qualities in hierarchical order.
Examples: Industrialization, Biology
The consideration of energy quality was a fundamental driver of industrialization from the 18th through 20th centuries. Consider for example the industrialization of New England in the 18th century. This refers to the construction of textile mills containing power looms for weaving cloth. The simplest, most economical and straightforward source of energy was provided by water wheels, extracting energy from a millpond behind a dam on a local creek. If another nearby landowner also decided to build a mill on the same creek, the construction of their dam would lower the overall hydraulic head to power the existing waterwheel, thus hurting power generation and efficiency. This eventually became an issue endemic to the entire region, reducing the overall profitability of older mills as newer ones were built. The search for higher quality energy was a major impetus throughout the 19th and 20th centuries. For example, burning coal to make steam to generate mechanical energy would not have been imaginable in the 18th century; by the end of the 19th century, the use of water wheels was long outmoded. Similarly, the quality of energy from elec
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What type energy sources emit fewer or no greenhouse gases?
A. renewable
B. animal
C. gas
D. fossil
Answer:
|
|
ai2_arc-155
|
multiple_choice
|
A girl walked for 30 minutes. She noticed that she traveled farther in the first 15 minutes of her walk than in the second 15 minutes. What can she conclude about her walk?
|
[
"She walked over many hills.",
"Her average speed was faster during the first half of her walk.",
"She walked in two different directions.",
"She was walking at a constant speed."
] |
B
|
Relavent Documents:
Document 0:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 1:::
Progress tests are longitudinal, feedback oriented educational assessment tools for the evaluation of development and sustainability of cognitive knowledge during a learning process. A progress test is a written knowledge exam (usually involving multiple choice questions) that is usually administered to all students in the "A" program at the same time and at regular intervals (usually twice to four times yearly) throughout the entire academic program. The test samples the complete knowledge domain expected of new graduates upon completion of their courses, regardless of the year level of the student). The differences between students’ knowledge levels show in the test scores; the further a student has progressed in the curriculum the higher the scores. As a result, these resultant scores provide a longitudinal, repeated measures, curriculum-independent assessment of the objectives (in knowledge) of the entire programme.
History
Since its inception in the late 1970s at both Maastricht University and the University of Missouri–Kansas City independently, the progress test of applied knowledge has been increasingly used in medical and health sciences programs across the globe. They are well established and increasingly used in medical education in both undergraduate and postgraduate medical education. They are used formatively and summatively.
Use in academic programs
The progress test is currently used by national progress test consortia in the United Kingdom, Italy, The Netherlands, in Germany (including Austria), and in individual schools in Africa, Saudi Arabia, South East Asia, the Caribbean, Australia, New Zealand, Sweden, Finland, UK, and the USA. The National Board of Medical Examiners in the USA also provides progress testing in various countries The feasibility of an international approach to progress testing has been recently acknowledged and was first demonstrated by Albano et al. in 1996, who compared test scores across German, Dutch and Italian medi
Document 2:::
The SAT Subject Test in Biology was the name of a one-hour multiple choice test given on biology by the College Board. A student chose whether to take the test depending upon college entrance requirements for the schools in which the student is planning to apply. Until 1994, the SAT Subject Tests were known as Achievement Tests; and from 1995 until January 2005, they were known as SAT IIs. Of all SAT subject tests, the Biology E/M test was the only SAT II that allowed the test taker a choice between the ecological or molecular tests. A set of 60 questions was taken by all test takers for Biology and a choice of 20 questions was allowed between either the E or M tests. This test was graded on a scale between 200 and 800. The average for Molecular is 630 while Ecological is 591.
On January 19 2021, the College Board discontinued all SAT Subject tests, including the SAT Subject Test in Biology E/M. This was effective immediately in the United States, and the tests were to be phased out by the following summer for international students. This was done as a response to changes in college admissions due to the impact of the COVID-19 pandemic on education.
Format
This test had 80 multiple-choice questions that were to be answered in one hour. All questions had five answer choices. Students received one point for each correct answer, lost ¼ of a point for each incorrect answer, and received 0 points for questions left blank. The student's score was based entirely on his or her performance in answering the multiple-choice questions.
The questions covered a broad range of topics in general biology. There were more specific questions related respectively on ecological concepts (such as population studies and general Ecology) on the E test and molecular concepts such as DNA structure, translation, and biochemistry on the M test.
Preparation
The College Board suggested a year-long course in biology at the college preparatory level, as well as a one-year course in algebra, a
Document 3:::
Test equating traditionally refers to the statistical process of determining comparable scores on different forms of an exam. It can be accomplished using either classical test theory or item response theory.
In item response theory, equating is the process of placing scores from two or more parallel test forms onto a common score scale. The result is that scores from two different test forms can be compared directly, or treated as though they came from the same test form. When the tests are not parallel, the general process is called linking. It is the process of equating the units and origins of two scales on which the abilities of students have been estimated from results on different tests. The process is analogous to equating degrees Fahrenheit with degrees Celsius by converting measurements from one scale to the other. The determination of comparable scores is a by-product of equating that results from equating the scales obtained from test results.
Purpose
Suppose that Dick and Jane both take a test to become licensed in a certain profession. Because the high stakes (you get to practice the profession if you pass the test) may create a temptation to cheat, the organization that oversees the test creates two forms. If we know that Dick scored 60% on form A and Jane scored 70% on form B, do we know for sure which one has a better grasp of the material? What if form A is composed of very difficult items, while form B is relatively easy? Equating analyses are performed to address this very issue, so that scores are as fair as possible.
Equating in item response theory
In item response theory, person "locations" (measures of some quality being assessed by a test) are estimated on an interval scale; i.e., locations are estimated in relation to a unit and origin. It is common in educational assessment to employ tests in order to assess different groups of students with the intention of establishing a common scale by equating the origins, and when appropri
Document 4:::
Adaptive comparative judgement is a technique borrowed from psychophysics which is able to generate reliable results for educational assessment – as such it is an alternative to traditional exam script marking. In the approach, judges are presented with pairs of student work and are then asked to choose which is better, one or the other. By means of an iterative and adaptive algorithm, a scaled distribution of student work can then be obtained without reference to criteria.
Introduction
Traditional exam script marking began in Cambridge 1792 when, with undergraduate numbers rising, the importance of proper ranking of students was growing. So in 1792 the new Proctor of Examinations, William Farish, introduced marking, a process in which every examiner gives a numerical score to each response by every student, and the overall total mark puts the students in the final rank order. Francis Galton (1869) noted that, in an unidentified year about 1863, the Senior Wrangler scored 7,634 out of a maximum of 17,000, while the Second Wrangler scored 4,123. (The 'Wooden Spoon' scored only 237.)
Prior to 1792, a team of Cambridge examiners convened at 5pm on the last day of examining, reviewed the 19 papers each student had sat – and published their rank order at midnight. Marking solved the problems of numbers and prevented unfair personal bias, and its introduction was a step towards modern objective testing, the format it is best suited to. But the technology of testing that followed, with its major emphasis on reliability and the automatisation of marking, has been an uncomfortable partner for some areas of educational achievement: assessing writing or speaking, and other kinds of performance need something more qualitative and judgemental.
The technique of Adaptive Comparative Judgement is an alternative to marking. It returns to the pre-1792 idea of sorting papers according to their quality, but retains the guarantee of reliability and fairness. It is by far the most rel
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
A girl walked for 30 minutes. She noticed that she traveled farther in the first 15 minutes of her walk than in the second 15 minutes. What can she conclude about her walk?
A. She walked over many hills.
B. Her average speed was faster during the first half of her walk.
C. She walked in two different directions.
D. She was walking at a constant speed.
Answer:
|
|
sciq-2133
|
multiple_choice
|
The fight or flight response and similar responses are controlled by what part of the nervous system?
|
[
"cells",
"autonomous",
"central nervous system",
"spinal cord"
] |
B
|
Relavent Documents:
Document 0:::
Neural top–down control of physiology concerns the direct regulation by the brain of physiological functions (in addition to smooth muscle and glandular ones). Cellular functions include the immune system’s production of T-lymphocytes and antibodies, and nonimmune related homeostatic functions such as liver gluconeogenesis, sodium reabsorption, osmoregulation, and brown adipose tissue nonshivering thermogenesis. This regulation occurs through the sympathetic and parasympathetic system (the autonomic nervous system), and their direct innervation of body organs and tissues that starts in the brainstem. There is also a noninnervation hormonal control through the hypothalamus and pituitary (HPA). These lower brain areas are under control of cerebral cortex ones. Such cortical regulation differs between its left and right sides. Pavlovian conditioning shows that brain control over basic cell level physiological function can be learned.
Higher brain
Cerebral cortex
Sympathetic and parasympathetic nervous systems and the hypothalamus are regulated by the higher brain. Through them, the higher cerebral cortex areas can control the immune system, and the body’s homeostatic and stress physiology. Areas doing this include the insular cortex, the orbital, and the medial prefrontal cortices. These cerebral areas also control smooth muscle and glandular physiological processes through the sympathetic and parasympathetic nervous system including blood circulation, urogenital, gastrointestinal functions, pancreatic gut secretions, respiration, coughing, vomiting, piloerection, pupil dilation, lacrimation and salivation.
Lateralization
The sympathetic nervous system is predominantly controlled by the right side of the brain (focused upon the insular cortex), while the left side predominantly controls the parasympathetic nervous system. The cerebral cortex in rodents shows lateral specialization in its regulation of immunity with immunosuppression being controlled by the righ
Document 1:::
The sensory nervous system is a part of the nervous system responsible for processing sensory information. A sensory system consists of sensory neurons (including the sensory receptor cells), neural pathways, and parts of the brain involved in sensory perception and interoception. Commonly recognized sensory systems are those for vision, hearing, touch, taste, smell, balance and visceral sensation. Sense organs are transducers that convert data from the outer physical world to the realm of the mind where people interpret the information, creating their perception of the world around them.
The receptive field is the area of the body or environment to which a receptor organ and receptor cells respond. For instance, the part of the world an eye can see, is its receptive field; the light that each rod or cone can see, is its receptive field. Receptive fields have been identified for the visual system, auditory system and somatosensory system.
Stimulus
Organisms need information to solve at least three kinds of problems: (a) to maintain an appropriate environment, i.e., homeostasis; (b) to time activities (e.g., seasonal changes in behavior) or synchronize activities with those of conspecifics; and (c) to locate and respond to resources or threats (e.g., by moving towards resources or evading or attacking threats). Organisms also need to transmit information in order to influence another's behavior: to identify themselves, warn conspecifics of danger, coordinate activities, or deceive.
Sensory systems code for four aspects of a stimulus; type (modality), intensity, location, and duration. Arrival time of a sound pulse and phase differences of continuous sound are used for sound localization. Certain receptors are sensitive to certain types of stimuli (for example, different mechanoreceptors respond best to different kinds of touch stimuli, like sharp or blunt objects). Receptors send impulses in certain patterns to send information about the intensity of a stimul
Document 2:::
The fight-or-flight or the fight-flight-or-freeze (also called hyperarousal or the acute stress response) is a physiological reaction that occurs in response to a perceived harmful event, attack, or threat to survival. It was first described by Walter Bradford Cannon. His theory states that animals react to threats with a general discharge of the sympathetic nervous system, preparing the animal for fighting or fleeing. More specifically, the adrenal medulla produces a hormonal cascade that results in the secretion of catecholamines, especially norepinephrine and epinephrine. The hormones estrogen, testosterone, and cortisol, as well as the neurotransmitters dopamine and serotonin, also affect how organisms react to stress. The hormone osteocalcin might also play a part.
This response is recognised as the first stage of the general adaptation syndrome that regulates stress responses among vertebrates and other organisms.
Name
Originally understood as the "fight-or-flight" response in Cannon's research, the state of hyperarousal results in several responses beyond fighting or fleeing. This has led people to calling it the "fight, flight, freeze" response, "fight-flight-freeze-fawn"
or "fight-flight-faint-or-freeze", among other variants. The wider array of responses, such as freezing, fainting, fleeing, or experiencing fright, has led researchers to use more neutral or accommodating terminology such as "hyperarousal" or the "acute stress response".
Physiology
Autonomic nervous system
The autonomic nervous system is a control system that acts largely unconsciously and regulates heart rate, digestion, respiratory rate, pupillary response, urination, and sexual arousal. This system is the primary mechanism in control of the fight-or-flight response and its role is mediated by two different components: the sympathetic nervous system and the parasympathetic nervous system.
Sympathetic nervous system
The sympathetic nervous system originates in the spinal cord and i
Document 3:::
The following diagram is provided as an overview of and topical guide to the human nervous system:
Human nervous system – the part of the human body that coordinates a person's voluntary and involuntary actions and transmits signals between different parts of the body. The human nervous system consists of two main parts: the central nervous system (CNS) and the peripheral nervous system (PNS). The CNS contains the brain and spinal cord. The PNS consists mainly of nerves, which are long fibers that connect the CNS to every other part of the body. The PNS includes motor neurons, mediating voluntary movement; the autonomic nervous system, comprising the sympathetic nervous system and the parasympathetic nervous system and regulating involuntary functions; and the enteric nervous system, a semi-independent part of the nervous system whose function is to control the gastrointestinal system.
Evolution of the human nervous system
Evolution of nervous systems
Evolution of human intelligence
Evolution of the human brain
Paleoneurology
Some branches of science that study the human nervous system
Neuroscience
Neurology
Paleoneurology
Central nervous system
The central nervous system (CNS) is the largest part of the nervous system and includes the brain and spinal cord.
Spinal cord
Brain
Brain – center of the nervous system.
Outline of the human brain
List of regions of the human brain
Principal regions of the vertebrate brain:
Peripheral nervous system
Peripheral nervous system (PNS) – nervous system structures that do not lie within the CNS.
Sensory system
A sensory system is a part of the nervous system responsible for processing sensory information. A sensory system consists of sensory receptors, neural pathways, and parts of the brain involved in sensory perception.
List of sensory systems
Sensory neuron
Perception
Visual system
Auditory system
Somatosensory system
Vestibular system
Olfactory system
Taste
Pain
Components of the nervous system
Neuron
I
Document 4:::
Physiological psychology is a subdivision of behavioral neuroscience (biological psychology) that studies the neural mechanisms of perception and behavior through direct manipulation of the brains of nonhuman animal subjects in controlled experiments. This field of psychology takes an empirical and practical approach when studying the brain and human behavior. Most scientists in this field believe that the mind is a phenomenon that stems from the nervous system. By studying and gaining knowledge about the mechanisms of the nervous system, physiological psychologists can uncover many truths about human behavior. Unlike other subdivisions within biological psychology, the main focus of psychological research is the development of theories that describe brain-behavior relationships.
Physiological psychology studies many topics relating to the body's response to a behavior or activity in an organism. It concerns the brain cells, structures, components, and chemical interactions that are involved in order to produce actions. Psychologists in this field usually focus their attention to topics such as sleep, emotion, ingestion, senses, reproductive behavior, learning/memory, communication, psychopharmacology, and neurological disorders. The basis for these studies all surround themselves around the notion of how the nervous system intertwines with other systems in the body to create a specific behavior.
Nervous system
The nervous system can be described as a control system that interconnects the other body systems. It consists of the brain, spinal cord, and other nerve tissues throughout the body. The system's primary function is to react to internal and external stimuli in the human body. It uses electrical and chemical signals to send out responses to different parts of the body, and it is made up of nerve cells called neurons. Through the system, messages are transmitted to body tissues such as a muscle. There are two major subdivisions in the nervous system known a
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
The fight or flight response and similar responses are controlled by what part of the nervous system?
A. cells
B. autonomous
C. central nervous system
D. spinal cord
Answer:
|
|
sciq-9246
|
multiple_choice
|
The saguaro doesn't have any leaves to lose water by which process?
|
[
"propagation",
"evaporation",
"respiration",
"transpiration"
] |
D
|
Relavent Documents:
Document 0:::
The soil-plant-atmosphere continuum (SPAC) is the pathway for water moving from soil through plants to the atmosphere. Continuum in the description highlights the continuous nature of water connection through the pathway. The low water potential of the atmosphere, and relatively higher (i.e. less negative) water potential inside leaves, leads to a diffusion gradient across the stomatal pores of leaves, drawing water out of the leaves as vapour. As water vapour transpires out of the leaf, further water molecules evaporate off the surface of mesophyll cells to replace the lost molecules since water in the air inside leaves is maintained at saturation vapour pressure. Water lost at the surface of cells is replaced by water from the xylem, which due to the cohesion-tension properties of water in the xylem of plants pulls additional water molecules through the xylem from the roots toward the leaf.
Components
The transport of water along this pathway occurs in components, variously defined among scientific disciplines:
Soil physics characterizes water in soil in terms of tension,
Physiology of plants and animals characterizes water in organisms in terms of diffusion pressure deficit, and
Meteorology uses vapour pressure or relative humidity to characterize atmospheric water.
SPAC integrates these components and is defined as a:
...concept recognising that the field with all its components (soil, plant, animals and the ambient atmosphere taken together) constitutes a physically integrated, dynamic system in which the various flow processes involving energy and matter occur simultaneously and independently like links in the chain.
This characterises the state of water in different components of the SPAC as expressions of the energy level or water potential of each. Modelling of water transport between components relies on SPAC, as do studies of water potential gradients between segments.
See also
Ecohydrology
Evapotranspiration
Hydraulic redistribution; a p
Document 1:::
Wilting is the loss of rigidity of non-woody parts of plants. This occurs when the turgor pressure in non-lignified plant cells falls towards zero, as a result of diminished water in the cells. Wilting also serves to reduce water loss, as it makes the leaves expose less surface area. The rate of loss of water from the plant is greater than the absorption of water in the plant. The process of wilting
modifies the leaf angle distribution of the plant (or canopy) towards more erectophile conditions.
Lower water availability may result from:
drought conditions, where the soil moisture drops below conditions most favorable for plant functioning;
the temperature falls to the point where the plant's vascular system cannot function;
high salinity, which causes water to diffuse from the plant cells and induce shrinkage;
saturated soil conditions, where roots are unable to obtain sufficient oxygen for cellular respiration, and so are unable to transport water into the plant; or
bacteria or fungi that clog the plant's vascular system.
Wilting diminishes the plant's ability to transpire and grow. Permanent wilting leads to plant death. Symptoms of wilting and blights resemble one another.
The plants may recover during the night when evaporation is reduced as the stomata closes.
In woody plants, reduced water availability leads to cavitation of the xylem.
Wilting occurs in plants such as balsam and holy basil. Wilting is an effect of the plant growth-inhibiting hormone, abscisic acid.
With cucurbits, wilting can be caused by the squash vine borer.
Document 2:::
Tissue hydration is the process of absorbing and retaining water in biological tissues.
Plants
Land plants maintain adequate tissue hydration by means of an outer waterproof layer. In soft or green tissues, this is usually a waxy cuticle over the outer epidermis. In older, woody tissues, waterproofing chemicals are present in the secondary cell wall that limit or inhibit the flow of water. Vascular plants also possess an internal vascular system that distributes fluids throughout the plant.
Some xerophytes, such as cacti and other desert plants, have mucilage in their tissues. This is a sticky substance that holds water within the plant, reducing the rate of dehydration. Some seeds and spores remain dormant until adequate moisture is present, at which time the seed or spore begins to germinate.
Animals
Animals maintain adequate tissue hydration by means of (1) an outer skin, shell, or cuticle; (2) a fluid-filled coelom cavity; and (3) a circulatory system.
Hydration of fat free tissues, ratio of total body water to fat free body mass, is stable at 0.73 in mammals.
In humans, a significant drop in tissue hydration can lead to the medical condition of dehydration. This may result from loss of water itself, loss of electrolytes, or a loss of blood plasma. Administration of hydrational fluids as part of sound dehydration management is necessary to avoid severe complications, and in some cases, death.
Some invertebrates are able to survive extreme desiccation of their tissues by entering a state of cryptobiosis.
See also
Osmoregulation
Document 3:::
A xerophyte (from Greek ξηρός xeros 'dry' + φυτόν phuton 'plant') is a species of plant that has adaptations to survive in an environment with little liquid water. Examples are typically desert regions like the Sahara, and places in the Alps or the Arctic. Popular examples of xerophytes are cacti, pineapple and some Gymnosperm plants.
The structural features (morphology) and fundamental chemical processes (physiology) of xerophytes are variously adapted to conserve water, also common to store large quantities of water, during dry periods. Other species are able to survive long periods of extreme dryness or desiccation of their tissues, during which their metabolic activity may effectively shut down. Plants with such morphological and physiological adaptations are . Xerophytes such as cacti are capable of withstanding extended periods of dry conditions as they have deep-spreading roots and capacity to store water. Their waxy, thorny leaves prevent loss of moisture. Even their fleshy stems can store water.
Introduction
Plants absorb water from the soil, which then evaporates from their shoots and leaves; this process is known as transpiration. If placed in a dry environment, a typical mesophytic plant would evaporate water faster than the rate of water uptake from the soil, leading to wilting and even death.
Xerophytic plants exhibit a diversity of specialized adaptations to survive in such water-limiting conditions. They may use water from their own storage, allocate water specifically to sites of new tissue growth, or lose less water to the atmosphere and so channel a greater proportion of water from the soil to photosynthesis and growth. Different plant species possess different qualities and mechanisms to manage water supply, enabling them to survive.
Cacti and other succulents are commonly found in deserts, where there is little rainfall. Other xerophytes, such as certain bromeliads, can survive through both extremely wet and extremely dry periods and can
Document 4:::
Maintenance respiration (or maintenance energy) refers to metabolism occurring in an organism that is needed to maintain that organism in a healthy, living state. Maintenance respiration contrasts with growth respiration, which is responsible for the synthesis of new structures in growth, nutrient uptake, nitrogen (N) reduction and phloem loading, whereas maintenance respiration is associated with protein and membrane turnover and maintenance of ion concentrations and gradients.
In plants
Maintenance respiration in plants refers to the amount of cellular respiration, measured by the carbon dioxide (CO2) released or oxygen (O2) consumed, during the generation of usable energy (mainly ATP, NADPH, and NADH) and metabolic intermediates used for (i) resynthesis of compounds that undergo renewal (turnover) in the normal process of metabolism (examples are enzymatic proteins, ribonucleic acids, and membrane lipids); (ii) maintenance of chemical gradients of ions and metabolites across cellular membranes that are necessary for cellular integrity and plant health; and (iii) operation of metabolic processes involved in physiological adjustment (i.e., acclimation) to a change in the plant's environment. The metabolic costs of the repair of injury from biotic or abiotic stress may also be considered a part of maintenance respiration.
Maintenance respiration is essential for biological health and growth of plants. It is estimated that about half of the respiration carried out by terrestrial plants during their lifetime is for the support of maintenance processes. Because typically more than half of global terrestrial plant photosynthesis (or gross primary production) is used for plant respiration, more than one quarter of global terrestrial plant photosynthesis is presumably consumed in maintenance respiration.
Maintenance respiration is a key component of most physiologically based mathematical models of plant growth, including models of crop growth and yield and models of
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
The saguaro doesn't have any leaves to lose water by which process?
A. propagation
B. evaporation
C. respiration
D. transpiration
Answer:
|
|
sciq-7399
|
multiple_choice
|
Practical limits of van de graaff generators arise because the large electric fields polarize and eventually do what to surrounding materials?
|
[
"decompose",
"isolate",
"ionize",
"displace"
] |
C
|
Relavent Documents:
Document 0:::
A Van de Graaff generator is an electrostatic generator which uses a moving belt to accumulate electric charge on a hollow metal globe on the top of an insulated column, creating very high electric potentials. It produces very high voltage direct current (DC) electricity at low current levels. It was invented by American physicist Robert J. Van de Graaff in 1929.
The potential difference achieved by modern Van de Graaff generators can be as much as 5 megavolts. A tabletop version can produce on the order of 100 kV and can store enough energy to produce visible electric sparks. Small Van de Graaff machines are produced for entertainment, and for physics education to teach electrostatics; larger ones are displayed in some science museums.
The Van de Graaff generator was originally developed as a particle accelerator for physics research, as its high potential can be used to accelerate subatomic particles to great speeds in an evacuated tube. It was the most powerful type of accelerator until the cyclotron was developed in the early 1930s. Van de Graaff generators are still used as accelerators to generate energetic particle and X-ray beams for nuclear research and nuclear medicine.
The voltage produced by an open-air Van de Graaff machine is limited by arcing and corona discharge to about 5 MV. Most modern industrial machines are enclosed in a pressurized tank of insulating gas; these can achieve potentials as large as about 25 MV.
History
Background
The concept of an electrostatic generator in which charge is mechanically transported in small amounts into the interior of a high-voltage electrode originated with the Kelvin water dropper, invented in 1867 by William Thomson (Lord Kelvin), in which charged drops of water fall into a bucket with the same polarity charge, adding to the charge.
In a machine of this type, the gravitational force moves the drops against the opposing electrostatic field of the bucket. Kelvin himself first suggested using a belt to carry
Document 1:::
An electrostatic generator, or electrostatic machine, is an electrical generator that produces static electricity, or electricity at high voltage and low continuous current. The knowledge of static electricity dates back to the earliest civilizations, but for millennia it remained merely an interesting and mystifying phenomenon, without a theory to explain its behavior and often confused with magnetism. By the end of the 17th century, researchers had developed practical means of generating electricity by friction, but the development of electrostatic machines did not begin in earnest until the 18th century, when they became fundamental instruments in the studies about the new science of electricity.
Electrostatic generators operate by using manual (or other) power to transform mechanical work into electric energy, or using electric currents. Manual electrostatic generators develop electrostatic charges of opposite signs rendered to two conductors, using only electric forces, and work by using moving plates, drums, or belts to carry electric charge to a high potential electrode.
Description
Electrostatic machines are typically used in science classrooms to safely demonstrate electrical forces and high voltage phenomena. The elevated potential differences achieved have been also used for a variety of practical applications, such as operating X-ray tubes, particle accelerators, spectroscopy, medical applications, sterilization of food, and nuclear physics experiments. Electrostatic generators such as the Van de Graaff generator, and variations as the Pelletron, also find use in physics research.
Electrostatic generators can be divided into categories depending on how the charge is generated:
Friction machines use the triboelectric effect (electricity generated by contact or friction)
Influence machines use electrostatic induction
Others
Friction machines
History
The first electrostatic generators are called friction machines because of the friction in the genera
Document 2:::
A Coulombic explosion is a condensed-matter physics process in which a molecule or crystal lattice is destroyed by the Coulombic repulsion between its constituent atoms. Coulombic explosions are a prominent technique in laser-based machining, and appear naturally in certain high-energy reactions.
Mechanism
A Coulombic explosion begins when an intense electric field (often from a laser) excites the valence electrons in a solid, ejecting them from the system and leaving behind positively charged ions. The chemical bonds holding the solid together are weakened by the loss of the electrons, enabling the Coulombic repulsion between the ions to overcome them. The result is an explosion of ions and electrons – a plasma.
The laser must be very intense to produce a Coulomb explosion. If it is too weak, the energy given to the electrons will be transferred to the ions via electron-phonon coupling. This will cause the entire material to heat up, melt, and thermally ablate away as a plasma. The end result is similar to Coulomb explosion, except that any fine structure in the material will be damaged by thermal melting.
It may be shown that the Coulomb explosion occurs in the same parameter regime as the superradiant phase transition i.e. when the destabilizing interactions become overwhelming and dominate over the oscillatory phonon-solid binding motions.
Technological use
A Coulomb explosion is a "cold" alternative to the dominant laser etching technique of thermal ablation, which depends on local heating, melting, and vaporization of molecules and atoms using less-intense beams. Pulse brevity down only to the nanosecond regime is sufficient to localize thermal ablation – before the heat is conducted far, the energy input (pulse) has ended. Nevertheless, thermally ablated materials may seal pores important in catalysis or battery operation, and recrystallize or even burn the substrate, thus changing the physical and chemical properties at the etch site. In contrast, even
Document 3:::
Wake Shield Facility (WSF) was a NASA experimental science platform that was placed in low Earth orbit by the Space Shuttle. It was a diameter, free-flying stainless steel disk.
The WSF was deployed using the Space Shuttle's Canadarm. The WSF then used nitrogen gas thrusters to position itself about behind the Space Shuttle, which was at an orbital altitude of over , within the thermosphere, where the atmosphere is exceedingly tenuous. The WSF's orbital speed was at least three to four times faster than the speed of thermospheric gas molecules in the area, which resulted in a cone behind the WSF that was entirely free of gas molecules. The WSF thus created an ultrahigh vacuum in its wake. The resulting vacuum was used to study epitaxial film growth. The WSF operated at a distance from the Space Shuttle to avoid contamination from the Shuttle's rocket thrusters and water dumped overboard from the Shuttle's Waste Collection System (space toilet). After two days, the Space Shuttle would rendezvous with the WSF and again use its robotic arm to collect the WSF and to store it in the Shuttle's payload bay for return to Earth.
The WSF was flown into space three times, aboard Shuttle flights STS-60 (WSF-1), STS-69 (WSF-2) and STS-80 (WSF-3). During STS-60, some hardware issues were experienced, and, as a result, the WSF-1 was only deployed at the end of the Shuttle's Canadarm. During the later missions, the WSF was deployed as a free-flying platform in the wake of the Shuttle.
These flights proved the vacuum wake concept and realized the space epitaxy concept by growing the first-ever crystalline semiconductor thin films in the vacuum of space. These included gallium arsenide (GaAs) and aluminum gallium arsenide (AlGaAs) depositions. These experiments have been used to develop better photocells and thin films. Among the potential resulting applications are artificial retinas made from tiny ceramic detectors.
Pre-flight calculations suggested that the pressure on the w
Document 4:::
The Röntgen Memorial Site in Würzburg, Germany, is dedicated to the work of the German physicist Wilhelm Conrad Röntgen (1845–1923) and his discovery of X-rays, for which he was granted the Nobel Prize in physics. It contains an exhibition of historical instruments, machines and documents.
Location
The Röntgen Memorial Site is in the foyer, corridors and two laboratory rooms of the former Physics Institute of the University of Würzburg in Röntgenring 8, a building that is now used by the University of Applied Sciences Würzburg-Schweinfurt. The road, where the building lies, was renamed in 1909 from Pleicherring to Röntgenring.
History
On the late Friday evening of 8. November 1895 Röntgen discovered for the first time the rays which penetrate through solid materials and gave them the name X-rays. He presented this in a lecture and publication On a new type of rays - Über eine neue Art von Strahlen on 23 January 1896 at the Physical Medical Society of Würzburg.
During the discussion of this lecture, the anatomist Albert von Kölliker proposed to call these rays Röntgen radiation after their inventor, a term that is still being used in Germany.
Exhibition
The Röntgen Memorial Site gives an insight into the particle physics of the late 19th century. It shows an experimental set-up of cathodic rays beside the apparatus of the discovery. An experiment of penetrating solid materials by X-rays is shown in the historic laboratory of Röntgen. A separate room shows various X-ray tubes, a medical X-ray machine of Siemens & Halske from 1912 and several original documents. In the foyer a short German movie explains the purpose of the Memorial Site and the life of Röntgen. In the corridor some personal belongings of Röntgen are displayed to give some background information on his personal and historical circumstances.
After remodeling in 2015 the tables and captures of the exhibition are now in English and German language.
Society
The site is managed by the non-profit
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Practical limits of van de graaff generators arise because the large electric fields polarize and eventually do what to surrounding materials?
A. decompose
B. isolate
C. ionize
D. displace
Answer:
|
|
sciq-4441
|
multiple_choice
|
To the extent that behaviors are controlled by what, they may evolve through natural selection?
|
[
"genes",
"mutation",
"rna",
"dna"
] |
A
|
Relavent Documents:
Document 0:::
During the latter half of the 20th century, the fields of genetics and molecular biology matured greatly, significantly increasing understanding of biological heredity. As with other complex and evolving fields of knowledge, the public awareness of these advances has primarily been through the mass media, and a number of common misunderstandings of genetics have arisen.
Genetic determinism
It is a popular misconception that all patterns of an animal's behaviour, and more generally its phenotype, are rigidly determined by its genes. Although many examples of animals exist that display certain well-defined behaviour that is genetically programmed, these examples cannot be extrapolated to all animal behaviour. There is good evidence that some basic aspects of human behaviour, such as circadian rhythms are genetically based, but it is clear that many other aspects are not.
In the first place, much phenotypic variability does not stem from genes themselves. For example:
Epigenetic inheritance. In the widest definition this includes all biological inheritance mechanisms that do not change the DNA sequence of the genome. In a narrower definition it excludes biological phenomena such as the effects of prions and maternal antibodies which are also inherited and have clear survival implications.
Learning from experience. This feature is obviously important for humans, but there is considerable evidence of learned behaviour in other animal species (vertebrates and invertebrates). There are even reports of learned behaviour in Drosophila larvae.
A gene for X
In the early years of genetics it was suggested that there might be "a gene for" a wide range of particular characteristics. This was partly because the examples studied from Mendel onwards inevitably focused on genes whose effects could be readily identified; partly that it was easier to teach science that way; and partly because the mathematics of evolutionary dynamics is simpler if there is a simple mapping between
Document 1:::
Developmental systems theory (DST) is an overarching theoretical perspective on biological development, heredity, and evolution. It emphasizes the shared contributions of genes, environment, and epigenetic factors on developmental processes. DST, unlike conventional scientific theories, is not directly used to help make predictions for testing experimental results; instead, it is seen as a collection of philosophical, psychological, and scientific models of development and evolution. As a whole, these models argue the inadequacy of the modern evolutionary synthesis on the roles of genes and natural selection as the principal explanation of living structures. Developmental systems theory embraces a large range of positions that expand biological explanations of organismal development and hold modern evolutionary theory as a misconception of the nature of living processes.
Overview
All versions of developmental systems theory espouse the view that:
All biological processes (including both evolution and development) operate by continually assembling new structures.
Each such structure transcends the structures from which it arose and has its own systematic characteristics, information, functions and laws.
Conversely, each such structure is ultimately irreducible to any lower (or higher) level of structure, and can be described and explained only on its own terms.
Furthermore, the major processes through which life as a whole operates, including evolution, heredity and the development of particular organisms, can only be accounted for by incorporating many more layers of structure and process than the conventional concepts of ‘gene’ and ‘environment’ normally allow for.
In other words, although it does not claim that all structures are equal, development systems theory is fundamentally opposed to reductionism of all kinds. In short, developmental systems theory intends to formulate a perspective which does not presume the causal (or ontological) priority of any p
Document 2:::
The term evolvability is used for a recent framework of computational learning introduced by Leslie Valiant in his paper of the same name and described below. The aim of this theory is to model biological evolution and categorize which types of mechanisms are evolvable. Evolution is an extension of PAC learning and learning from statistical queries.
General framework
Let and be collections of functions on variables. Given an ideal function , the goal is to find by local search a representation that closely approximates . This closeness is measured by the performance of with respect to .
As is the case in the biological world, there is a difference between genotype and phenotype. In general, there can be multiple representations (genotypes) that correspond to the same function (phenotype). That is, for some , with , still for all . However, this need not be the case. The goal then, is to find a representation that closely matches the phenotype of the ideal function, and the spirit of the local search is to allow only small changes in the genotype. Let the neighborhood of a representation be the set of possible mutations of .
For simplicity, consider Boolean functions on , and let be a probability distribution on . Define the performance in terms of this. Specifically,
Note that In general, for non-Boolean functions, the performance will not correspond directly to the probability that the functions agree, although it will have some relationship.
Throughout an organism's life, it will only experience a limited number of environments, so its performance cannot be determined exactly. The empirical performance is defined by
where is a multiset of independent selections from according to . If is large enough, evidently will be close to the actual performance .
Given an ideal function , initial representation , sample size , and tolerance , the mutator is a random variable defined as follows. Each is classified as beneficial, neutral, or deleteriou
Document 3:::
Tinbergen's four questions, named after 20th century biologist Nikolaas Tinbergen, are complementary categories of explanations for animal behaviour. These are also commonly referred to as levels of analysis. It suggests that an integrative understanding of behaviour must include ultimate (evolutionary) explanations, in particular:
behavioural adaptive functions
phylogenetic history; and the proximate explanations
underlying physiological mechanisms
ontogenetic/developmental history.
Four categories of questions and explanations
When asked about the purpose of sight in humans and animals, even elementary-school children can answer that animals have vision to help them find food and avoid danger (function/adaptation). Biologists have three additional explanations: sight is caused by a particular series of evolutionary steps (phylogeny), the mechanics of the eye (mechanism/causation), and even the process of an individual's development (ontogeny).
This schema constitutes a basic framework of the overlapping behavioural fields of ethology, behavioural ecology, comparative psychology, sociobiology, evolutionary psychology, and anthropology. Julian Huxley identified the first three questions. Niko Tinbergen gave only the fourth question, as Huxley's questions failed to distinguish between survival value and evolutionary history; Tinbergen's fourth question helped resolve this problem.
Evolutionary (ultimate) explanations
First question: Function (adaptation)
Darwin's theory of evolution by natural selection is the only scientific explanation for why an animal's behaviour is usually well adapted for survival and reproduction in its environment. However, claiming that a particular mechanism is well suited to the present environment is different from claiming that this mechanism was selected for in the past due to its history of being adaptive.
The literature conceptualizes the relationship between function and evolution in two ways. On the one hand, function
Document 4:::
Evolutionary biology is the subfield of biology that studies the evolutionary processes (natural selection, common descent, speciation) that produced the diversity of life on Earth. It is also defined as the study of the history of life forms on Earth. Evolution holds that all species are related and gradually change over generations. In a population, the genetic variations affect the phenotypes (physical characteristics) of an organism. These changes in the phenotypes will be an advantage to some organisms, which will then be passed onto their offspring. Some examples of evolution in species over many generations are the peppered moth and flightless birds. In the 1930s, the discipline of evolutionary biology emerged through what Julian Huxley called the modern synthesis of understanding, from previously unrelated fields of biological research, such as genetics and ecology, systematics, and paleontology.
The investigational range of current research has widened to encompass the genetic architecture of adaptation, molecular evolution, and the different forces that contribute to evolution, such as sexual selection, genetic drift, and biogeography. Moreover, the newer field of evolutionary developmental biology ("evo-devo") investigates how embryogenesis is controlled, thus yielding a wider synthesis that integrates developmental biology with the fields of study covered by the earlier evolutionary synthesis.
Subfields
Evolution is the central unifying concept in biology. Biology can be divided into various ways. One way is by the level of biological organization, from molecular to cell, organism to population. Another way is by perceived taxonomic group, with fields such as zoology, botany, and microbiology, reflecting what was once seen as the major divisions of life. A third way is by approaches, such as field biology, theoretical biology, experimental evolution, and paleontology. These alternative ways of dividing up the subject have been combined with evolution
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
To the extent that behaviors are controlled by what, they may evolve through natural selection?
A. genes
B. mutation
C. rna
D. dna
Answer:
|
|
sciq-3156
|
multiple_choice
|
What are liverworts with a flattened, ribbon-like body called?
|
[
"hepatic liverworts",
"kidney liverworts",
"thallose liverworts",
"hornworts"
] |
C
|
Relavent Documents:
Document 0:::
Succubous is a manner in which the leaves of a liverwort overlap. If one were to look down from above (dorsal side) on a plant where the leaf attachment is succubous, the upper edge of each leaf would be covered by the next leaf along the stem. The lower edge of each leaf is visible from above, but the edge of the leaf closer to the tip of the stem is obscured by a neighboring leaf. The opposite of succubous is incubous.
Document 1:::
An elater is a cell (or structure attached to a cell) that is hygroscopic, and therefore will change shape in response to changes in moisture in the environment. Elaters come in a variety of forms, but are always associated with plant spores. In many plants that do not have seeds, they function in dispersing the spores to a new location. Mosses do not have elaters, but peristomes which change shape with changes in humidity or moisture to allow for a gradual release of spores.
Horsetail elaters
In the horsetails, elaters are four ribbon-like appendages attached to the spores. These appendages develop from an outer spiral layer of the spore wall. At maturity, the four strips peel away from the inner wall, except at a single point on the spore where all four strips are attached.
Under moist conditions, the elaters curl tightly around the spore. The wet spores tend to stick to each other and to nearby surfaces because of surface tension. When conditions are dry, the spores no longer stick to each other and are more easily dispersed. At that time, the elaters uncoil to extend out from the spore and will catch air currents. The fact that they are extended only when conditions are dry means that successful spore dispersal is more likely.
Liverwort elaters
In the liverworts also known as hepaticopsida [example Riccia,Marchantia], elaters are cells that develop in the sporophyte alongside the spores. They are complete cells, usually with helical thickenings at maturity that respond to moisture content.
In most liverworts, the elaters are unattached, but in some leafy species (such as Frullania) a few elaters will remain attached to the inside of the sporangium (spore capsule).
Hornwort pseudo-elaters
In the hornworts, elaters are branched clusters of cells that develop in the sporophyte alongside the spores. They are complete cells, usually without helical thickenings (except in the Dendrocerotaceae).
Document 2:::
Offshoots are lateral shoots that are produced on the main stem of a plant. They may be known colloquially as "suckers", “pups” or “sister plants”
See also
Stolon or runners
Plant anatomy
Plant morphology
Document 3:::
The term incubous describes the way in which the leaves of a liverwort are attached to the stem. If one were to look down from above (dorsal side) on a plant where the leaf attachment is incubous, the upper edge of each leaf would overlap the next higher leaf along the stem. Because of this, the upper edge of each leaf is visible from above, but the lower edge of each leaf is obscured by its neighboring leaf. The opposite of incubous is succubous.
Document 4:::
Arachnoid organs, such as leaves or stems, have an external appearance similar to cobwebs – the appearance of being covered with fine, white, usually tangled hairs. This can cause plants to appear grey or white. The arachnoid appearance is common on the leaves and stems of various sclerophyllous members of the family Asteraceae, such as some thistles. The arachnoid appearance of Haworthia arachnoidea arises from the spinescent leaf denticles, and the arachnoid appearance of the cactus Cephalocereus senilis is from long-lasting hairy spines.
See also
Arachnoid (disambiguation)
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What are liverworts with a flattened, ribbon-like body called?
A. hepatic liverworts
B. kidney liverworts
C. thallose liverworts
D. hornworts
Answer:
|
|
sciq-11291
|
multiple_choice
|
What law shows the relationships among temperature, volume, and pressure?
|
[
"Murphy's Law",
"Law of Conservation",
"combined gas",
"Newton's law"
] |
C
|
Relavent Documents:
Document 0:::
Scientific laws or laws of science are statements, based on repeated experiments or observations, that describe or predict a range of natural phenomena. The term law has diverse usage in many cases (approximate, accurate, broad, or narrow) across all fields of natural science (physics, chemistry, astronomy, geoscience, biology). Laws are developed from data and can be further developed through mathematics; in all cases they are directly or indirectly based on empirical evidence. It is generally understood that they implicitly reflect, though they do not explicitly assert, causal relationships fundamental to reality, and are discovered rather than invented.
Scientific laws summarize the results of experiments or observations, usually within a certain range of application. In general, the accuracy of a law does not change when a new theory of the relevant phenomenon is worked out, but rather the scope of the law's application, since the mathematics or statement representing the law does not change. As with other kinds of scientific knowledge, scientific laws do not express absolute certainty, as mathematical theorems or identities do. A scientific law may be contradicted, restricted, or extended by future observations.
A law can often be formulated as one or several statements or equations, so that it can predict the outcome of an experiment. Laws differ from hypotheses and postulates, which are proposed during the scientific process before and during validation by experiment and observation. Hypotheses and postulates are not laws, since they have not been verified to the same degree, although they may lead to the formulation of laws. Laws are narrower in scope than scientific theories, which may entail one or several laws. Science distinguishes a law or theory from facts. Calling a law a fact is ambiguous, an overstatement, or an equivocation. The nature of scientific laws has been much discussed in philosophy, but in essence scientific laws are simply empirical
Document 1:::
The laws describing the behaviour of gases under fixed pressure, volume and absolute temperature conditions are called Gas Laws. The basic gas laws were discovered by the end of the 18th century when scientists found out that relationships between pressure, volume and temperature of a sample of gas could be obtained which would hold to approximation for all gases. These macroscopic gas laws were found to be consistent with atomic and kinetic theory.
History
Following the invention of the Torricelli mercury barometer in mid 17th century, the pressure-volume gas law was soon revealed by Robert Boyle while keeping temperature constant. Marriott, however, did notice small temperature dependence. It took another century and a half to develop thermometry and recognise the absolute zero temperature scale before the discovery of temperature-dependent gas laws.
Boyle's law
In 1662, Robert Boyle systematically studied the relationship between the volume and pressure of a fixed amount of gas at a constant temperature. He observed that the volume of a given mass of a gas is inversely proportional to its pressure at a constant temperature.
Boyle's law, published in 1662, states that, at a constant temperature, the product of the pressure and volume of a given mass of an ideal gas in a closed system is always constant. It can be verified experimentally using a pressure gauge and a variable volume container. It can also be derived from the kinetic theory of gases: if a container, with a fixed number of molecules inside, is reduced in volume, more molecules will strike a given area of the sides of the container per unit time, causing a greater pressure.
Statement
Boyle's law states that:
The concept can be represented with these formulae:
, meaning "Volume is inversely proportional to Pressure", or
, meaning "Pressure is inversely proportional to Volume", or
, or
where is the pressure, is the volume of a gas, and is the constant in this equation (and is not the same as
Document 2:::
Charles's law (also known as the law of volumes) is an experimental gas law that describes how gases tend to expand when heated. A modern statement of Charles's law is:
When the pressure on a sample of a dry gas is held constant, the Kelvin temperature and the volume will be in direct proportion.
This relationship of direct proportion can be written as:
So this means:
where:
is the volume of the gas,
is the temperature of the gas (measured in kelvins), and
is a non-zero constant.
This law describes how a gas expands as the temperature increases; conversely, a decrease in temperature will lead to a decrease in volume. For comparing the same substance under two different sets of conditions, the law can be written as:
The equation shows that, as absolute temperature increases, the volume of the gas also increases in proportion.
History
The law was named after scientist Jacques Charles, who formulated the original law in his unpublished work from the 1780s.
In two of a series of four essays presented between 2 and 30 October 1801, John Dalton demonstrated by experiment that all the gases and vapours that he studied expanded by the same amount between two fixed points of temperature. The French natural philosopher Joseph Louis Gay-Lussac confirmed the discovery in a presentation to the French National Institute on 31 Jan 1802, although he credited the discovery to unpublished work from the 1780s by Jacques Charles. The basic principles had already been described by Guillaume Amontons and Francis Hauksbee a century earlier.
Dalton was the first to demonstrate that the law applied generally to all gases, and to the vapours of volatile liquids if the temperature was well above the boiling point. Gay-Lussac concurred. With measurements only at the two thermometric fixed points of water (0°C and 100°C), Gay-Lussac was unable to show that the equation relating volume to temperature was a linear function. On mathematical grounds alone, Gay-Lussac's paper do
Document 3:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 4:::
This is a list of topics that are included in high school physics curricula or textbooks.
Mathematical Background
SI Units
Scalar (physics)
Euclidean vector
Motion graphs and derivatives
Pythagorean theorem
Trigonometry
Motion and forces
Motion
Force
Linear motion
Linear motion
Displacement
Speed
Velocity
Acceleration
Center of mass
Mass
Momentum
Newton's laws of motion
Work (physics)
Free body diagram
Rotational motion
Angular momentum (Introduction)
Angular velocity
Centrifugal force
Centripetal force
Circular motion
Tangential velocity
Torque
Conservation of energy and momentum
Energy
Conservation of energy
Elastic collision
Inelastic collision
Inertia
Moment of inertia
Momentum
Kinetic energy
Potential energy
Rotational energy
Electricity and magnetism
Ampère's circuital law
Capacitor
Coulomb's law
Diode
Direct current
Electric charge
Electric current
Alternating current
Electric field
Electric potential energy
Electron
Faraday's law of induction
Ion
Inductor
Joule heating
Lenz's law
Magnetic field
Ohm's law
Resistor
Transistor
Transformer
Voltage
Heat
Entropy
First law of thermodynamics
Heat
Heat transfer
Second law of thermodynamics
Temperature
Thermal energy
Thermodynamic cycle
Volume (thermodynamics)
Work (thermodynamics)
Waves
Wave
Longitudinal wave
Transverse waves
Transverse wave
Standing Waves
Wavelength
Frequency
Light
Light ray
Speed of light
Sound
Speed of sound
Radio waves
Harmonic oscillator
Hooke's law
Reflection
Refraction
Snell's law
Refractive index
Total internal reflection
Diffraction
Interference (wave propagation)
Polarization (waves)
Vibrating string
Doppler effect
Gravity
Gravitational potential
Newton's law of universal gravitation
Newtonian constant of gravitation
See also
Outline of physics
Physics education
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What law shows the relationships among temperature, volume, and pressure?
A. Murphy's Law
B. Law of Conservation
C. combined gas
D. Newton's law
Answer:
|
|
sciq-5429
|
multiple_choice
|
The speed of a wave is a product of its wavelength and what else?
|
[
"frequency",
"density",
"voltage",
"magnitude"
] |
A
|
Relavent Documents:
Document 0:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 1:::
In the physical sciences, the wavenumber (or wave number), also known as repetency, is the spatial frequency of a wave, measured in cycles per unit distance (ordinary wavenumber) or radians per unit distance (angular wavenumber). It is analogous to temporal frequency, which is defined as the number of wave cycles per unit time (ordinary frequency) or radians per unit time (angular frequency).
In multidimensional systems, the wavenumber is the magnitude of the wave vector. The space of wave vectors is called reciprocal space. Wave numbers and wave vectors play an essential role in optics and the physics of wave scattering, such as X-ray diffraction, neutron diffraction, electron diffraction, and elementary particle physics. For quantum mechanical waves, the wavenumber multiplied by the reduced Planck's constant is the canonical momentum.
Wavenumber can be used to specify quantities other than spatial frequency. For example, in optical spectroscopy, it is often used as a unit of temporal frequency assuming a certain speed of light.
Definition
Wavenumber, as used in spectroscopy and most chemistry fields, is defined as the number of wavelengths per unit distance, typically centimeters (cm−1):
where λ is the wavelength. It is sometimes called the "spectroscopic wavenumber". It equals the spatial frequency.
For example, a wavenumber in inverse centimeters can be converted to a frequency in gigahertz by multiplying by 29.9792458 cm/ns (the speed of light, in centimeters per nanosecond); conversely, an electromagnetic wave at 29.9792458 GHz has a wavelength of 1 cm in free space.
In theoretical physics, a wave number, defined as the number of radians per unit distance, sometimes called "angular wavenumber", is more often used:
When wavenumber is represented by the symbol , a frequency is still being represented, albeit indirectly. As described in the spectroscopy section, this is done through the relationship , where s is a frequency in hertz. This is done for con
Document 2:::
This is a list of wave topics.
0–9
21 cm line
A
Abbe prism
Absorption spectroscopy
Absorption spectrum
Absorption wavemeter
Acoustic wave
Acoustic wave equation
Acoustics
Acousto-optic effect
Acousto-optic modulator
Acousto-optics
Airy disc
Airy wave theory
Alfvén wave
Alpha waves
Amphidromic point
Amplitude
Amplitude modulation
Animal echolocation
Antarctic Circumpolar Wave
Antiphase
Aquamarine Power
Arrayed waveguide grating
Artificial wave
Atmospheric diffraction
Atmospheric wave
Atmospheric waveguide
Atom laser
Atomic clock
Atomic mirror
Audience wave
Autowave
Averaged Lagrangian
B
Babinet's principle
Backward wave oscillator
Bandwidth-limited pulse
beat
Berry phase
Bessel beam
Beta wave
Black hole
Blazar
Bloch's theorem
Blueshift
Boussinesq approximation (water waves)
Bow wave
Bragg diffraction
Bragg's law
Breaking wave
Bremsstrahlung, Electromagnetic radiation
Brillouin scattering
Bullet bow shockwave
Burgers' equation
Business cycle
C
Capillary wave
Carrier wave
Cherenkov radiation
Chirp
Ernst Chladni
Circular polarization
Clapotis
Closed waveguide
Cnoidal wave
Coherence (physics)
Coherence length
Coherence time
Cold wave
Collimated light
Collimator
Compton effect
Comparison of analog and digital recording
Computation of radiowave attenuation in the atmosphere
Continuous phase modulation
Continuous wave
Convective heat transfer
Coriolis frequency
Coronal mass ejection
Cosmic microwave background radiation
Coulomb wave function
Cutoff frequency
Cutoff wavelength
Cymatics
D
Damped wave
Decollimation
Delta wave
Dielectric waveguide
Diffraction
Direction finding
Dispersion (optics)
Dispersion (water waves)
Dispersion relation
Dominant wavelength
Doppler effect
Doppler radar
Douglas Sea Scale
Draupner wave
Droplet-shaped wave
Duhamel's principle
E
E-skip
Earthquake
Echo (phenomenon)
Echo sounding
Echolocation (animal)
Echolocation (human)
Eddy (fluid dynamics)
Edge wave
Eikonal equation
Ekman layer
Ekman spiral
Ekman transport
El Niño–Southern Oscillation
El
Document 3:::
A wavenumber–frequency diagram is a plot displaying the relationship between the wavenumber (spatial frequency) and the frequency (temporal frequency) of certain phenomena. Usually frequencies are placed on the vertical axis, while wavenumbers are placed on the horizontal axis.
In the atmospheric sciences, these plots are a common way to visualize atmospheric waves.
In the geosciences, especially seismic data analysis, these plots also called f–k plot, in which energy density within a given time interval is contoured on a frequency-versus-wavenumber basis. They are used to examine the direction and apparent velocity of seismic waves and in velocity filter design.
Origins
In general, the relationship between wavelength , frequency , and the phase velocity of a sinusoidal wave is:
Using the wavenumber () and angular frequency () notation, the previous equation can be rewritten as
On the other hand, the group velocity is equal to the slope of the wavenumber–frequency diagram:
Analyzing such relationships in detail often yields information on the physical properties of the medium, such as density, composition, etc.
See also
Dispersion relation
Document 4:::
In signal processing, the energy of a continuous-time signal x(t) is defined as the area under the squared magnitude of the considered signal i.e., mathematically
Unit of will be (unit of signal)2.
And the energy of a discrete-time signal x(n) is defined mathematically as
Relationship to energy in physics
Energy in this context is not, strictly speaking, the same as the conventional notion of energy in physics and the other sciences. The two concepts are, however, closely related, and it is possible to convert from one to the other:
where Z represents the magnitude, in appropriate units of measure, of the load driven by the signal.
For example, if x(t) represents the potential (in volts) of an electrical signal propagating across a transmission line, then Z would represent the characteristic impedance (in ohms) of the transmission line. The units of measure for the signal energy would appear as volt2·seconds, which is not dimensionally correct for energy in the sense of the physical sciences. After dividing by Z, however, the dimensions of E would become volt2·seconds per ohm,
which is equivalent to joules, the SI unit for energy as defined in the physical sciences.
Spectral energy density
Similarly, the spectral energy density of signal x(t) is
where X(f) is the Fourier transform of x(t).
For example, if x(t) represents the magnitude of the electric field component (in volts per meter) of an optical signal propagating through free space, then the dimensions of X(f) would become volt·seconds per meter and would represent the signal's spectral energy density (in volts2·second2 per meter2) as a function of frequency f (in hertz). Again, these units of measure are not dimensionally correct in the true sense of energy density as defined in physics. Dividing by Zo, the characteristic impedance of free space (in ohms), the dimensions become joule-seconds per meter2 or, equivalently, joules per meter2 per hertz, which is dimensionally correct in SI
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
The speed of a wave is a product of its wavelength and what else?
A. frequency
B. density
C. voltage
D. magnitude
Answer:
|
|
sciq-6293
|
multiple_choice
|
What kind of mouthparts do aphids have?
|
[
"siphoning mouthparts",
"sponging mouthparts",
"piercing-sucking mouthparts",
"chewing mouthparts"
] |
C
|
Relavent Documents:
Document 0:::
This glossary describes the terms used in formal descriptions of spiders; where applicable these terms are used in describing other arachnids.
Links within the glossary are shown .
Terms
A
Abdomen or opisthosoma: One of the two main body parts (tagmata), located towards the posterior end; see also Abdomen § Other animals
Accessory claw: Modified at the tip of the in web-building spiders; used with to grip strands of the web
Anal tubercle: A small protuberance (tubercule) above the through which the anus opens
Apodeme: see
Apophysis (plural apophyses): An outgrowth or process changing the general shape of a body part, particularly the appendages; often used in describing the male : see
Atrium (plural atria): An internal chamber at the entrance to the in female haplogyne spiders
B
Bidentate: Having two
Book lungs: Respiratory organs on the ventral side (underside) of the , in front of the , opening through narrow slits; see also Book lungs
Branchial operculum: see
Bulbus: see
C
Calamistrum (plural calamistra): Modified setae (bristles) on the of the fourth leg of spiders with a , arranged in one or more rows or in an oval shape, used to comb silk produced by the cribellum; see also Calamistrum
Caput (plural capita): see
Carapace: A hardened plate (sclerite) covering the upper (dorsal) portion of the ; see also Carapace
Carpoblem: The principal on the male ; also just called the tibial apophysis
Cephalic region or caput: The front part of the , separated from the thoracic region by the
Cephalothorax or prosoma: One of the two main body parts (tagmata), located towards the anterior end, composed of the head ( or caput) and the thorax (thoracic region), the two regions being separated by the ; covered by the and bearing the , legs, and mouthparts
Cervical groove: A shallow U-shaped groove, separating the and thoracic regions of the
Chelate: A description of a where the closes against a tooth-like process
Chelic
Document 1:::
Trapdoor spider is a common name that is used to refer to various spiders from several different groups that create burrows with a silk-hinged trapdoor to help them ambush prey.
Several families within the infraorder Mygalomorphae contain trapdoor spiders:
Actinopodidae, a family otherwise known as 'mouse-spiders', in South America and Australia
Antrodiaetidae, a family of 'folding trapdoor spiders' from the United States and Japan
Barychelidae, a family of 'brush-footed trapdoor spiders' with pantropical distribution
Ctenizidae, a family of 'cork-lid trapdoor spiders' in tropical and subtropical regions
Cyrtaucheniidae, a family of 'wafer-lid trapdoor spiders, with wide distribution except cooler regions
Euctenizidae, a family of spiders that make wafer-like or cork-like trapdoors
Halonoproctidae, a family of spiders that make wafer-like or cork-like trapdoors and includes the phragmotic genus Cyclocosmia
Idiopidae, a family of 'spurred-trapdoor spiders' or 'armoured trapdoors' mostly in Southern Hemisphere
Migidae, also known as 'ridge fanged trapdoor spiders' or 'tree trapdoor spiders', in the Southern Hemisphere
Nemesiidae, a family of 'tube trapdoor spiders', with both tropical and temperate species worldwide
Theraphosidae, a family of tarantulas (where just a few species make trapdoors), also with wide distribution
There is also one family of trapdoor spiders in the suborder Mesothelae:
Liphistiidae, an unusual and unique family of spiders with armoured abdomens from Southeast Asia, China and Japan
Set index articles on spiders
Set index articles on animal common names
Document 2:::
This is a list of honeydew sources. Honeydew is a sugary excretion from plant sap sucking insects such as aphids or scales. There are many trees that are hosts to aphids and scale insects that produce honeydew
Honeydew sources
Document 3:::
A leaf litter sieve is a piece of equipment used by entomologists, in particular by coleopterists (beetle collectors) (Cooter 1991, page 7) as an aid to finding invertebrates in leaf litter.
A typical leaf litter sieve consists of a gauze with holes of approximately 5 to 10 mm width. The entomologist places handfuls of leaf litter into the sieve, which is placed above a white sheet or tray. The sieve is shaken, and insects are separated from the leaf litter and fall out for inspection. Charles Valentine Riley details use of a simple sieve with a cloth bag.
A more complex combination sieve is described by Hongfu.
See also
Tullgren funnel
Document 4:::
This glossary of entomology describes terms used in the formal study of insect species by entomologists.
A–C
A synthetic chlorinated hydrocarbon insecticide, toxic to vertebrates. Though its phytotoxicity is low, solvents in some formulations may damage certain crops. cf. the related Dieldrin, Endrin, Isodrin
D–F
A synthetic chlorinated hydrocarbon insecticide, toxic to vertebrates. cf. the related Aldrin, Endrin, Isodrin
A synthetic chlorinated hydrocarbon insecticide, toxic to vertebrates. Though its phytotoxicity is low, solvents in some formulations may damage certain crops. cf. the related Dieldrin, Aldrin, Isodrin
G–L
A synthetic chlorinated hydrocarbon insecticide, toxic to vertebrates. Though its phytotoxicity is low, solvents in some formulations may damage certain crops. cf. the related Dieldrin, Aldrin, Endrin
M–O
P–R
S–Z
Figures
See also
Anatomical terms of location
Butterfly
Caterpillar
Comstock–Needham system
External morphology of Lepidoptera
Glossary of ant terms
Glossary of spider terms
Glossary of scientific names
Insect wing
Pupa
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What kind of mouthparts do aphids have?
A. siphoning mouthparts
B. sponging mouthparts
C. piercing-sucking mouthparts
D. chewing mouthparts
Answer:
|
|
sciq-8568
|
multiple_choice
|
What is the layer above the troposphere?
|
[
"earth",
"condensation",
"stratosphere",
"sediment"
] |
C
|
Relavent Documents:
Document 0:::
Atmospheric temperature is a measure of temperature at different levels of the Earth's atmosphere. It is governed by many factors, including incoming solar radiation, humidity and altitude. When discussing surface air temperature, the annual atmospheric temperature range at any geographical location depends largely upon the type of biome, as measured by the Köppen climate classification
Temperature versus altitude
Temperature varies greatly at different heights relative to Earth's surface and this variation in temperature characterizes the four layers that exist in the atmosphere. These layers include the troposphere, stratosphere, mesosphere, and thermosphere.
The troposphere is the lowest of the four layers, extending from the surface of the Earth to about into the atmosphere where the tropopause (the boundary between the troposphere stratosphere) is located. The width of the troposphere can vary depending on latitude, for example, the troposphere is thicker in the tropics (about ) because the tropics are generally warmer, and thinner at the poles (about ) because the poles are colder. Temperatures in the atmosphere decrease with height at an average rate of 6.5°C (11.7°F) per kilometer. Because the troposphere experiences its warmest temperatures closer to Earth's surface, there is great vertical movement of heat and water vapour, causing turbulence. This turbulence, in conjunction with the presence of water vapour, is the reason that weather occurs within the troposphere.
Following the tropopause is the stratosphere. This layer extends from the tropopause to the stratopause which is located at an altitude of about . Temperatures remain constant with height from the tropopause to an altitude of , after which they start to increase with height. This happening is referred to as an inversion and It is because of this inversion that the stratosphere is not characterised as turbulent. The stratosphere receives its warmth from the sun and the ozone layer which ab
Document 1:::
Aeronomy is the scientific study of the upper atmosphere of the Earth and corresponding regions of the atmospheres of other planets. It is a branch of both atmospheric chemistry and atmospheric physics. Scientists specializing in aeronomy, known as aeronomers, study the motions and chemical composition and properties of the Earth's upper atmosphere and regions of the atmospheres of other planets that correspond to it, as well as the interaction between upper atmospheres and the space environment. In atmospheric regions aeronomers study, chemical dissociation and ionization are important phenomena.
History
The mathematician Sydney Chapman introduced the term aeronomy to describe the study of the Earth's upper atmosphere in 1946 in a letter to the editor of Nature entitled "Some Thoughts on Nomenclature." The term became official in 1954 when the International Union of Geodesy and Geophysics adopted it. "Aeronomy" later also began to refer to the study of the corresponding regions of the atmospheres of other planets.
Branches
Aeronomy can be divided into three main branches: terrestrial aeronomy, planetary aeronomy, and comparative aeronomy.
Terrestrial aeronomy
Terrestrial aeronomy focuses on the Earth's upper atmosphere, which extends from the stratopause to the atmosphere's boundary with outer space and is defined as consisting of the mesosphere, thermosphere, and exosphere and their ionized component, the ionosphere. Terrestrial aeronomy contrasts with meteorology, which is the scientific study of the Earth's lower atmosphere, defined as the troposphere and stratosphere. Although terrestrial aeronomy and meteorology once were completely separate fields of scientific study, cooperation between terrestrial aeronomers and meteorologists has grown as discoveries made since the early 1990s have demonstrated that the upper and lower atmospheres have an impact on one another's physics, chemistry, and biology.
Terrestrial aeronomers study atmospheric tides and upper-
Document 2:::
In aviation, ceiling is a measurement of the height of the base of the lowest clouds (not to be confused with cloud base which has a specific definition) that cover more than half of the sky (more than 4 oktas) relative to the ground. Ceiling is not specifically reported as part of the METAR (METeorological Aviation Report) used for flight planning by pilots worldwide, but can be deduced from the lowest height with broken (BKN) or overcast (OVC) reported. A ceiling listed as "unlimited" means either that the sky is mostly free of cloud cover, or that the cloud is high enough not to impede Visual Flight Rules (VFR) operation.
Definitions
ICAO The height above the ground or water of the base of the lowest layer of cloud below 6000 meters (20,000 feet) covering more than half the sky.
United Kingdom The vertical distance from the elevation of an aerodrome to the lowest part of any cloud visible from the aerodrome which is sufficient to obscure more than half of the sky.
United States The height above the Earth's surface of the lowest layer of clouds or obscuring phenomena that is reported as broken, overcast, or obscuration, and not classified as thin or partial.
See also
Cloud base
Document 3:::
A pre-STEM program is a course of study at any two-year college that prepares a student to transfer to a four-year school to earn a bachelor's degree in a STEM field.
Overview
The concept of a pre-STEM program is being developed to address America's need for more college-trained professionals in science, technology, engineering, and mathematics (STEM). It is an innovation meant to fill a gap at community colleges that do not have 'major' degree paths that students identify with on their way to earning an Associates degree. Students must complete a considerable amount of STEM coursework before transferring from a two-year school to a four-year school and earn a baccalaureate degree in a STEM field. Schools with a pre-STEM program are able to identify those students and support them with STEM-specific academic and career advising, increasing the student's chances of going on to earn a STEM baccalaureate degree in a timely fashion.
With over 50% of America's college-bound students starting their college career at public or private two-year school, and with a very small proportion of students who start college at a two-year school matriculating to and earning STEM degrees from four-year schools, pre-STEM programs have great potential for broadening participation in baccalaureate STEM studies.
Example programs
The effectiveness of pre-STEM programs is being investigated by a consortium of schools in Missouri: Moberly Area Community College, St. Charles Community College, Metropolitan Community College, and Truman State University.
A larger group of schools met at the Belknap Springs Meetings in October 2009 to discuss the challenges and opportunities presented by STEM-focused partnerships between 2-year and 4-year schools. Each program represented a two-year school and a four-year school that were trying to increase the number of people who earn a baccalaureate degree in a STEM area through various means, some of which were pre-STEM programs. Other methods includes
Document 4:::
Planetary oceanography also called astro-oceanography or exo-oceanography is the study of oceans on planets and moons other than Earth. Unlike other planetary sciences like astrobiology, astrochemistry and planetary geology, it only began after the discovery of underground oceans in Saturn's moon Titan and Jupiter's moon Europa. This field remains speculative until further missions reach the oceans beneath the rock or ice layer of the moons. There are many theories about oceans or even ocean worlds of celestial bodies in the Solar System, from oceans made of diamond in Neptune to a gigantic ocean of liquid hydrogen that may exist underneath Jupiter's surface.
Early in their geologic histories, Mars and Venus are theorized to have had large water oceans. The Mars ocean hypothesis suggests that nearly a third of the surface of Mars was once covered by water, and a runaway greenhouse effect may have boiled away the global ocean of Venus. Compounds such as salts and ammonia dissolved in water lower its freezing point so that water might exist in large quantities in extraterrestrial environments as brine or convecting ice. Unconfirmed oceans are speculated beneath the surface of many dwarf planets and natural satellites; notably, the ocean of the moon Europa is estimated to have over twice the water volume of Earth's. The Solar System's giant planets are also thought to have liquid atmospheric layers of yet to be confirmed compositions. Oceans may also exist on exoplanets and exomoons, including surface oceans of liquid water within a circumstellar habitable zone. Ocean planets are a hypothetical type of planet with a surface completely covered with liquid.
Extraterrestrial oceans may be composed of water or other elements and compounds. The only confirmed large stable bodies of extraterrestrial surface liquids are the lakes of Titan, which are made of hydrocarbons instead of water. However, there is strong evidence for subsurface water oceans' existence elsewhere in t
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is the layer above the troposphere?
A. earth
B. condensation
C. stratosphere
D. sediment
Answer:
|
|
sciq-4962
|
multiple_choice
|
The earth’s biomes are categorized into two major groups named what?
|
[
"icelandic and aquatic",
"terrestrial and galactic",
"aquatic and galactic",
"terrestrial and aquatic"
] |
D
|
Relavent Documents:
Document 0:::
Ecological classification or ecological typology is the classification of land or water into geographical units that represent variation in one or more ecological features. Traditional approaches focus on geology, topography, biogeography, soils, vegetation, climate conditions, living species, habitats, water resources, and sometimes also anthropic factors. Most approaches pursue the cartographical delineation or regionalisation of distinct areas for mapping and planning.
Approaches to classifications
Different approaches to ecological classifications have been developed in terrestrial, freshwater and marine disciplines. Traditionally these approaches have focused on biotic components (vegetation classification), abiotic components (environmental approaches) or implied ecological and evolutionary processes (biogeographical approaches). Ecosystem classifications are specific kinds of ecological classifications that consider all four elements of the definition of ecosystems: a biotic component, an abiotic complex, the interactions between and within them, and the physical space they occupy (ecotope).
Vegetation classification
Vegetation is often used to classify terrestrial ecological units. Vegetation classification can be based on vegetation structure and floristic composition. Classifications based entirely on vegetation structure overlap with land cover mapping categories.
Many schemes of vegetation classification are in use by the land, resource and environmental management agencies of different national and state jurisdictions. The International Vegetation Classification (IVC or EcoVeg) has been recently proposed but has not been yet widely adopted.
Vegetation classifications have limited use in aquatic systems, since only a handful of freshwater or marine habitats are dominated by plants (e.g. kelp forests or seagrass meadows). Also, some extreme terrestrial environments, like subterranean or cryogenic ecosystems, are not properly described in vegetation c
Document 1:::
A biochore is a subdivision of the biosphere consisting of biotopes that resemble one another and thus are colonized by similar biota. The concept is relevant in biogeography to refer to a units regardless it rank (regardless the scale).
Document 2:::
A biome () is a biogeographical unit consisting of a biological community that has formed in response to the physical environment in which they are found and a shared regional climate. Biomes may span more than one continent. Biome is a broader term than habitat and can comprise a variety of habitats.
While a biome can cover small areas, a microbiome is a mix of organisms that coexist in a defined space on a much smaller scale. For example, the human microbiome is the collection of bacteria, viruses, and other microorganisms that are present on or in a human body.
A biota is the total collection of organisms of a geographic region or a time period, from local geographic scales and instantaneous temporal scales all the way up to whole-planet and whole-timescale spatiotemporal scales. The biotas of the Earth make up the biosphere.
Etymology
The term was suggested in 1916 by Clements, originally as a synonym for biotic community of Möbius (1877). Later, it gained its current definition, based on earlier concepts of phytophysiognomy, formation and vegetation (used in opposition to flora), with the inclusion of the animal element and the exclusion of the taxonomic element of species composition. In 1935, Tansley added the climatic and soil aspects to the idea, calling it ecosystem. The International Biological Program (1964–74) projects popularized the concept of biome.
However, in some contexts, the term biome is used in a different manner. In German literature, particularly in the Walter terminology, the term is used similarly as biotope (a concrete geographical unit), while the biome definition used in this article is used as an international, non-regional, terminology—irrespectively of the continent in which an area is present, it takes the same biome name—and corresponds to his "zonobiome", "orobiome" and "pedobiome" (biomes determined by climate zone, altitude or soil).
In Brazilian literature, the term "biome" is sometimes used as synonym of biogeographic pr
Document 3:::
Biogeoclimatic ecosystem classification (BEC) is an ecological classification framework used in British Columbia to define, describe, and map ecosystem-based units at various scales, from broad, ecologically-based climatic regions down to local ecosystems or sites. BEC is termed an ecosystem classification as the approach integrates site, soil, and vegetation characteristics to develop and characterize all units. BEC has a strong application focus and guides to classification and management of forests, grasslands and wetlands are available for much of the province to aid in identification of the ecosystem units.
History
The biogeoclimatic ecosystem classification (BEC) system evolved from the work of Vladimir J. Krajina, a Czech-trained professor of ecology and botany at the University of British Columbia and his students, from 1949 - 1970. Krajina conceptualized the biogeoclimatic approach as an attempt to describe the ecologically diverse and largely undescribed landscape of British Columbia, the mountainous western-most province of Canada, using a unique blend of various contemporary traditions. These included the American tradition of community change and climax, the state factor concept of Jenny, the Braun-Blanquet approach, the Russian biogeocoenose, and environmental grids, and the European microscopic pedology approach
The biogeoclimatic approach was subsequently adopted by the Forest Service of British Columbia in 1976—initially as a five-year program to develop the classification to assist with tree species selection in reforestation. The classification concepts adopted from Krajina were modified by the staff of the B.C. Forest Service in the implementation of a provincial classification. Over the past 40 years, the BEC approach has been expanded and applied to all regions of British Columbia. It has developed into a comprehensive framework for understanding ecosystems in a climatically and topographically complex region.
Classification Framework
Biog
Document 4:::
Earth system science (ESS) is the application of systems science to the Earth. In particular, it considers interactions and 'feedbacks', through material and energy fluxes, between the Earth's sub-systems' cycles, processes and "spheres"—atmosphere, hydrosphere, cryosphere, geosphere, pedosphere, lithosphere, biosphere, and even the magnetosphere—as well as the impact of human societies on these components. At its broadest scale, Earth system science brings together researchers across both the natural and social sciences, from fields including ecology, economics, geography, geology, glaciology, meteorology, oceanography, climatology, paleontology, sociology, and space science. Like the broader subject of systems science, Earth system science assumes a holistic view of the dynamic interaction between the Earth's spheres and their many constituent subsystems fluxes and processes, the resulting spatial organization and time evolution of these systems, and their variability, stability and instability. Subsets of Earth System science include systems geology and systems ecology, and many aspects of Earth System science are fundamental to the subjects of physical geography and climate science.
Definition
The Science Education Resource Center, Carleton College, offers the following description: "Earth System science embraces chemistry, physics, biology, mathematics and applied sciences in transcending disciplinary boundaries to treat the Earth as an integrated system. It seeks a deeper understanding of the physical, chemical, biological and human interactions that determine the past, current and future states of the Earth. Earth System science provides a physical basis for understanding the world in which we live and upon which humankind seeks to achieve sustainability".
Earth System science has articulated four overarching, definitive and critically important features of the Earth System, which include:
Variability: Many of the Earth System's natural 'modes' and variab
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
The earth’s biomes are categorized into two major groups named what?
A. icelandic and aquatic
B. terrestrial and galactic
C. aquatic and galactic
D. terrestrial and aquatic
Answer:
|
|
sciq-7288
|
multiple_choice
|
What is the term for the decayed remains of living organisms?
|
[
"intermediate material",
"necrosis",
"humus",
"waste"
] |
C
|
Relavent Documents:
Document 0:::
In biology, detritus () is dead particulate organic material, as distinguished from dissolved organic material. Detritus typically includes the bodies or fragments of bodies of dead organisms, and fecal material. Detritus typically hosts communities of microorganisms that colonize and decompose (i.e. remineralize) it. In terrestrial ecosystems it is present as leaf litter and other organic matter that is intermixed with soil, which is denominated "soil organic matter". The detritus of aquatic ecosystems is organic substances that is suspended in the water and accumulates in depositions on the floor of the body of water; when this floor is a seabed, such a deposition is denominated "marine snow".
Theory
The corpses of dead plants or animals, material derived from animal tissues (e.g. molted skin), and fecal matter gradually lose their form due to physical processes and the action of decomposers, including grazers, bacteria, and fungi. Decomposition, the process by which organic matter is decomposed, occurs in several phases. Micro- and macro-organisms that feed on it rapidly consume and absorb materials such as proteins, lipids, and sugars that are low in molecular weight, while other compounds such as complex carbohydrates are decomposed more slowly. The decomposing microorganisms degrade the organic materials so as to gain the resources they require for their survival and reproduction. Accordingly, simultaneous to microorganisms' decomposition of the materials of dead plants and animals is their assimilation of decomposed compounds to construct more of their biomass (i.e. to grow their own bodies). When microorganisms die, fine organic particles are produced, and if small animals that feed on microorganisms eat these particles they collect inside the intestines of the consumers, and change shape into large pellets of dung. As a result of this process, most of the materials of dead organisms disappear and are not visible and recognizable in any form, but are pres
Document 1:::
Microbiology of decomposition is the study of all microorganisms involved in decomposition, the chemical and physical processes during which organic matter is broken down and reduced to its original elements.
Decomposition microbiology can be divided into two fields of interest, namely the decomposition of plant materials and the decomposition of cadavers and carcasses.
The decomposition of plant materials is commonly studied in order to understand the cycling of carbon within a given environment and to understand the subsequent impacts on soil quality. Plant material decomposition is also often referred to as composting. The decomposition of cadavers and carcasses has become an important field of study within forensic taphonomy.
Decomposition microbiology of plant materials
The breakdown of vegetation is highly dependent on oxygen and moisture levels. During decomposition, microorganisms require oxygen for their respiration. If anaerobic conditions dominate the decomposition environment, microbial activity will be slow and thus decomposition will be slow. Appropriate moisture levels are required for microorganisms to proliferate and to actively decompose organic matter. In arid environments, bacteria and fungi dry out and are unable to take part in decomposition. In wet environments, anaerobic conditions will develop and decomposition can also be considerably slowed down. Decomposing microorganisms also require the appropriate plant substrates in order to achieve good levels of decomposition. This usually translates to having appropriate carbon to nitrogen ratios (C:N). The ideal composting carbon-to-nitrogen ratio is thought to be approximately 30:1. As in any microbial process, the decomposition of plant litter by microorganisms will also be dependent on temperature. For example, leaves on the ground will not undergo decomposition during the winter months where snow cover occurs as temperatures are too low to sustain microbial activities.
Decomposition mi
Document 2:::
Decomposition is the process in which the organs and complex molecules of animal and human bodies break down into simple organic matter over time. In vertebrates, five stages of decomposition are typically recognized: fresh, bloat, active decay, advanced decay, and dry/skeletonized. Knowing the different stages of decomposition can help investigators in determining the post-mortem interval (PMI). The rate of decomposition of human remains can vary due to environmental factors and other factors. Environmental factors include temperature, burning, humidity, and the availability of oxygen. Other factors include body size, clothing, and the cause of death.
Stages and characteristics
The five stages of decomposition—fresh (aka autolysis), bloat, active decay, advanced decay, and dry/skeletonized—have specific characteristics that are used to identify which stage the remains are in. These stages are illustrated by reference to an experimental study of the decay of a pig corpse.
Fresh
At this stage the remains are usually intact and free of insects. The corpse progresses through algor mortis (a reduction in body temperature until ambient temperature is reached), rigor mortis (the temporary stiffening of the limbs due to chemical changes in the muscles), and livor mortis (pooling of the blood on the side of the body that is closest to the ground).
Bloat
At this stage, the microorganisms residing in the digestive system begin to digest the tissues of the body, excreting gases that cause the torso and limbs to bloat, and producing foul-smelling chemicals including putrescine and cadaverine. Cells in tissues break down and release hydrolytic enzymes, and the top layer of skin may become loosened, leading to skin slippage. Decomposition of the gastrointestinal tract results in a dark, foul-smelling liquid called "purge fluid" that is forced out of the nose and mouth due to gas pressure in the intestine. The bloat stage is characterized by a shift in the bacterial popul
Document 3:::
The necrobiome has been defined as the community of species associated with decaying corpse remains. The process of decomposition is complex. Microbes decompose cadavers, but other organisms including fungi, nematodes, insects, and larger scavenger animals also contribute. Once the immune system is no longer active, microbes colonizing the intestines and lungs decompose their respective tissues and then travel throughout the body via the blood and lymphatic systems to break down other tissue and bone. During this process, gases are released as a by-product and accumulate, causing bloating. Eventually, the gases seep through the body's wounds and natural openings, providing a way for some microbes to exit from the inside of the cadaver and inhabit the outside. The microbial communities colonizing the internal organs of a cadaver are referred to as the thanatomicrobiome. The region outside of the cadaver that is exposed to the external environment is referred to as the epinecrotic portion of the necrobiome, and is especially important when determining the time and location of death for an individual. Different microbes play specific roles during each stage of the decomposition process. The microbes that will colonize the cadaver and the rate of their activity are determined by the cadaver itself and the cadaver's surrounding environmental conditions.
History
There is textual evidence that human cadavers were first studied around the third century BC to gain an understanding of human anatomy. Many of the first human cadaver studies took place in Italy, where the earliest record of determining the cause of death from a human corpse dates back to 1286. However, understanding of the human body progressed slowly, in part because the spread of Christianity and other religious beliefs resulted in human dissection becoming illegal. Thus, non-human animals were solely dissected for anatomical understanding until the 13th century when officials realized human cadavers were ne
Document 4:::
Decomposers are organisms that break down dead or decaying organisms; they carry out decomposition, a process possible by only certain kingdoms, such as fungi. Like herbivores and predators, decomposers are heterotrophic, meaning that they use organic substrates to get their energy, carbon and nutrients for growth and development. While the terms decomposer and detritivore are often interchangeably used, detritivores ingest and digest dead matter internally, while decomposers directly absorb nutrients through external chemical and biological processes. Thus, invertebrates such as earthworms, woodlice, and sea cucumbers are technically detritivores, not decomposers, since they are unable to absorb nutrients without ingesting them.
Fungi
The primary decomposer of litter in many ecosystems is fungi. Unlike bacteria, which are unicellular organisms and are decomposers as well, most saprotrophic fungi grow as a branching network of hyphae. While bacteria are restricted to growing and feeding on the exposed surfaces of organic matter, fungi can use their hyphae to penetrate larger pieces of organic matter, below the surface. Additionally, only wood-decay fungi have evolved the enzymes necessary to decompose lignin, a chemically complex substance found in wood. These two factors make fungi the primary decomposers in forests, where litter has high concentrations of lignin and often occurs in large pieces. Fungi decompose organic matter by releasing enzymes to break down the decaying material, after which they absorb the nutrients in the decaying material. Hyphae are used to break down matter and absorb nutrients and are also used in reproduction. When two compatible fungi hyphae grow close to each other, they will then fuse together for reproduction, and form another fungus.
See also
Chemotroph
Micro-animals
Microorganism
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is the term for the decayed remains of living organisms?
A. intermediate material
B. necrosis
C. humus
D. waste
Answer:
|
|
sciq-9109
|
multiple_choice
|
Insect wings are part of the exoskeleton and attached to the what?
|
[
"head",
"spines",
"thorax",
"neck"
] |
C
|
Relavent Documents:
Document 0:::
Hybrid Insect Micro-Electro-Mechanical Systems (HI-MEMS) is a project of DARPA, a unit of the United States Department of Defense. Created in 2006, the unit's goal is the creation of tightly coupled machine-insect interfaces by placing micro-mechanical systems inside the insects during the early stages of metamorphosis. After implantation, the "insect cyborgs" could be controlled by sending electrical impulses to their muscles. The primary application is surveillance. The project was created with the ultimate goal of delivering an insect within 5 meters of a target located 100 meters away from its starting point. In 2008, a team from the University of Michigan demonstrated a cyborg unicorn beetle at an academic conference in Tucson, Arizona. The beetle was able to take off and land, turn left or right, and demonstrate other flight behaviors. Researchers at Cornell University demonstrated the successful implantation of electronic probes into tobacco hornworms in the pupal stage.
Document 1:::
Arthropods are covered with a tough, resilient integument or exoskeleton of chitin. Generally the exoskeleton will have thickened areas in which the chitin is reinforced or stiffened by materials such as minerals or hardened proteins. This happens in parts of the body where there is a need for rigidity or elasticity. Typically the mineral crystals, mainly calcium carbonate, are deposited among the chitin and protein molecules in a process called biomineralization. The crystals and fibres interpenetrate and reinforce each other, the minerals supplying the hardness and resistance to compression, while the chitin supplies the tensile strength. Biomineralization occurs mainly in crustaceans. In insects and arachnids, the main reinforcing materials are various proteins hardened by linking the fibres in processes called sclerotisation and the hardened proteins are called sclerotin. The dorsal tergum, ventral sternum, and the lateral pleura form the hardened plates or sclerites of a typical body segment.
In either case, in contrast to the carapace of a tortoise or the cranium of a vertebrate, the exoskeleton has little ability to grow or change its form once it has matured. Except in special cases, whenever the animal needs to grow, it moults, shedding the old skin after growing a new skin from beneath.
Microscopic structure
A typical arthropod exoskeleton is a multi-layered structure with four functional regions: epicuticle, procuticle, epidermis and basement membrane. Of these, the epicuticle is a multi-layered external barrier that, especially in terrestrial arthropods, acts as a barrier against desiccation. The strength of the exoskeleton is provided by the underlying procuticle, which is in turn secreted by the epidermis. Arthropod cuticle is a biological composite material, consisting of two main portions: fibrous chains of alpha-chitin within a matrix of silk-like and globular proteins, of which the best-known is the rubbery protein called resilin. The rel
Document 2:::
This glossary describes the terms used in formal descriptions of spiders; where applicable these terms are used in describing other arachnids.
Links within the glossary are shown .
Terms
A
Abdomen or opisthosoma: One of the two main body parts (tagmata), located towards the posterior end; see also Abdomen § Other animals
Accessory claw: Modified at the tip of the in web-building spiders; used with to grip strands of the web
Anal tubercle: A small protuberance (tubercule) above the through which the anus opens
Apodeme: see
Apophysis (plural apophyses): An outgrowth or process changing the general shape of a body part, particularly the appendages; often used in describing the male : see
Atrium (plural atria): An internal chamber at the entrance to the in female haplogyne spiders
B
Bidentate: Having two
Book lungs: Respiratory organs on the ventral side (underside) of the , in front of the , opening through narrow slits; see also Book lungs
Branchial operculum: see
Bulbus: see
C
Calamistrum (plural calamistra): Modified setae (bristles) on the of the fourth leg of spiders with a , arranged in one or more rows or in an oval shape, used to comb silk produced by the cribellum; see also Calamistrum
Caput (plural capita): see
Carapace: A hardened plate (sclerite) covering the upper (dorsal) portion of the ; see also Carapace
Carpoblem: The principal on the male ; also just called the tibial apophysis
Cephalic region or caput: The front part of the , separated from the thoracic region by the
Cephalothorax or prosoma: One of the two main body parts (tagmata), located towards the anterior end, composed of the head ( or caput) and the thorax (thoracic region), the two regions being separated by the ; covered by the and bearing the , legs, and mouthparts
Cervical groove: A shallow U-shaped groove, separating the and thoracic regions of the
Chelate: A description of a where the closes against a tooth-like process
Chelic
Document 3:::
Cephalization is an evolutionary trend in which, over many generations, the mouth, sense organs, and nerve ganglia become concentrated at the front end of an animal, producing a head region. This is associated with movement and bilateral symmetry, such that the animal has a definite head end. This led to the formation of a highly sophisticated brain in three groups of animals, namely the arthropods, cephalopod molluscs, and vertebrates.
Animals without bilateral symmetry
Cnidaria, such as the radially symmetrical Hydrozoa, show some degree of cephalization. The Anthomedusae have a head end with their mouth, photoreceptive cells, and a concentration of neural cells.
Bilateria
Cephalization is a characteristic feature of the Bilateria, a large group containing the majority of animal phyla. These have the ability to move, using muscles, and a body plan with a front end that encounters stimuli first as the animal moves forwards, and accordingly has evolved to contain many of the body's sense organs, able to detect light, chemicals, and gravity. There is often also a collection of nerve cells able to process the information from these sense organs, forming a brain in several phyla and one or more ganglia in others.
Acoela
The Acoela are basal bilaterians, part of the Xenacoelomorpha. They are small and simple animals, and have very slightly more nerve cells at the head end than elsewhere, not forming a distinct and compact brain. This represents an early stage in cephalization.
Flatworms
The Platyhelminthes (flatworms) have a more complex nervous system than the Acoela, and are lightly cephalized, for instance having an eyespot above the brain, near the front end.
Complex active bodies
The philosopher Michael Trestman noted that three bilaterian phyla, namely the arthropods, the molluscs in the shape of the cephalopods, and the chordates, were distinctive in having "complex active bodies", something that the acoels and flatworms did not have. Any such animal, whe
Document 4:::
The protocerebrum is the first segment of the panarthropod brain.
Recent studies suggest that it comprises two regions.
Region associated with the expression of six3
six3 is a transcription factor that marks the anteriormost part of the developing body in a whole host of Metazoa.
In the panarthropod brain, the anteriormost (rostralmost) part of the germband expresses six3. This region is described as medial, and corresponds to the annelid prostomium.
In arthropods, it contains the pars intercerebralis and pars lateralis.
six3 is associated with the euarthropod labrum and the onychophoran frontal appendages (antennae).
Region associated with the expression of orthodenticle
The other region expresses homologues of orthodenticle, Otx or otd. This region is more caudal and lateral, and bears the eyes.
Orthodenticle is associated with the protocerebral bridge, part of the central complex, traditionally a marker of the prosocerebrum.
In the annelid brain, Otx expression characterises the peristomium, but also creeps forwards into the regions of the prostomium that bear the larval eyes.
Names of regions
Inconsistent use of the terms archicerebrum and the prosocerebrum makes them confusing.
The regions were defined by Siewing (1963): the archicerebrum as containing the ocular lobes and the mushroom bodies (= corpora pedunculata), and the prosocerebrum as comprising the central complex.
The archicerebrum has traditionally been equated with the anteriormost, 'non-segmental' part of the protocerebrum, equivalent to the acron in older terminology.
The prosocerebrum is then equivalent to the 'segmental' part of the protocerebrum, bordered by segment polarity genes such as engrailed, and (on one interpretation) bearing modified segmental appendages (= camera-type eyes).
But Urbach and Technau (2003) complicate the matter by seeing the prosocerebrum (central complex) + labrum as the anteriormost region
Strausfeld 2016 identifies the anteriormost part of the b
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Insect wings are part of the exoskeleton and attached to the what?
A. head
B. spines
C. thorax
D. neck
Answer:
|
|
scienceQA-11748
|
multiple_choice
|
What do these two changes have in common?
cooking an egg
rust forming on a metal gate
|
[
"Both are chemical changes.",
"Both are only physical changes.",
"Both are caused by heating.",
"Both are caused by cooling."
] |
A
|
Step 1: Think about each change.
Cooking an egg is a chemical change. The heat causes the matter in the egg to change. Cooked egg and raw egg are different types of matter.
Rust forming on a metal gate is a chemical change. As the gate rusts, the metal turns into a different type of matter called rust. Rust is reddish-brown and falls apart easily.
Step 2: Look at each answer choice.
Both are only physical changes.
Both changes are chemical changes. They are not physical changes.
Both are chemical changes.
Both changes are chemical changes. The type of matter before and after each change is different.
Both are caused by heating.
Cooking is caused by heating. But rust forming on a metal gate is not.
Both are caused by cooling.
Neither change is caused by cooling.
|
Relavent Documents:
Document 0:::
Physical changes are changes affecting the form of a chemical substance, but not its chemical composition. Physical changes are used to separate mixtures into their component compounds, but can not usually be used to separate compounds into chemical elements or simpler compounds.
Physical changes occur when objects or substances undergo a change that does not change their chemical composition. This contrasts with the concept of chemical change in which the composition of a substance changes or one or more substances combine or break up to form new substances. In general a physical change is reversible using physical means. For example, salt dissolved in water can be recovered by allowing the water to evaporate.
A physical change involves a change in physical properties. Examples of physical properties include melting, transition to a gas, change of strength, change of durability, changes to crystal form, textural change, shape, size, color, volume and density.
An example of a physical change is the process of tempering steel to form a knife blade. A steel blank is repeatedly heated and hammered which changes the hardness of the steel, its flexibility and its ability to maintain a sharp edge.
Many physical changes also involve the rearrangement of atoms most noticeably in the formation of crystals. Many chemical changes are irreversible, and many physical changes are reversible, but reversibility is not a certain criterion for classification. Although chemical changes may be recognized by an indication such as odor, color change, or production of a gas, every one of these indicators can result from physical change.
Examples
Heating and cooling
Many elements and some compounds change from solids to liquids and from liquids to gases when heated and the reverse when cooled. Some substances such as iodine and carbon dioxide go directly from solid to gas in a process called sublimation.
Magnetism
Ferro-magnetic materials can become magnetic. The process is reve
Document 1:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 2:::
In physics, a dynamical system is said to be mixing if the phase space of the system becomes strongly intertwined, according to at least one of several mathematical definitions. For example, a measure-preserving transformation T is said to be strong mixing if
whenever A and B are any measurable sets and μ is the associated measure. Other definitions are possible, including weak mixing and topological mixing.
The mathematical definition of mixing is meant to capture the notion of physical mixing. A canonical example is the Cuba libre: suppose one is adding rum (the set A) to a glass of cola. After stirring the glass, the bottom half of the glass (the set B) will contain rum, and it will be in equal proportion as it is elsewhere in the glass. The mixing is uniform: no matter which region B one looks at, some of A will be in that region. A far more detailed, but still informal description of mixing can be found in the article on mixing (mathematics).
Every mixing transformation is ergodic, but there are ergodic transformations which are not mixing.
Physical mixing
The mixing of gases or liquids is a complex physical process, governed by a convective diffusion equation that may involve non-Fickian diffusion as in spinodal decomposition. The convective portion of the governing equation contains fluid motion terms that are governed by the Navier–Stokes equations. When fluid properties such as viscosity depend on composition, the governing equations may be coupled. There may also be temperature effects. It is not clear that fluid mixing processes are mixing in the mathematical sense.
Small rigid objects (such as rocks) are sometimes mixed in a rotating drum or tumbler. The 1969 Selective Service draft lottery was carried out by mixing plastic capsules which contained a slip of paper (marked with a day of the year).
See also
Miscibility
Document 3:::
Adaptive comparative judgement is a technique borrowed from psychophysics which is able to generate reliable results for educational assessment – as such it is an alternative to traditional exam script marking. In the approach, judges are presented with pairs of student work and are then asked to choose which is better, one or the other. By means of an iterative and adaptive algorithm, a scaled distribution of student work can then be obtained without reference to criteria.
Introduction
Traditional exam script marking began in Cambridge 1792 when, with undergraduate numbers rising, the importance of proper ranking of students was growing. So in 1792 the new Proctor of Examinations, William Farish, introduced marking, a process in which every examiner gives a numerical score to each response by every student, and the overall total mark puts the students in the final rank order. Francis Galton (1869) noted that, in an unidentified year about 1863, the Senior Wrangler scored 7,634 out of a maximum of 17,000, while the Second Wrangler scored 4,123. (The 'Wooden Spoon' scored only 237.)
Prior to 1792, a team of Cambridge examiners convened at 5pm on the last day of examining, reviewed the 19 papers each student had sat – and published their rank order at midnight. Marking solved the problems of numbers and prevented unfair personal bias, and its introduction was a step towards modern objective testing, the format it is best suited to. But the technology of testing that followed, with its major emphasis on reliability and the automatisation of marking, has been an uncomfortable partner for some areas of educational achievement: assessing writing or speaking, and other kinds of performance need something more qualitative and judgemental.
The technique of Adaptive Comparative Judgement is an alternative to marking. It returns to the pre-1792 idea of sorting papers according to their quality, but retains the guarantee of reliability and fairness. It is by far the most rel
Document 4:::
At equilibrium, the relationship between water content and equilibrium relative humidity of a material can be displayed graphically by a curve, the so-called moisture sorption isotherm.
For each humidity value, a sorption isotherm indicates the corresponding water content value at a given, constant temperature. If the composition or quality of the material changes, then its sorption behaviour also changes. Because of the complexity of sorption process the isotherms cannot be determined explicitly by calculation, but must be recorded experimentally for each product.
The relationship between water content and water activity (aw) is complex. An increase in aw is usually accompanied by an increase in water content, but in a non-linear fashion. This relationship between water activity and moisture content at a given temperature is called the moisture sorption isotherm. These curves are determined experimentally and constitute the fingerprint of a food system.
BET theory (Brunauer-Emmett-Teller) provides a calculation to describe the physical adsorption of gas molecules on a solid surface. Because of the complexity of the process, these calculations are only moderately successful; however, Stephen Brunauer was able to classify sorption isotherms into five generalized shapes as shown in Figure 2. He found that Type II and Type III isotherms require highly porous materials or desiccants, with first monolayer adsorption, followed by multilayer adsorption and finally leading to capillary condensation, explaining these materials high moisture capacity at high relative humidity.
Care must be used in extracting data from isotherms, as the representation for each axis may vary in its designation. Brunauer provided the vertical axis as moles of gas adsorbed divided by the moles of the dry material, and on the horizontal axis he used the ratio of partial pressure of the gas just over the sample, divided by its partial pressure at saturation. More modern isotherms showing the
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What do these two changes have in common?
cooking an egg
rust forming on a metal gate
A. Both are chemical changes.
B. Both are only physical changes.
C. Both are caused by heating.
D. Both are caused by cooling.
Answer:
|
sciq-6276
|
multiple_choice
|
All animals have specialized types of what basic structures, which can then do different jobs?
|
[
"proteins",
"atoms",
"muscles",
"cells"
] |
D
|
Relavent Documents:
Document 0:::
Animal science is described as "studying the biology of animals that are under the control of humankind". It can also be described as the production and management of farm animals. Historically, the degree was called animal husbandry and the animals studied were livestock species, like cattle, sheep, pigs, poultry, and horses. Today, courses available look at a broader area, including companion animals, like dogs and cats, and many exotic species. Degrees in Animal Science are offered at a number of colleges and universities. Animal science degrees are often offered at land-grant universities, which will often have on-campus farms to give students hands-on experience with livestock animals.
Education
Professional education in animal science prepares students for careers in areas such as animal breeding, food and fiber production, nutrition, animal agribusiness, animal behavior, and welfare. Courses in a typical Animal Science program may include genetics, microbiology, animal behavior, nutrition, physiology, and reproduction. Courses in support areas, such as genetics, soils, agricultural economics and marketing, legal aspects, and the environment also are offered.
Bachelor degree
At many universities, a Bachelor of Science (BS) degree in Animal Science allows emphasis in certain areas. Typical areas are species-specific or career-specific. Species-specific areas of emphasis prepare students for a career in dairy management, beef management, swine management, sheep or small ruminant management, poultry production, or the horse industry. Other career-specific areas of study include pre-veterinary medicine studies, livestock business and marketing, animal welfare and behavior, animal nutrition science, animal reproduction science, or genetics. Youth programs are also an important part of animal science programs.
Pre-veterinary emphasis
Many schools that offer a degree option in Animal Science also offer a pre-veterinary emphasis such as Iowa State University, th
Document 1:::
This list of life sciences comprises the branches of science that involve the scientific study of life – such as microorganisms, plants, and animals including human beings. This science is one of the two major branches of natural science, the other being physical science, which is concerned with non-living matter. Biology is the overall natural science that studies life, with the other life sciences as its sub-disciplines.
Some life sciences focus on a specific type of organism. For example, zoology is the study of animals, while botany is the study of plants. Other life sciences focus on aspects common to all or many life forms, such as anatomy and genetics. Some focus on the micro-scale (e.g. molecular biology, biochemistry) other on larger scales (e.g. cytology, immunology, ethology, pharmacy, ecology). Another major branch of life sciences involves understanding the mindneuroscience. Life sciences discoveries are helpful in improving the quality and standard of life and have applications in health, agriculture, medicine, and the pharmaceutical and food science industries. For example, it has provided information on certain diseases which has overall aided in the understanding of human health.
Basic life science branches
Biology – scientific study of life
Anatomy – study of form and function, in plants, animals, and other organisms, or specifically in humans
Astrobiology – the study of the formation and presence of life in the universe
Bacteriology – study of bacteria
Biotechnology – study of combination of both the living organism and technology
Biochemistry – study of the chemical reactions required for life to exist and function, usually a focus on the cellular level
Bioinformatics – developing of methods or software tools for storing, retrieving, organizing and analyzing biological data to generate useful biological knowledge
Biolinguistics – the study of the biology and evolution of language.
Biological anthropology – the study of humans, non-hum
Document 2:::
A cell type is a classification used to identify cells that share morphological or phenotypical features. A multicellular organism may contain cells of a number of widely differing and specialized cell types, such as muscle cells and skin cells, that differ both in appearance and function yet have identical genomic sequences. Cells may have the same genotype, but belong to different cell types due to the differential regulation of the genes they contain. Classification of a specific cell type is often done through the use of microscopy (such as those from the cluster of differentiation family that are commonly used for this purpose in immunology). Recent developments in single cell RNA sequencing facilitated classification of cell types based on shared gene expression patterns. This has led to the discovery of many new cell types in e.g. mouse cortex, hippocampus, dorsal root ganglion and spinal cord.
Animals have evolved a greater diversity of cell types in a multicellular body (100–150 different cell types), compared
with 10–20 in plants, fungi, and protists. The exact number of cell types is, however, undefined, and the Cell Ontology, as of 2021, lists over 2,300 different cell types.
Multicellular organisms
All higher multicellular organisms contain cells specialised for different functions. Most distinct cell types arise from a single totipotent cell that differentiates into hundreds of different cell types during the course of development. Differentiation of cells is driven by different environmental cues (such as cell–cell interaction) and intrinsic differences (such as those caused by the uneven distribution of molecules during division). Multicellular organisms are composed of cells that fall into two fundamental types: germ cells and somatic cells. During development, somatic cells will become more specialized and form the three primary germ layers: ectoderm, mesoderm, and endoderm. After formation of the three germ layers, cells will continue to special
Document 3:::
The following outline is provided as an overview of and topical guide to biophysics:
Biophysics – interdisciplinary science that uses the methods of physics to study biological systems.
Nature of biophysics
Biophysics is
An academic discipline – branch of knowledge that is taught and researched at the college or university level. Disciplines are defined (in part), and recognized by the academic journals in which research is published, and the learned societies and academic departments or faculties to which their practitioners belong.
A scientific field (a branch of science) – widely recognized category of specialized expertise within science, and typically embodies its own terminology and nomenclature. Such a field will usually be represented by one or more scientific journals, where peer-reviewed research is published.
A natural science – one that seeks to elucidate the rules that govern the natural world using empirical and scientific methods.
A biological science – concerned with the study of living organisms, including their structure, function, growth, evolution, distribution, and taxonomy.
A branch of physics – concerned with the study of matter and its motion through space and time, along with related concepts such as energy and force.
An interdisciplinary field – field of science that overlaps with other sciences
Scope of biophysics research
Biomolecular scale
Biomolecule
Biomolecular structure
Organismal scale
Animal locomotion
Biomechanics
Biomineralization
Motility
Environmental scale
Biophysical environment
Biophysics research overlaps with
Agrophysics
Biochemistry
Biophysical chemistry
Bioengineering
Biogeophysics
Nanotechnology
Systems biology
Branches of biophysics
Astrobiophysics – field of intersection between astrophysics and biophysics concerned with the influence of the astrophysical phenomena upon life on planet Earth or some other planet in general.
Medical biophysics – interdisciplinary field that applies me
Document 4:::
In a multicellular organism, an organ is a collection of tissues joined in a structural unit to serve a common function. In the hierarchy of life, an organ lies between tissue and an organ system. Tissues are formed from same type cells to act together in a function. Tissues of different types combine to form an organ which has a specific function. The intestinal wall for example is formed by epithelial tissue and smooth muscle tissue. Two or more organs working together in the execution of a specific body function form an organ system, also called a biological system or body system.
An organ's tissues can be broadly categorized as parenchyma, the functional tissue, and stroma, the structural tissue with supportive, connective, or ancillary functions. For example, the gland's tissue that makes the hormones is the parenchyma, whereas the stroma includes the nerves that innervate the parenchyma, the blood vessels that oxygenate and nourish it and carry away its metabolic wastes, and the connective tissues that provide a suitable place for it to be situated and anchored. The main tissues that make up an organ tend to have common embryologic origins, such as arising from the same germ layer. Organs exist in most multicellular organisms. In single-celled organisms such as members of the eukaryotes, the functional analogue of an organ is known as an organelle. In plants, there are three main organs.
The number of organs in any organism depends on the definition used. By one widely adopted definition, 79 organs have been identified in the human body.
Animals
Except for placozoans, multicellular animals including humans have a variety of organ systems. These specific systems are widely studied in human anatomy. The functions of these organ systems often share significant overlap. For instance, the nervous and endocrine system both operate via a shared organ, the hypothalamus. For this reason, the two systems are combined and studied as the neuroendocrine system. The sam
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
All animals have specialized types of what basic structures, which can then do different jobs?
A. proteins
B. atoms
C. muscles
D. cells
Answer:
|
|
sciq-5994
|
multiple_choice
|
When scientists work in natural settings rather than a lab, it is called what?
|
[
"extracurricular activity",
"exploratory",
"outside work",
"fieldwork"
] |
D
|
Relavent Documents:
Document 0:::
A pre-STEM program is a course of study at any two-year college that prepares a student to transfer to a four-year school to earn a bachelor's degree in a STEM field.
Overview
The concept of a pre-STEM program is being developed to address America's need for more college-trained professionals in science, technology, engineering, and mathematics (STEM). It is an innovation meant to fill a gap at community colleges that do not have 'major' degree paths that students identify with on their way to earning an Associates degree. Students must complete a considerable amount of STEM coursework before transferring from a two-year school to a four-year school and earn a baccalaureate degree in a STEM field. Schools with a pre-STEM program are able to identify those students and support them with STEM-specific academic and career advising, increasing the student's chances of going on to earn a STEM baccalaureate degree in a timely fashion.
With over 50% of America's college-bound students starting their college career at public or private two-year school, and with a very small proportion of students who start college at a two-year school matriculating to and earning STEM degrees from four-year schools, pre-STEM programs have great potential for broadening participation in baccalaureate STEM studies.
Example programs
The effectiveness of pre-STEM programs is being investigated by a consortium of schools in Missouri: Moberly Area Community College, St. Charles Community College, Metropolitan Community College, and Truman State University.
A larger group of schools met at the Belknap Springs Meetings in October 2009 to discuss the challenges and opportunities presented by STEM-focused partnerships between 2-year and 4-year schools. Each program represented a two-year school and a four-year school that were trying to increase the number of people who earn a baccalaureate degree in a STEM area through various means, some of which were pre-STEM programs. Other methods includes
Document 1:::
Female education in STEM refers to child and adult female representation in the educational fields of science, technology, engineering, and mathematics (STEM). In 2017, 33% of students in STEM fields were women.
The organization UNESCO has stated that this gender disparity is due to discrimination, biases, social norms and expectations that influence the quality of education women receive and the subjects they study. UNESCO also believes that having more women in STEM fields is desirable because it would help bring about sustainable development.
Current status of girls and women in STEM education
Overall trends in STEM education
Gender differences in STEM education participation are already visible in early childhood care and education in science- and math-related play, and become more pronounced at higher levels of education. Girls appear to lose interest in STEM subjects with age, particularly between early and late adolescence. This decreased interest affects participation in advanced studies at the secondary level and in higher education. Female students represent 35% of all students enrolled in STEM-related fields of study at this level globally. Differences are also observed by disciplines, with female enrollment lowest in engineering, manufacturing and construction, natural science, mathematics and statistics and ICT fields. Significant regional and country differences in female representation in STEM studies can be observed, though, suggesting the presence of contextual factors affecting girls’ and women's engagement in these fields. Women leave STEM disciplines in disproportionate numbers during their higher education studies, in their transition to the world of work and even in their career cycle.
Learning achievement in STEM education
Data on gender differences in learning achievement present a complex picture, depending on what is measured (subject, knowledge acquisition against knowledge application), the level of education/age of students, and
Document 2:::
The Science, Technology, Engineering and Mathematics Network or STEMNET is an educational charity in the United Kingdom that seeks to encourage participation at school and college in science and engineering-related subjects (science, technology, engineering, and mathematics) and (eventually) work.
History
It is based at Woolgate Exchange near Moorgate tube station in London and was established in 1996. The chief executive is Kirsten Bodley. The STEMNET offices are housed within the Engineering Council.
Function
Its chief aim is to interest children in science, technology, engineering and mathematics. Primary school children can start to have an interest in these subjects, leading secondary school pupils to choose science A levels, which will lead to a science career. It supports the After School Science and Engineering Clubs at schools. There are also nine regional Science Learning Centres.
STEM ambassadors
To promote STEM subjects and encourage young people to take up jobs in these areas, STEMNET have around 30,000 ambassadors across the UK. these come from a wide selection of the STEM industries and include TV personalities like Rob Bell.
Funding
STEMNET used to receive funding from the Department for Education and Skills. Since June 2007, it receives funding from the Department for Children, Schools and Families and Department for Innovation, Universities and Skills, since STEMNET sits on the chronological dividing point (age 16) of both of the new departments.
See also
The WISE Campaign
Engineering and Physical Sciences Research Council
National Centre for Excellence in Teaching Mathematics
Association for Science Education
Glossary of areas of mathematics
Glossary of astronomy
Glossary of biology
Glossary of chemistry
Glossary of engineering
Glossary of physics
Document 3:::
Science, technology, engineering, and mathematics (STEM) is an umbrella term used to group together the distinct but related technical disciplines of science, technology, engineering, and mathematics. The term is typically used in the context of education policy or curriculum choices in schools. It has implications for workforce development, national security concerns (as a shortage of STEM-educated citizens can reduce effectiveness in this area), and immigration policy, with regard to admitting foreign students and tech workers.
There is no universal agreement on which disciplines are included in STEM; in particular, whether or not the science in STEM includes social sciences, such as psychology, sociology, economics, and political science. In the United States, these are typically included by organizations such as the National Science Foundation (NSF), the Department of Labor's O*Net online database for job seekers, and the Department of Homeland Security. In the United Kingdom, the social sciences are categorized separately and are instead grouped with humanities and arts to form another counterpart acronym HASS (Humanities, Arts, and Social Sciences), rebranded in 2020 as SHAPE (Social Sciences, Humanities and the Arts for People and the Economy). Some sources also use HEAL (health, education, administration, and literacy) as the counterpart of STEM.
Terminology
History
Previously referred to as SMET by the NSF, in the early 1990s the acronym STEM was used by a variety of educators, including Charles E. Vela, the founder and director of the Center for the Advancement of Hispanics in Science and Engineering Education (CAHSEE). Moreover, the CAHSEE started a summer program for talented under-represented students in the Washington, D.C., area called the STEM Institute. Based on the program's recognized success and his expertise in STEM education, Charles Vela was asked to serve on numerous NSF and Congressional panels in science, mathematics, and engineering edu
Document 4:::
Animal science is described as "studying the biology of animals that are under the control of humankind". It can also be described as the production and management of farm animals. Historically, the degree was called animal husbandry and the animals studied were livestock species, like cattle, sheep, pigs, poultry, and horses. Today, courses available look at a broader area, including companion animals, like dogs and cats, and many exotic species. Degrees in Animal Science are offered at a number of colleges and universities. Animal science degrees are often offered at land-grant universities, which will often have on-campus farms to give students hands-on experience with livestock animals.
Education
Professional education in animal science prepares students for careers in areas such as animal breeding, food and fiber production, nutrition, animal agribusiness, animal behavior, and welfare. Courses in a typical Animal Science program may include genetics, microbiology, animal behavior, nutrition, physiology, and reproduction. Courses in support areas, such as genetics, soils, agricultural economics and marketing, legal aspects, and the environment also are offered.
Bachelor degree
At many universities, a Bachelor of Science (BS) degree in Animal Science allows emphasis in certain areas. Typical areas are species-specific or career-specific. Species-specific areas of emphasis prepare students for a career in dairy management, beef management, swine management, sheep or small ruminant management, poultry production, or the horse industry. Other career-specific areas of study include pre-veterinary medicine studies, livestock business and marketing, animal welfare and behavior, animal nutrition science, animal reproduction science, or genetics. Youth programs are also an important part of animal science programs.
Pre-veterinary emphasis
Many schools that offer a degree option in Animal Science also offer a pre-veterinary emphasis such as Iowa State University, th
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
When scientists work in natural settings rather than a lab, it is called what?
A. extracurricular activity
B. exploratory
C. outside work
D. fieldwork
Answer:
|
|
sciq-5214
|
multiple_choice
|
What is the process bacteria use to break down chemicals into food?
|
[
"gametogenesis",
"biosynthesis",
"chemosynthesis",
"cellular respiration"
] |
C
|
Relavent Documents:
Document 0:::
The branches of microbiology can be classified into pure and applied sciences. Microbiology can be also classified based on taxonomy, in the cases of bacteriology, mycology, protozoology, and phycology. There is considerable overlap between the specific branches of microbiology with each other and with other disciplines, and certain aspects of these branches can extend beyond the traditional scope of microbiology
In general the field of microbiology can be divided in the more fundamental branch (pure microbiology) and the applied microbiology (biotechnology). In the more fundamental field the organisms are studied as the subject itself on a deeper (theoretical) level.
Applied microbiology refers to the fields where the micro-organisms are applied in certain processes such as brewing or fermentation. The organisms itself are often not studied as such, but applied to sustain certain processes.
Pure microbiology
Bacteriology: the study of bacteria
Mycology: the study of fungi
Protozoology: the study of protozoa
Phycology/algology: the study of algae
Parasitology: the study of parasites
Immunology: the study of the immune system
Virology: the study of viruses
Nematology: the study of nematodes
Microbial cytology: the study of microscopic and submicroscopic details of microorganisms
Microbial physiology: the study of how the microbial cell functions biochemically. Includes the study of microbial growth, microbial metabolism and microbial cell structure
Microbial pathogenesis: the study of pathogens which happen to be microbes
Microbial ecology: the relationship between microorganisms and their environment
Microbial genetics: the study of how genes are organized and regulated in microbes in relation to their cellular functions Closely related to the field of molecular biology
Cellular microbiology: a discipline bridging microbiology and cell biology
Evolutionary microbiology: the study of the evolution of microbes. This field can be subdivided into:
Micr
Document 1:::
The Society for Industrial Microbiology and Biotechnology (SIMB) is a nonprofit, international association dedicated to the advancement of microbiological sciences, especially as they apply to industrial products, biotechnology, materials, and processes. SIMB promotes the exchange of scientific information through its meetings and publications, and serves as liaison among the specialized fields of microbiology. SIMB was established in 1949 as the Society for Industrial Microbiology (SIM) by Walter Ezekiel, Charles Thom, and Charles L. Porter.
Governance
The SIMB is governed by a Constitution and Bylaws. The membership of SIMB elects a Board of Directors that consists of a President, President-Elect, Past-President, Secretary, Treasurer and four Directors.
Publications
SIMB has two publications, the Journal of Industrial Microbiology and Biotechnology and SIMB News.
Scientific Meetings
SIMB Annual Meeting
Symposium on Biomaterials, Fuels and Chemicals (SBFC)
The first Symposium on Biotechnology for Fuels and Chemicals was held in 1978 and hosted by Oak Ridge National Laboratory (Oak Ridge, TN). It was the first technical meeting focusing exclusively on the biotechnologically-‐mediated conversion of renewable feedstocks, especially lignocellulosic plant biomass, to fuels and chemicals. This annual meeting soon became large enough to be co-‐hosted by the predecessor of the National Renewable Energy Laboratory (Golden, CO) and the Symposium's location alternated yearly between Tennessee and Colorado. In 2008, SIMB began handling the logistics of the meeting and locations were expanded to include other states, with the Symposium being held in alternate years in the eastern or western United States.
Recent Advances in Fermentation Technology (RAFT)
Industrial Microbiology Meets Microbiome (IMMM)
Natural Products
Although there has been a steady decline in natural product discovery efforts in the pharmaceutical industry over the last three decades, natural pro
Document 2:::
List of Useful Microorganisms Used In preparation Of Food And Beverage
See also
Fermentation (food)
Food microbiology
Document 3:::
Biotransformation is the biochemical modification of one chemical compound or a mixture of chemical compounds. Biotransformations can be conducted with whole cells, their lysates, or purified enzymes. Increasingly, biotransformations are effected with purified enzymes. Major industries and life-saving technologies depend on biotransformations.
Advantages and disadvantages
Compared to the conventional production of chemicals, biotransformations are often attractive because their selectivities can be high, limiting the coproduction of undesirable coproducts. Generally operating under mild temperatures and pressures in aqueous solutions, many biotransformations are "green". The catalysts, i.e. the enzymes, are amenable to improvement by genetic manipulation.
Biotechnology usually is restrained by substrate scope. Petrochemicals for example are often not amenable to biotransformations, especially on the scale required for some applications, e.g. fuels. Biotransformations can be slow and are often incompatible with high temperatures, which are employed in traditional chemical synthesis to increase rates. Enzymes are generally only stable <100 °C, and usually much lower. Enzymes, like other catalysts are poisonable. In some cases, performance or recyclability can be improved by using immobilized enzymes.
Historical
Wine and beer making are examples of biotransformations that have been practiced since ancient times. Vinegar has long been produced by fermentation, involving the oxidation of ethanol to acetic acid. Cheesemaking traditionally relies on microbes to convert dairy precursors. Yogurt is produced by inoculating heat-treated milk with microorganisms such as Streptococcus thermophilus and Lactobacillus bulgaricus.
Modern examples
Pharmaceuticals
Beta-lactam antibiotics, e.g., penicillin and cephalosporin are produced by biotransformations in an industry valued several billions of dollars. Processes are conducted in vessels up to 60,000 gal in volum
Document 4:::
Industrial microbiology is a branch of biotechnology that applies microbial sciences to create industrial products in mass quantities, often using microbial cell factories. There are multiple ways to manipulate a microorganism in order to increase maximum product yields. Introduction of mutations into an organism may be accomplished by introducing them to mutagens. Another way to increase production is by gene amplification, this is done by the use of plasmids, and vectors. The plasmids and/ or vectors are used to incorporate multiple copies of a specific gene that would allow more enzymes to be produced that eventually cause more product yield. The manipulation of organisms in order to yield a specific product has many applications to the real world like the production of some antibiotics, vitamins, enzymes, amino acids, solvents, alcohol and daily products. Microorganisms play a big role in the industry, with multiple ways to be used. Medicinally, microbes can be used for creating antibiotics in order to treat infection. Microbes can also be used for the food industry as well. Microbes are very useful in creating some of the mass produced products that are consumed by people. The chemical industry also uses microorganisms in order to synthesize amino acids and organic solvents. Microbes can also be used in an agricultural application for use as a biopesticide instead of using dangerous chemicals and or inoculants to help plant proliferation.
Medical application
The medical application to industrial microbiology is the production of new drugs synthesized in a specific organism for medical purposes. Production of antibiotics is necessary for the treatment of many bacterial infections. Some natural occurring antibiotics and precursors, are produced through a process called fermentation. The microorganisms grow in a liquid media where the population size is controlled in order to yield the greatest amount of product. In this environment nutrient, pH, temperature, an
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is the process bacteria use to break down chemicals into food?
A. gametogenesis
B. biosynthesis
C. chemosynthesis
D. cellular respiration
Answer:
|
|
sciq-1052
|
multiple_choice
|
Surface tension is a property of matter that is in what state?
|
[
"Base",
"water",
"liquid",
"mixture"
] |
C
|
Relavent Documents:
Document 0:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 1:::
The Szyszkowski Equation has been used by Meissner and Michaels to describe the decrease in surface tension of aqueous solutions of carboxylic acids, alcohols and esters at varying mole fractions. It describes the exponential decrease of the surface tension at low concentrations reasonably but should be used only at concentrations below 1 mole%.
Equation
with:
σm is surface tension of the mixture
σw is surface tension of pure water
a is component specific constant (see table below)
x is mole fraction of the solvated component
The equation can be rearranged to be explicit in a:
This allows the direct calculation of that component specific parameter a from experimental data.
The equation can also be written as:
with:
γ is surface tension of the mixture
γ0 is surface tension of pure water
R is ideal gas constant 8.31 J/(mol*K)
T is temperature in K
ω is cross-sectional area of the surfactant molecules at the surface
The surface tension of pure water is dependent on temperature. At room temperature (298 K), it is equal to 71.97 mN/m
Parameters
Meissner and Michaels published the following a constants:
Example
The following table and diagram show experimentally determined surface tensions in the mixture of water and propionic acid.
This example shows a good agreement between the published value a=2.6*10−3 and the calculated value a=2.59*10−3 at the smallest given mole fraction of 0.00861 but at higher concentrations of propionic acid the value of an increases considerably, showing deviations from the predicted value.
See also
Bohdan Szyszkowski
Document 2:::
Surface tension is the tendency of liquid surfaces at rest to shrink into the minimum surface area possible. Surface tension is what allows objects with a higher density than water such as razor blades and insects (e.g. water striders) to float on a water surface without becoming even partly submerged.
At liquid–air interfaces, surface tension results from the greater attraction of liquid molecules to each other (due to cohesion) than to the molecules in the air (due to adhesion).
There are two primary mechanisms in play. One is an inward force on the surface molecules causing the liquid to contract. Second is a tangential force parallel to the surface of the liquid. This tangential force is generally referred to as the surface tension. The net effect is the liquid behaves as if its surface were covered with a stretched elastic membrane. But this analogy must not be taken too far as the tension in an elastic membrane is dependent on the amount of deformation of the membrane while surface tension is an inherent property of the liquid–air or liquid–vapour interface.
Because of the relatively high attraction of water molecules to each other through a web of hydrogen bonds, water has a higher surface tension (72.8 millinewtons (mN) per meter at 20 °C) than most other liquids. Surface tension is an important factor in the phenomenon of capillarity.
Surface tension has the dimension of force per unit length, or of energy per unit area. The two are equivalent, but when referring to energy per unit of area, it is common to use the term surface energy, which is a more general term in the sense that it applies also to solids.
In materials science, surface tension is used for either surface stress or surface energy.
Causes
Due to the cohesive forces, a molecule located away from the surface is pulled equally in every direction by neighboring liquid molecules, resulting in a net force of zero. The molecules at the surface do not have the same molecules on all sides of th
Document 3:::
In physics, the Young–Laplace equation () is an algebraic equation that describes the capillary pressure difference sustained across the interface between two static fluids, such as water and air, due to the phenomenon of surface tension or wall tension, although use of the latter is only applicable if assuming that the wall is very thin. The Young–Laplace equation relates the pressure difference to the shape of the surface or wall and it is fundamentally important in the study of static capillary surfaces. It's a statement of normal stress balance for static fluids meeting at an interface, where the interface is treated as a surface (zero thickness):
where is the Laplace pressure, the pressure difference across the fluid interface (the exterior pressure minus the interior pressure), is the surface tension (or wall tension), is the unit normal pointing out of the surface, is the mean curvature, and and are the principal radii of curvature. Note that only normal stress is considered, this is because it has been shown that a static interface is possible only in the absence of tangential stress.
The equation is named after Thomas Young, who developed the qualitative theory of surface tension in 1805, and Pierre-Simon Laplace who completed the mathematical description in the following year. It is sometimes also called the Young–Laplace–Gauss equation, as Carl Friedrich Gauss unified the work of Young and Laplace in 1830, deriving both the differential equation and boundary conditions using Johann Bernoulli's virtual work principles.
Soap films
If the pressure difference is zero, as in a soap film without gravity, the interface will assume the shape of a minimal surface.
Emulsions
The equation also explains the energy required to create an emulsion. To form the small, highly curved droplets of an emulsion, extra energy is required to overcome the large pressure that results from their small radius.
The Laplace pressure, which is greater for smaller droplets,
Document 4:::
In physics, ultradivided matter is a family of states of matter characterised by a heterogeneous mixture of two or more different materials, where the interaction energy between the suspended phase is larger than kT. The term 'ultradivided matter' encapsulates several types of substance including: soap micelles, emulsions, and suspensions of solids such as colloids.).
An ultradivided state differs from a solution. In a steady-state solution, all interactions between a solution's constituent molecules are on the order of the thermal energy kT. Thus any otherwise aggregative force between similar molecules in a solution is subordinate to thermal fluctuations, and the solution does allow flocculation of one of the constituent components. Ultradivided matter, however, is characterised by large interfacial energies where the intermolecular interactions of one or more constituents of the substance are stronger than kT. This leads to different behaviour. An intuitive example of this can be seen in the tendency of a biphasic mixture of water (a polar liquid) and oil (a non-polar liquid) to spontaneously separate into two phases. This may seem to imply a change from a state with higher disorder or higher entropy (a suspension of oil droplets in water) to a lower-entropy arrangement (a neat separation into two regions of different material). Such a transition would seem to violate the second law of thermodynamics, which is impossible. The resolution to this apparent paradox is that the interface between oil and water necessitates an ordered alignment of oil and water molecules at the interface. Minimisation of the surface area between the two phases thus correlates with an increase of the entropy of the system. The highest entropy state thus has a minimum interfacial surface area between the two phases and thus a neat separation is created, into two regions of different material.
See also
Colloid
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Surface tension is a property of matter that is in what state?
A. Base
B. water
C. liquid
D. mixture
Answer:
|
|
sciq-1446
|
multiple_choice
|
What disease is described as a respiratory disorder characterized by wheezing, coughing, and a feeling of constriction in the chest?
|
[
"asthma",
"indigestion",
"dementia",
"inflammation"
] |
A
|
Relavent Documents:
Document 0:::
Acute bronchitis, also known as a chest cold, is short-term bronchitis – inflammation of the bronchi (large and medium-sized airways) of the lungs. The most common symptom is a cough. Other symptoms include coughing up mucus, wheezing, shortness of breath, fever, and chest discomfort. The infection may last from a few to ten days. The cough may persist for several weeks afterward with the total duration of symptoms usually around three weeks. Some have symptoms for up to six weeks.
In more than 90% of cases, the cause is a viral infection. These viruses may be spread through the air when people cough or by direct contact. Risk factors include exposure to tobacco smoke, dust, and other air pollution. A small number of cases are due to high levels of air pollution or bacteria such as Mycoplasma pneumoniae or Bordetella pertussis. Diagnosis is typically based on a person's signs and symptom. The color of the sputum does not indicate if the infection is viral or bacterial. Determining the underlying organism is typically not needed. Other causes of similar symptoms include asthma, pneumonia, bronchiolitis, bronchiectasis, and COPD. A chest X-ray may be useful to detect pneumonia.
Prevention is by not smoking and avoiding other lung irritants. Frequent hand washing and flu vaccination may also be protective. Treatment of acute bronchitis typically involves rest, paracetamol (acetaminophen), and NSAIDs to help with the fever. Cough medicine has little support for its use and is not recommended in children less than six years of age. Antibiotics should generally not be used. An exception is when acute bronchitis is due to pertussis. Tentative evidence supports honey and pelargonium to help with symptoms.
Acute bronchitis is one of the most common diseases. About 5% of adults are affected and about 6% of children have at least one episode a year. It occurs more often in the winter. More than 10 million people in the United States visit a doctor each year for this conditi
Document 1:::
Pneumonia is an inflammatory condition of the lung primarily affecting the small air sacs known as alveoli. Symptoms typically include some combination of productive or dry cough, chest pain, fever, and difficulty breathing. The severity of the condition is variable.
Pneumonia is usually caused by infection with viruses or bacteria, and less commonly by other microorganisms. Identifying the responsible pathogen can be difficult. Diagnosis is often based on symptoms and physical examination. Chest X-rays, blood tests, and culture of the sputum may help confirm the diagnosis. The disease may be classified by where it was acquired, such as community- or hospital-acquired or healthcare-associated pneumonia.
Risk factors for pneumonia include cystic fibrosis, chronic obstructive pulmonary disease (COPD), sickle cell disease, asthma, diabetes, heart failure, a history of smoking, a poor ability to cough (such as following a stroke), and a weak immune system.
Vaccines to prevent certain types of pneumonia (such as those caused by Streptococcus pneumoniae bacteria, linked to influenza, or linked to COVID-19) are available. Other methods of prevention include hand washing to prevent infection, not smoking, and social distancing.
Treatment depends on the underlying cause. Pneumonia believed to be due to bacteria is treated with antibiotics. If the pneumonia is severe, the affected person is generally hospitalized. Oxygen therapy may be used if oxygen levels are low.
Each year, pneumonia affects about 450 million people globally (7% of the population) and results in about 4 million deaths. With the introduction of antibiotics and vaccines in the 20th century, survival has greatly improved. Nevertheless, pneumonia remains a leading cause of death in developing countries, and also among the very old, the very young, and the chronically ill. Pneumonia often shortens the period of suffering among those already close to death and has thus been called "the old man's friend".
S
Document 2:::
Bronchiolitis is inflammation of the small airways in the lungs. Acute bronchiolitis is due to a viral infection usually affecting children younger than two years of age. Symptoms may include fever, cough, runny nose, wheezing, and breathing problems. More severe cases may be associated with nasal flaring, grunting, or the skin between the ribs pulling in with breathing. If the child has not been able to feed properly, signs of dehydration may be present.
Chronic bronchiolitis is the general term used for small airways disease in adults, notably in chronic obstructive pulmonary disease.
Acute bronchiolitis is usually the result of infection by respiratory syncytial virus (72% of cases) or human rhinovirus (26% of cases). Diagnosis is generally based on symptoms. Tests such as a chest X-ray or viral testing are not routinely needed.
There is no specific treatment. Symptomatic treatment at home is generally sufficient. Occasionally, hospital admission for oxygen, support with feeding, or intravenous fluids is required. Tentative evidence supports nebulized hypertonic saline. Evidence for antibiotics, antivirals, bronchodilators, or nebulized epinephrine is either unclear or not supportive.
About 10% to 30% of children under the age of two years are affected by bronchiolitis at some point in time. It commonly occurs in the winter in the Northern Hemisphere. It is the leading cause of hospitalizations in those less than one year of age in the United States. The risk of death among those who are admitted to hospital is about 1%. Outbreaks of the condition were first described in the 1940s.
Signs and symptoms
Bronchiolitis typically presents in children under two years old and is characterized by a constellation of respiratory symptoms that consists of fever, rhinorrhea, cough, wheeze, tachypnea and increased work of breathing such as nasal flaring or grunting that develops over one to three days. Crackles or wheeze are typical findings on listening to the chest wi
Document 3:::
Lung receptors sense irritation or inflammation in the bronchi and alveoli.
Document 4:::
Pulmonary pathology is the subspecialty of surgical pathology which deals with the diagnosis and characterization of neoplastic and non-neoplastic diseases of the lungs and thoracic pleura. Diagnostic specimens are often obtained via bronchoscopic transbronchial biopsy, CT-guided percutaneous biopsy, or video-assisted thoracic surgery (VATS). The diagnosis of inflammatory or fibrotic diseases of the lungs is considered by many pathologists to be particularly challenging.
Anatomical pathology
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What disease is described as a respiratory disorder characterized by wheezing, coughing, and a feeling of constriction in the chest?
A. asthma
B. indigestion
C. dementia
D. inflammation
Answer:
|
|
sciq-8835
|
multiple_choice
|
Organelles whose membranes are specialized for aerobic respiration are called what?
|
[
"chloroplasts",
"mitochondria",
"mitosis",
"vacuoles"
] |
B
|
Relavent Documents:
Document 0:::
Cell physiology is the biological study of the activities that take place in a cell to keep it alive. The term physiology refers to normal functions in a living organism. Animal cells, plant cells and microorganism cells show similarities in their functions even though they vary in structure.
General characteristics
There are two types of cells: prokaryotes and eukaryotes.
Prokaryotes were the first of the two to develop and do not have a self-contained nucleus. Their mechanisms are simpler than later-evolved eukaryotes, which contain a nucleus that envelops the cell's DNA and some organelles.
Prokaryotes
Prokaryotes have DNA located in an area called the nucleoid, which is not separated from other parts of the cell by a membrane. There are two domains of prokaryotes: bacteria and archaea. Prokaryotes have fewer organelles than eukaryotes. Both have plasma membranes and ribosomes (structures that synthesize proteins and float free in cytoplasm). Two unique characteristics of prokaryotes are fimbriae (finger-like projections on the surface of a cell) and flagella (threadlike structures that aid movement).
Eukaryotes
Eukaryotes have a nucleus where DNA is contained. They are usually larger than prokaryotes and contain many more organelles. The nucleus, the feature of a eukaryote that distinguishes it from a prokaryote, contains a nuclear envelope, nucleolus and chromatin. In cytoplasm, endoplasmic reticulum (ER) synthesizes membranes and performs other metabolic activities. There are two types, rough ER (containing ribosomes) and smooth ER (lacking ribosomes). The Golgi apparatus consists of multiple membranous sacs, responsible for manufacturing and shipping out materials such as proteins. Lysosomes are structures that use enzymes to break down substances through phagocytosis, a process that comprises endocytosis and exocytosis. In the mitochondria, metabolic processes such as cellular respiration occur. The cytoskeleton is made of fibers that support the str
Document 1:::
Chloroplasts contain several important membranes, vital for their function. Like mitochondria, chloroplasts have a double-membrane envelope, called the chloroplast envelope, but unlike mitochondria, chloroplasts also have internal membrane structures called thylakoids. Furthermore, one or two additional membranes may enclose chloroplasts in organisms that underwent secondary endosymbiosis, such as the euglenids and chlorarachniophytes.
The chloroplasts come via endosymbiosis by engulfment of a photosynthetic cyanobacterium by the eukaryotic, already mitochondriate cell. Over millions of years the endosymbiotic cyanobacterium evolved structurally and functionally, retaining its own DNA and the ability to divide by binary fission (not mitotically) but giving up its autonomy by the transfer of some of its genes to the nuclear genome.
Envelope membranes
Each of the envelope membranes is a lipid bilayer that is between 6 and 8 nm thick. The lipid composition of the outer membrane has been found to be 48% phospholipids, 46% galactolipids and 7% sulfolipids, while the inner membrane has been found to contain 16% phospholipids, 79% galactolipids and 5% sulfolipids in spinach chloroplasts.
The outer membrane is permeable to most ions and metabolites, but the inner membrane of the chloroplast is highly specialised with transport proteins. For example, carbohydrates are transported across the inner envelope membrane by a triose phosphate translocator. The two envelope membranes are separated by a gap of 10–20 nm, called the intermembrane space.
Thylakoid membrane
Within the envelope membranes, in the region called the stroma, there is a system of interconnecting flattened membrane compartments, called the thylakoids. The thylakoid membrane is quite similar in lipid composition to the inner envelope membrane, containing 78% galactolipids, 15.5% phospholipids and 6.5% sulfolipids in spinach chloroplasts. The thylakoid membrane encloses a single, continuous aqueous compartme
Document 2:::
Cellular compartments in cell biology comprise all of the closed parts within the cytosol of a eukaryotic cell, usually surrounded by a single or double lipid layer membrane. These compartments are often, but not always, defined as membrane-bound organelles. The formation of cellular compartments is called compartmentalization.
Both organelles, the mitochondria and chloroplasts (in photosynthetic organisms), are compartments that are believed to be of endosymbiotic origin. Other compartments such as peroxisomes, lysosomes, the endoplasmic reticulum, the cell nucleus or the Golgi apparatus are not of endosymbiotic origin. Smaller elements like vesicles, and sometimes even microtubules can also be counted as compartments.
It was thought that compartmentalization is not found in prokaryotic cells., but the discovery of carboxysomes and many other metabolosomes revealed that prokaryotic cells are capable of making compartmentalized structures, albeit these are in most cases not surrounded by a lipid bilayer, but of pure proteinaceous built.
Types
In general there are 4 main cellular compartments, they are:
The nuclear compartment comprising the nucleus
The intercisternal space which comprises the space between the membranes of the endoplasmic reticulum (which is continuous with the nuclear envelope)
Organelles (the mitochondrion in all eukaryotes and the plastid in phototrophic eukaryotes)
The cytosol
Function
Compartments have three main roles. One is to establish physical boundaries for biological processes that enables the cell to carry out different metabolic activities at the same time. This may include keeping certain biomolecules within a region, or keeping other molecules outside. Within the membrane-bound compartments, different intracellular pH, different enzyme systems, and other differences are isolated from other organelles and cytosol. With mitochondria, the cytosol has an oxidizing environment which converts NADH to NAD+. With these cases, the
Document 3:::
A protocell (or protobiont) is a self-organized, endogenously ordered, spherical collection of lipids proposed as a stepping stone toward the origin of life. A central question in evolution is how simple protocells first arose and how they could differ in reproductive output, thus enabling the accumulation of novel biological emergences over time, i.e. biological evolution. Although a functional protocell has not yet been achieved in a laboratory setting, the goal to understand the process appears well within reach.
Overview
Compartmentalization was important in the origins of life. Membranes form enclosed compartments that are separate from the external environment, thus providing the cell with functionally specialized aqueous spaces. As the lipid bilayer of membranes is impermeable to most hydrophilic molecules (dissolved by water), cells have membrane transport-systems that achieve the import of nutritive molecules as well as the export of waste. It is very challenging to construct protocells from molecular assemblies. An important step in this challenge is the achievement of vesicle dynamics that are relevant to cellular functions, such as membrane trafficking and self-reproduction, using amphiphilic molecules. On the primitive Earth, numerous chemical reactions of organic compounds produced the ingredients of life. Of these substances, amphiphilic molecules might be the first player in the evolution from molecular assembly to cellular life. A step from vesicle toward protocell might be to develop self-reproducing vesicles coupled with the metabolic system.
Another approach to the notion of a protocell concerns the term "chemoton" (short for 'chemical automaton') which refers to an abstract model for the fundamental unit of life introduced by Hungarian theoretical biologist Tibor Gánti. It is the oldest known computational abstract of a protocell. Gánti conceived the basic idea in 1952 and formulated the concept in 1971 in his book The Principles of Life (orig
Document 4:::
Paracellular transport refers to the transfer of substances across an epithelium by passing through the intercellular space between the cells. It is in contrast to transcellular transport, where the substances travel through the cell, passing through both the apical membrane and basolateral membrane.
The distinction has particular significance in renal physiology and intestinal physiology. Transcellular transport often involves energy expenditure whereas paracellular transport is unmediated and passive down a concentration gradient, or by osmosis (for water) and solvent drag for solutes. Paracellular transport also has the benefit that absorption rate is matched to load because it has no transporters that can be saturated.
In most mammals, intestinal absorption of nutrients is thought to be dominated by transcellular transport, e.g., glucose is primarily absorbed via the SGLT1 transporter and other glucose transporters. Paracellular absorption therefore plays only a minor role in glucose absorption, although there is evidence that paracellular pathways become more available when nutrients are present in the intestinal lumen. In contrast, small flying vertebrates (small birds and bats) rely on the paracellular pathway for the majority of glucose absorption in the intestine. This has been hypothesized to compensate for an evolutionary pressure to reduce mass in flying animals, which resulted in a reduction in intestine size and faster transit time of food through the gut.
Capillaries of the blood–brain barrier have only transcellular transport, in contrast with normal capillaries which have both transcellular and paracellular transport.
The paracellular pathway of transport is also important for the absorption of drugs in the gastrointestinal tract. The paracellular pathway allows the permeation of hydrophilic molecules that are not able to permeate through the lipid membrane by the transcellular pathway of absorption. This is particularly important for hydrophi
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Organelles whose membranes are specialized for aerobic respiration are called what?
A. chloroplasts
B. mitochondria
C. mitosis
D. vacuoles
Answer:
|
|
sciq-1210
|
multiple_choice
|
What forms the pathway of water and nutrients from roots to leaves and flower?
|
[
"the stem",
"the stamen",
"the stigma",
"the pistil"
] |
A
|
Relavent Documents:
Document 0:::
A stem is one of two main structural axes of a vascular plant, the other being the root. It supports leaves, flowers and fruits, transports water and dissolved substances between the roots and the shoots in the xylem and phloem, photosynthesis takes place here, stores nutrients, and produces new living tissue. The stem can also be called halm or haulm or culms.
The stem is normally divided into nodes and internodes:
The nodes are the points of attachment for leaves and can hold one or more leaves. There are sometimes axillary buds between the stem and leaf which can grow into branches (with leaves, conifer cones, or flowers). Adventitious roots may also be produced from the nodes. Vines may produce tendrils from nodes.
The internodes distance one node from another.
The term "shoots" is often confused with "stems"; "shoots" generally refers to new fresh plant growth, including both stems and other structures like leaves or flowers.
In most plants, stems are located above the soil surface, but some plants have underground stems.
Stems have several main functions:
Support for and the elevation of leaves, flowers, and fruits. The stems keep the leaves in the light and provide a place for the plant to keep its flowers and fruits.
Transport of fluids between the roots and the shoots in the xylem and phloem.
Storage of nutrients.
Production of new living tissue. The normal lifespan of plant cells is one to three years. Stems have cells called meristems that annually generate new living tissue.
Photosynthesis.
Stems have two pipe-like tissues called xylem and phloem. The xylem tissue arises from the cell facing inside and transports water by the action of transpiration pull, capillary action, and root pressure. The phloem tissue arises from the cell facing outside and consists of sieve tubes and their companion cells. The function of phloem tissue is to distribute food from photosynthetic tissue to other tissues. The two tissues are separated by cambium, a tis
Document 1:::
Transfer cells are specialized parenchyma cells that have an increased surface area, due to infoldings of the plasma membrane. They facilitate the transport of sugars from a sugar source, mainly mature leaves, to a sugar sink, often developing leaves or fruits. They are found in nectaries of flowers and some carnivorous plants.
Transfer cells are specially found in plants in the region of absorption or secretion of nutrients.
The term transfer cell was coined by Brian Gunning and John Stewart Pate. Their presence is generally correlated with the existence of extensive solute influxes across the plasma membrane.
Document 2:::
The soil-plant-atmosphere continuum (SPAC) is the pathway for water moving from soil through plants to the atmosphere. Continuum in the description highlights the continuous nature of water connection through the pathway. The low water potential of the atmosphere, and relatively higher (i.e. less negative) water potential inside leaves, leads to a diffusion gradient across the stomatal pores of leaves, drawing water out of the leaves as vapour. As water vapour transpires out of the leaf, further water molecules evaporate off the surface of mesophyll cells to replace the lost molecules since water in the air inside leaves is maintained at saturation vapour pressure. Water lost at the surface of cells is replaced by water from the xylem, which due to the cohesion-tension properties of water in the xylem of plants pulls additional water molecules through the xylem from the roots toward the leaf.
Components
The transport of water along this pathway occurs in components, variously defined among scientific disciplines:
Soil physics characterizes water in soil in terms of tension,
Physiology of plants and animals characterizes water in organisms in terms of diffusion pressure deficit, and
Meteorology uses vapour pressure or relative humidity to characterize atmospheric water.
SPAC integrates these components and is defined as a:
...concept recognising that the field with all its components (soil, plant, animals and the ambient atmosphere taken together) constitutes a physically integrated, dynamic system in which the various flow processes involving energy and matter occur simultaneously and independently like links in the chain.
This characterises the state of water in different components of the SPAC as expressions of the energy level or water potential of each. Modelling of water transport between components relies on SPAC, as do studies of water potential gradients between segments.
See also
Ecohydrology
Evapotranspiration
Hydraulic redistribution; a p
Document 3:::
Hydraulic redistribution is a passive mechanism where water is transported from moist to dry soils via subterranean networks. It occurs in vascular plants that commonly have roots in both wet and dry soils, especially plants with both taproots that grow vertically down to the water table, and lateral roots that sit close to the surface. In the late 1980s, there was a movement to understand the full extent of these subterranean networks. Since then it was found that vascular plants are assisted by fungal networks which grow on the root system to promote water redistribution.
Process
Hot, dry periods, when the surface soil dries out to the extent that the lateral roots exude whatever water they contain, will result in the death of such lateral roots unless the water is replaced. Similarly, under extremely wet conditions when lateral roots are inundated by flood waters, oxygen deprivation will also lead to root peril. In plants that exhibit hydraulic redistribution, there are xylem pathways from the taproots to the laterals, such that the absence or abundance of water at the laterals creates a pressure potential analogous to that of transpirational pull. In drought conditions, ground water is drawn up through the taproot to the laterals and exuded into the surface soil, replenishing that which was lost. Under flooding conditions, plant roots perform a similar function in the opposite direction.
Though often referred to as hydraulic lift, movement of water by the plant roots has been shown to occur in any direction. This phenomenon has been documented in over sixty plant species spanning a variety of plant types (from herbs and grasses to shrubs and trees) and over a range of environmental conditions (from the Kalahari Desert to the Amazon Rainforest).
Causes
The movement of this water can be explained by a water transport theory throughout a plant. This well-established water transport theory is called the cohesion-tension theory. In brief, it explains the movement
Document 4:::
In plants, the transpiration stream is the uninterrupted stream of water and solutes which is taken up by the roots and transported via the xylem to the leaves where it evaporates into the air/apoplast-interface of the substomatal cavity. It is driven by capillary action and in some plants by root pressure. The main driving factor is the difference in water potential between the soil and the substomatal cavity caused by transpiration.
Transpiration
Transpiration can be regulated through stomatal closure or opening. It allows for plants to efficiently transport water up to their highest body organs, regulate the temperature of stem and leaves and it allows for upstream signaling such as the dispersal of an apoplastic alkalinization during local oxidative stress.
Summary of water movement:
Soil
Roots and Root Hair
Xylem
Leaves
Stomata
Air
Osmosis
The water passes from the soil to the root by osmosis. The long and thin shape of root hairs maximizes surface area so that more water can enter. There is greater water potential in the soil than in the cytoplasm of the root hair cells. As the cell's surface membrane of the root hair cell is semi-permeable, osmosis can take place; and water passes from the soil to the root hairs.
The next stage in the transpiration stream is water passing into the xylem vessels. The water either goes through the cortex cells (between the root cells and the xylem vessels) or it bypasses them – going through their cell walls.
After this, the water moves up the xylem vessels to the leaves through diffusion: A pressure change between the top and bottom of the vessel. Diffusion takes place because there is a water potential gradient between water in the xylem vessel and the leaf (as water is transpiring out of the leaf). This means that water diffuses up the leaf. There is also a pressure change between the top and bottom of the xylem vessels, due to water loss from the leaves. This reduces the pressure of water at the top of the vessels. T
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What forms the pathway of water and nutrients from roots to leaves and flower?
A. the stem
B. the stamen
C. the stigma
D. the pistil
Answer:
|
|
sciq-1377
|
multiple_choice
|
What carries the instructions from the nucleus to the cytoplasm?
|
[
"messenger rna",
"anderson rna",
"dirscriptor rna",
"concept rna"
] |
A
|
Relavent Documents:
Document 0:::
Genomic deoxyribonucleic acid (abbreviated as gDNA) is chromosomal DNA, in contrast to extra-chromosomal DNAs like plasmids. Most organisms have the same genomic DNA in every cell; however, only certain genes are active in each cell to allow for cell function and differentiation within the body.
The genome of an organism (encoded by the genomic DNA) is the (biological) information of heredity which is passed from one generation of organism to the next. That genome is transcribed to produce various RNAs, which are necessary for the function of the organism. Precursor mRNA (pre-mRNA) is transcribed by RNA polymerase II in the nucleus. pre-mRNA is then processed by splicing to remove introns, leaving the exons in the mature messenger RNA (mRNA). Additional processing includes the addition of a 5' cap and a poly(A) tail to the pre-mRNA. The mature mRNA may then be transported to the cytosol and translated by the ribosome into a protein. Other types of RNA include ribosomal RNA (rRNA) and transfer RNA (tRNA). These types are transcribed by RNA polymerase I and RNA polymerase III, respectively, and are essential for protein synthesis. However 5s rRNA is the only rRNA which is transcribed by RNA Polymerase III.
Document 1:::
Eukaryotic transcription is the elaborate process that eukaryotic cells use to copy genetic information stored in DNA into units of transportable complementary RNA replica. Gene transcription occurs in both eukaryotic and prokaryotic cells. Unlike prokaryotic RNA polymerase that initiates the transcription of all different types of RNA, RNA polymerase in eukaryotes (including humans) comes in three variations, each translating a different type of gene. A eukaryotic cell has a nucleus that separates the processes of transcription and translation. Eukaryotic transcription occurs within the nucleus where DNA is packaged into nucleosomes and higher order chromatin structures. The complexity of the eukaryotic genome necessitates a great variety and complexity of gene expression control.
Eukaryotic transcription proceeds in three sequential stages: initiation, elongation, and termination.
The RNAs transcribed serve diverse functions. For example, structural components of the ribosome are transcribed by RNA polymerase I. Protein coding genes are transcribed by RNA polymerase II into messenger RNAs (mRNAs) that carry the information from DNA to the site of protein synthesis. More abundantly made are the so-called non-coding RNAs account for the large majority of the transcriptional output of a cell. These non-coding RNAs perform a variety of important cellular functions.
RNA polymerase
Eukaryotes have three nuclear RNA polymerases, each with distinct roles and properties.
RNA polymerase I (Pol I) catalyses the transcription of all rRNA genes except 5S. These rRNA genes are organised into a single transcriptional unit and are transcribed into a continuous transcript. This precursor is then processed into three rRNAs: 18S, 5.8S, and 28S. The transcription of rRNA genes takes place in a specialised structure of the nucleus called the nucleolus, where the transcribed rRNAs are combined with proteins to form ribosomes.
RNA polymerase II (Pol II) is responsible for the tra
Document 2:::
This lecture, named in memory of Keith R. Porter, is presented to an eminent cell biologist each year at the ASCB Annual Meeting. The ASCB Program Committee and the ASCB President recommend the Porter Lecturer to the Porter Endowment each year.
Lecturers
Source: ASCB
See also
List of biology awards
Document 3:::
Cell physiology is the biological study of the activities that take place in a cell to keep it alive. The term physiology refers to normal functions in a living organism. Animal cells, plant cells and microorganism cells show similarities in their functions even though they vary in structure.
General characteristics
There are two types of cells: prokaryotes and eukaryotes.
Prokaryotes were the first of the two to develop and do not have a self-contained nucleus. Their mechanisms are simpler than later-evolved eukaryotes, which contain a nucleus that envelops the cell's DNA and some organelles.
Prokaryotes
Prokaryotes have DNA located in an area called the nucleoid, which is not separated from other parts of the cell by a membrane. There are two domains of prokaryotes: bacteria and archaea. Prokaryotes have fewer organelles than eukaryotes. Both have plasma membranes and ribosomes (structures that synthesize proteins and float free in cytoplasm). Two unique characteristics of prokaryotes are fimbriae (finger-like projections on the surface of a cell) and flagella (threadlike structures that aid movement).
Eukaryotes
Eukaryotes have a nucleus where DNA is contained. They are usually larger than prokaryotes and contain many more organelles. The nucleus, the feature of a eukaryote that distinguishes it from a prokaryote, contains a nuclear envelope, nucleolus and chromatin. In cytoplasm, endoplasmic reticulum (ER) synthesizes membranes and performs other metabolic activities. There are two types, rough ER (containing ribosomes) and smooth ER (lacking ribosomes). The Golgi apparatus consists of multiple membranous sacs, responsible for manufacturing and shipping out materials such as proteins. Lysosomes are structures that use enzymes to break down substances through phagocytosis, a process that comprises endocytosis and exocytosis. In the mitochondria, metabolic processes such as cellular respiration occur. The cytoskeleton is made of fibers that support the str
Document 4:::
The D arm is a feature in the tertiary structure of transfer RNA (tRNA). It is composed of the two D stems and the D loop. The D loop contains the base dihydrouridine, for which the arm is named. The D loop's main function is that of recognition. It is widely believed that it acts as a recognition site for aminoacyl-tRNA synthetase, an enzyme involved in the aminoacylation of the tRNA molecule. The D stem is also believed to have a recognition role although this has yet to be verified.
It is a highly variable region and is notable for its unusual conformation due to the over-crowding on one of the guanosine residues. It appears to play a large role in the stabilization of the tRNA's tertiary structure.
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What carries the instructions from the nucleus to the cytoplasm?
A. messenger rna
B. anderson rna
C. dirscriptor rna
D. concept rna
Answer:
|
|
sciq-6342
|
multiple_choice
|
During which phase do sister chromatids separate and the centromeres divide?
|
[
"gap cycle",
"passivation",
"cell phase",
"anaphase"
] |
D
|
Relavent Documents:
Document 0:::
Chromosome segregation is the process in eukaryotes by which two sister chromatids formed as a consequence of DNA replication, or paired homologous chromosomes, separate from each other and migrate to opposite poles of the nucleus. This segregation process occurs during both mitosis and meiosis. Chromosome segregation also occurs in prokaryotes. However, in contrast to eukaryotic chromosome segregation, replication and segregation are not temporally separated. Instead segregation occurs progressively following replication.
Mitotic chromatid segregation
During mitosis chromosome segregation occurs routinely as a step in cell division (see mitosis diagram). As indicated in the mitosis diagram, mitosis is preceded by a round of DNA replication, so that each chromosome forms two copies called chromatids. These chromatids separate to opposite poles, a process facilitated by a protein complex referred to as cohesin. Upon proper segregation, a complete set of chromatids ends up in each of two nuclei, and when cell division is completed, each DNA copy previously referred to as a chromatid is now called a chromosome.
Meiotic chromosome and chromatid segregation
Chromosome segregation occurs at two separate stages during meiosis called anaphase I and anaphase II (see meiosis diagram). In a diploid cell there are two sets of homologous chromosomes of different parental origin (e.g. a paternal and a maternal set). During the phase of meiosis labeled “interphase s” in the meiosis diagram there is a round of DNA replication, so that each of the chromosomes initially present is now composed of two copies called chromatids. These chromosomes (paired chromatids) then pair with the homologous chromosome (also paired chromatids) present in the same nucleus (see prophase I in the meiosis diagram). The process of alignment of paired homologous chromosomes is called synapsis (see Synapsis). During synapsis, genetic recombination usually occurs. Some of the recombination even
Document 1:::
A kinetochore (, ) is a disc-shaped protein structure associated with duplicated chromatids in eukaryotic cells where the spindle fibers attach during cell division to pull sister chromatids apart. The kinetochore assembles on the centromere and links the chromosome to microtubule polymers from the mitotic spindle during mitosis and meiosis. The term kinetochore was first used in a footnote in a 1934 Cytology book by Lester W. Sharp and commonly accepted in 1936. Sharp's footnote reads: "The convenient term kinetochore (= movement place) has been suggested to the author by J. A. Moore", likely referring to John Alexander Moore who had joined Columbia University as a freshman in 1932.
Monocentric organisms, including vertebrates, fungi, and most plants, have a single centromeric region on each chromosome which assembles a single, localized kinetochore. Holocentric organisms, such as nematodes and some plants, assemble a kinetochore along the entire length of a chromosome.
Kinetochores start, control, and supervise the striking movements of chromosomes during cell division. During mitosis, which occurs after the amount of DNA is doubled in each chromosome (while maintaining the same number of chromosomes) in S phase, two sister chromatids are held together by a centromere. Each chromatid has its own kinetochore, which face in opposite directions and attach to opposite poles of the mitotic spindle apparatus. Following the transition from metaphase to anaphase, the sister chromatids separate from each other, and the individual kinetochores on each chromatid drive their movement to the spindle poles that will define the two new daughter cells. The kinetochore is therefore essential for the chromosome segregation that is classically associated with mitosis and meiosis.
Structure of Kinetochore
The kinetochore contains two regions:
an inner kinetochore, which is tightly associated with the centromere DNA and assembled in a specialized form of chromatin that persists t
Document 2:::
Sister chromatid cohesion refers to the process by which sister chromatids are paired and held together during certain phases of the cell cycle. Establishment of sister chromatid cohesion is the process by which chromatin-associated cohesin protein becomes competent to physically bind together the sister chromatids. In general, cohesion is established during S phase as DNA is replicated, and is lost when chromosomes segregate during mitosis and meiosis. Some studies have suggested that cohesion aids in aligning the kinetochores during mitosis by forcing the kinetochores to face opposite cell poles.
Cohesin loading
Cohesin first associates with the chromosomes during G1 phase. The cohesin ring is composed of two SMC (structural maintenance of chromosomes) proteins and two additional Scc proteins. Cohesin may originally interact with chromosomes via the ATPase domains of the SMC proteins. In yeast, the loading of cohesin on the chromosomes depends on proteins Scc2 and Scc4.
Cohesin interacts with the chromatin at specific loci. High levels of cohesin binding are observed at the centromere. Cohesin is also loaded at cohesin attachment regions (CARs) along the length of the chromosomes. CARs are approximately 500-800 base pair regions spaced at approximately 9 kilobase intervals along the chromosomes. In yeast, CARs tend to be rich in adenine-thymine base pairs. CARs are independent of origins of replication.
Establishment of cohesion
Establishment of cohesion refers to the process by which chromatin-associated cohesin becomes cohesion-competent. Chromatin association of cohesin is not sufficient for cohesion. Cohesin must undergo subsequent modification ("establishment") to be capable of physically holding the sister chromosomes together. Though cohesin can associate with chromatin earlier in the cell cycle, cohesion is established during S phase. Early data suggesting that S phase is crucial to cohesion was based on the fact that after S phase, sister chromatids
Document 3:::
Neocentromeres are new centromeres that form at a place on the chromosome that is usually not centromeric. They typically arise due to disruption of the normal centromere. These neocentromeres should not be confused with “knobs”, which were also described as “neocentromeres” in maize in the 1950s. Unlike most normal centromeres, neocentromeres do not contain satellite sequences that are highly repetitive but instead consist of unique sequences. Despite this, most neocentromeres are still able to carry out the functions of normal centromeres in regulating chromosome segregation and inheritance. This raises many questions on what is necessary versus what is sufficient for constituting a centromere.
As neocentromeres are still a relatively new phenomenon in cell biology and genetics, it may be useful to keep in mind that neocentromeres may be somewhat related to point centromeres, holocentromeres, and regional centromeres. Whereas point centromeres are defined by sequence, regional and holocentromeres are epigenetically defined by where a specific type of nucleosome (the one containing the centromeric histone H3) is located.
It may also be analytically helpful to take into account that the centromere is generally defined in relation to the kinetochore, specifically as the “part of the chromosome that links two sister chromatids together via the kinetochore”. However, the emergence of research in neocentromeres troubles this conventional definition and questions the function of a centromere beyond being a “landing pad” for kinetochore formation. This expands the scope of the centromere's function to include regulating the function of the kinetochore and the mitotic spindle.
History
Neocentromeres were discovered relatively recently. They were first observed by Andy Choo in a human karyotype clinic case in 1997, using fluorescent in situ hybridization (FISH) and cytogenetic analysis. The neocentromeres were observed on chromosome 10 of a patient, who was a child with
Document 4:::
An isochromosome is an unbalanced structural abnormality in which the arms of the chromosome are mirror images of each other. The chromosome consists of two copies of either the long (q) arm or the short (p) arm because isochromosome formation is equivalent to a simultaneous duplication and deletion of genetic material. Consequently, there is partial trisomy of the genes present in the isochromosome and partial monosomy of the genes in the lost arm.
Nomenclature
An isochromosome can be abbreviated as i(chromosome number)(centromeric breakpoint). For example, an isochromosome of chromosome 17 containing two q arms can be identified as i(17)(q10).(Medulloblastoma)
Mechanism
Isochromosomes can be created during mitosis and meiosis through a misdivision of the centromere or U-type strand exchange.
Centromere misdivision
Under normal separation of sister chromatids in anaphase, the centromere will divide longitudinally, or parallel to the long axis of the chromosome. An isochromosome is created when the centromere is divided transversely, or perpendicular to the long axis of the chromosome. The division is usually not occurring in the centromere itself, but in an area surrounding the centromere, also known as a pericentric region. It is proposed that these sites of exchange contain homologous sequences between sister chromatids. Although the resulting chromosome may appear monocentric with only one centromere, it is isodicentric with two centromeres very close to each other; resulting in a potential loss of genetic material found on the other arms. Misdivision of the centromere can also produce monocentric isochromosomes, but they are not as common as dicentric isochromosomes.
U-type strand exchange
A more common mechanism in the formation of isochromosomes is through the breakage and fusion of sister chromatids, most likely occurring in early anaphase of mitosis or meiosis. A double-stranded break in the pericentric region of the chromosome is repaired when the sist
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
During which phase do sister chromatids separate and the centromeres divide?
A. gap cycle
B. passivation
C. cell phase
D. anaphase
Answer:
|
|
sciq-7916
|
multiple_choice
|
What forms when there is a difference in temperature between the land and the air?
|
[
"smoke",
"fog",
"tsunamis",
"sandstorms"
] |
B
|
Relavent Documents:
Document 0:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 1:::
This is a list of meteorology topics. The terms relate to meteorology, the interdisciplinary scientific study of the atmosphere that focuses on weather processes and forecasting. (see also: List of meteorological phenomena)
A
advection
aeroacoustics
aerobiology
aerography (meteorology)
aerology
air parcel (in meteorology)
air quality index (AQI)
airshed (in meteorology)
American Geophysical Union (AGU)
American Meteorological Society (AMS)
anabatic wind
anemometer
annular hurricane
anticyclone (in meteorology)
apparent wind
Atlantic Oceanographic and Meteorological Laboratory (AOML)
Atlantic hurricane season
atmometer
atmosphere
Atmospheric Model Intercomparison Project (AMIP)
Atmospheric Radiation Measurement (ARM)
(atmospheric boundary layer [ABL]) planetary boundary layer (PBL)
atmospheric chemistry
atmospheric circulation
atmospheric convection
atmospheric dispersion modeling
atmospheric electricity
atmospheric icing
atmospheric physics
atmospheric pressure
atmospheric sciences
atmospheric stratification
atmospheric thermodynamics
atmospheric window (see under Threats)
B
ball lightning
balloon (aircraft)
baroclinity
barotropity
barometer ("to measure atmospheric pressure")
berg wind
biometeorology
blizzard
bomb (meteorology)
buoyancy
Bureau of Meteorology (in Australia)
C
Canada Weather Extremes
Canadian Hurricane Centre (CHC)
Cape Verde-type hurricane
capping inversion (in meteorology) (see "severe thunderstorms" in paragraph 5)
carbon cycle
carbon fixation
carbon flux
carbon monoxide (see under Atmospheric presence)
ceiling balloon ("to determine the height of the base of clouds above ground level")
ceilometer ("to determine the height of a cloud base")
celestial coordinate system
celestial equator
celestial horizon (rational horizon)
celestial navigation (astronavigation)
celestial pole
Celsius
Center for Analysis and Prediction of Storms (CAPS) (in Oklahoma in the US)
Center for the Study o
Document 2:::
In atmospheric science, equivalent temperature is the temperature of air in a parcel from which all the water vapor has been extracted by an adiabatic process.
Air contains water vapor that has been evaporated into it from liquid sources (lakes, sea, etc...). The energy needed to do that has been taken from the air. Taking a volume of air at temperature and mixing ratio of , drying it by condensation will restore energy to the airmass. This will depend on the latent heat release as:
where:
: latent heat of evaporation (2400 kJ/kg at 25°C to 2600 kJ/kg at −40°C)
: specific heat at constant pressure for air (≈ 1004 J/(kg·K))
Tables exist for exact values of the last two coefficients.
See also
Wet-bulb temperature
Potential temperature
Atmospheric thermodynamics
Equivalent potential temperature
Bibliography
M Robitzsch, Aequivalenttemperatur und Aequivalentthemometer, Meteorologische Zeitschrift, 1928, pp. 313-315.
M K Yau and R.R. Rogers, Short Course in Cloud Physics, Third Edition, published by Butterworth-Heinemann, January 1, 1989, 304 pages.
J.V. Iribarne and W.L. Godson, Atmospheric Thermodynamics, published by D. Reidel Publishing Company, Dordrecht, Holland, 1973, 222 pages
Atmospheric thermodynamics
Atmospheric temperature
Meteorological quantities
Document 3:::
In atmospheric science, an atmospheric model is a mathematical model constructed around the full set of primitive, dynamical equations which govern atmospheric motions. It can supplement these equations with parameterizations for turbulent diffusion, radiation, moist processes (clouds and precipitation), heat exchange, soil, vegetation, surface water, the kinematic effects of terrain, and convection. Most atmospheric models are numerical, i.e. they discretize equations of motion. They can predict microscale phenomena such as tornadoes and boundary layer eddies, sub-microscale turbulent flow over buildings, as well as synoptic and global flows. The horizontal domain of a model is either global, covering the entire Earth, or regional (limited-area), covering only part of the Earth. The different types of models run are thermotropic, barotropic, hydrostatic, and nonhydrostatic. Some of the model types make assumptions about the atmosphere which lengthens the time steps used and increases computational speed.
Forecasts are computed using mathematical equations for the physics and dynamics of the atmosphere. These equations are nonlinear and are impossible to solve exactly. Therefore, numerical methods obtain approximate solutions. Different models use different solution methods. Global models often use spectral methods for the horizontal dimensions and finite-difference methods for the vertical dimension, while regional models usually use finite-difference methods in all three dimensions. For specific locations, model output statistics use climate information, output from numerical weather prediction, and current surface weather observations to develop statistical relationships which account for model bias and resolution issues.
Types
The main assumption made by the thermotropic model is that while the magnitude of the thermal wind may change, its direction does not change with respect to height, and thus the baroclinicity in the atmosphere can be simulated usi
Document 4:::
An urban thermal plume describes rising air in the lower altitudes of the Earth's atmosphere caused by urban areas being warmer than surrounding areas. Over the past thirty years there has been increasing interest in what have been called urban heat islands (UHI), but it is only since 2007 that thought has been given to the rising columns of warm air, or ‘thermal plumes’ that they produce. Common on-shore breezes at the seaside on a warm day, and off-shore breezes at night are caused by the land heating up faster on a sunny day and cooling faster after sunset, respectively. Thermals, or warm airs, that rise from the land and sea affect the local microscale meteorology; and perhaps at times the mesometeorology. Urban thermal plumes have as powerful although less localized an effect.
London is generally 3 to 9 Celsius hotter than the Home Counties. London’s meteorological aberrations were first studied by Luke Howard, FRS in the 1810s, but the notion that this large warm area would produce a significant urban thermal plume was not seriously proposed until very recently.
Microscale thermal plumes, whose diameters may be measured in tens of metres, such as those produced by industrial chimney stacks, have been extensively investigated, but largely from the point of view of the plumes dispersal by local micrometeorology. Though their velocity is generally less, their very much greater magnitude (diameter) means that urban thermal plumes will have a more significant effect upon the mesometeorology and even continental macrometeorology.
Climate change
Decreasing Arctic sea ice cover is one of the most visible manifestations of climate change, often linked to rising global temperatures. However, there are several reports that shrinking polar ice is due more to changes in ambient wind direction than to increasing environmental temperatures per se.
In 2006-07, a team led by Son Nghiem of NASA Jet Propulsion Laboratory, Pasadena, California, studied trends in Arctic perenn
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What forms when there is a difference in temperature between the land and the air?
A. smoke
B. fog
C. tsunamis
D. sandstorms
Answer:
|
|
sciq-6252
|
multiple_choice
|
What are the membrane-bound organelles that are usually larger than vesicles and can have secretory, excretory, and storage functions?
|
[
"vacuoles",
"tubules",
"ribosomes",
"nucleolus"
] |
A
|
Relavent Documents:
Document 0:::
Cell physiology is the biological study of the activities that take place in a cell to keep it alive. The term physiology refers to normal functions in a living organism. Animal cells, plant cells and microorganism cells show similarities in their functions even though they vary in structure.
General characteristics
There are two types of cells: prokaryotes and eukaryotes.
Prokaryotes were the first of the two to develop and do not have a self-contained nucleus. Their mechanisms are simpler than later-evolved eukaryotes, which contain a nucleus that envelops the cell's DNA and some organelles.
Prokaryotes
Prokaryotes have DNA located in an area called the nucleoid, which is not separated from other parts of the cell by a membrane. There are two domains of prokaryotes: bacteria and archaea. Prokaryotes have fewer organelles than eukaryotes. Both have plasma membranes and ribosomes (structures that synthesize proteins and float free in cytoplasm). Two unique characteristics of prokaryotes are fimbriae (finger-like projections on the surface of a cell) and flagella (threadlike structures that aid movement).
Eukaryotes
Eukaryotes have a nucleus where DNA is contained. They are usually larger than prokaryotes and contain many more organelles. The nucleus, the feature of a eukaryote that distinguishes it from a prokaryote, contains a nuclear envelope, nucleolus and chromatin. In cytoplasm, endoplasmic reticulum (ER) synthesizes membranes and performs other metabolic activities. There are two types, rough ER (containing ribosomes) and smooth ER (lacking ribosomes). The Golgi apparatus consists of multiple membranous sacs, responsible for manufacturing and shipping out materials such as proteins. Lysosomes are structures that use enzymes to break down substances through phagocytosis, a process that comprises endocytosis and exocytosis. In the mitochondria, metabolic processes such as cellular respiration occur. The cytoskeleton is made of fibers that support the str
Document 1:::
Paramural bodies are membranous or vesicular structures located between the cell walls and cell membranes of plant and fungal cells. When these are continuous with the cell wall, they are termed lomasomes, while they are referred to as plasmalemmasomes if associated with the plasmalemma.
Function
While their function has not yet been studied in great detail, it has been speculated that due to the morphological similarity of paramural bodies to the exosomes produced by mammalian cells, they may perform similar functions such as membrane vesicle trafficking between cells. Current evidence suggests that, like exosomes, paramural bodies are derived from multivesicular bodies.
See also
Exosome
Endosome
Golgi apparatus
Document 2:::
Endoplasm generally refers to the inner (often granulated), dense part of a cell's cytoplasm. This is opposed to the ectoplasm which is the outer (non-granulated) layer of the cytoplasm, which is typically watery and immediately adjacent to the plasma membrane. The nucleus is separated from the endoplasm by the nuclear envelope. The different makeups/viscosities of the endoplasm and ectoplasm contribute to the amoeba's locomotion through the formation of a pseudopod. However, other types of cells have cytoplasm divided into endo- and ectoplasm. The endoplasm, along with its granules, contains water, nucleic acids, amino acids, carbohydrates, inorganic ions, lipids, enzymes, and other molecular compounds. It is the site of most cellular processes as it houses the organelles that make up the endomembrane system, as well as those that stand alone. The endoplasm is necessary for most metabolic activities, including cell division.
The endoplasm, like the cytoplasm, is far from static. It is in a constant state of flux through intracellular transport, as vesicles are shuttled between organelles and to/from the plasma membrane. Materials are regularly both degraded and synthesized within the endoplasm based on the needs of the cell and/or organism. Some components of the cytoskeleton run throughout the endoplasm though most are concentrated in the ectoplasm - towards the cells edges, closer to the plasma membrane. The endoplasm's granules are suspended in cytosol.
Granules
The term granule refers to a small particle within the endoplasm, typically the secretory vesicles. The granule is the defining characteristic of the endoplasm, as they are typically not present within the ectoplasm. These offshoots of the endomembrane system are enclosed by a phospholipid bilayer and can fuse with other organelles as well as the plasma membrane. Their membrane is only semipermeable and allows them to house substances that could be harmful to the cell if they were allowed to flow fre
Document 3:::
Cellular compartments in cell biology comprise all of the closed parts within the cytosol of a eukaryotic cell, usually surrounded by a single or double lipid layer membrane. These compartments are often, but not always, defined as membrane-bound organelles. The formation of cellular compartments is called compartmentalization.
Both organelles, the mitochondria and chloroplasts (in photosynthetic organisms), are compartments that are believed to be of endosymbiotic origin. Other compartments such as peroxisomes, lysosomes, the endoplasmic reticulum, the cell nucleus or the Golgi apparatus are not of endosymbiotic origin. Smaller elements like vesicles, and sometimes even microtubules can also be counted as compartments.
It was thought that compartmentalization is not found in prokaryotic cells., but the discovery of carboxysomes and many other metabolosomes revealed that prokaryotic cells are capable of making compartmentalized structures, albeit these are in most cases not surrounded by a lipid bilayer, but of pure proteinaceous built.
Types
In general there are 4 main cellular compartments, they are:
The nuclear compartment comprising the nucleus
The intercisternal space which comprises the space between the membranes of the endoplasmic reticulum (which is continuous with the nuclear envelope)
Organelles (the mitochondrion in all eukaryotes and the plastid in phototrophic eukaryotes)
The cytosol
Function
Compartments have three main roles. One is to establish physical boundaries for biological processes that enables the cell to carry out different metabolic activities at the same time. This may include keeping certain biomolecules within a region, or keeping other molecules outside. Within the membrane-bound compartments, different intracellular pH, different enzyme systems, and other differences are isolated from other organelles and cytosol. With mitochondria, the cytosol has an oxidizing environment which converts NADH to NAD+. With these cases, the
Document 4:::
The food vacuole, or digestive vacuole, is an organelle found in simple eukaryotes such as protists. This organelle is essentially a lysosome. During the stage of the symbiont parasites' lifecycle where it resides within a human (or other mammalian) red blood cell, it is the site of haemoglobin digestion and the formation of the large haemozoin crystals that can be seen under a light microscope.
See also
Protists
Eukaryote
Amoeba
Lysosome
Enzymes
Euglenids
Paramecia
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What are the membrane-bound organelles that are usually larger than vesicles and can have secretory, excretory, and storage functions?
A. vacuoles
B. tubules
C. ribosomes
D. nucleolus
Answer:
|
|
sciq-5838
|
multiple_choice
|
This knocks electrons from atoms and turns them into ions?
|
[
"convection",
"chemical reactions",
"radiation",
"evaporation"
] |
C
|
Relavent Documents:
Document 0:::
Ionization (or ionisation) is the process by which an atom or a molecule acquires a negative or positive charge by gaining or losing electrons, often in conjunction with other chemical changes. The resulting electrically charged atom or molecule is called an ion. Ionization can result from the loss of an electron after collisions with subatomic particles, collisions with other atoms, molecules and ions, or through the interaction with electromagnetic radiation. Heterolytic bond cleavage and heterolytic substitution reactions can result in the formation of ion pairs. Ionization can occur through radioactive decay by the internal conversion process, in which an excited nucleus transfers its energy to one of the inner-shell electrons causing it to be ejected.
Uses
Everyday examples of gas ionization are such as within a fluorescent lamp or other electrical discharge lamps. It is also used in radiation detectors such as the Geiger-Müller counter or the ionization chamber. The ionization process is widely used in a variety of equipment in fundamental science (e.g., mass spectrometry) and in industry (e.g., radiation therapy). It is also widely used for air purification, though studies have shown harmful effects of this application.
Production of ions
Negatively charged ions are produced when a free electron collides with an atom and is subsequently trapped inside the electric potential barrier, releasing any excess energy. The process is known as electron capture ionization.
Positively charged ions are produced by transferring an amount of energy to a bound electron in a collision with charged particles (e.g. ions, electrons or positrons) or with photons. The threshold amount of the required energy is known as ionization potential. The study of such collisions is of fundamental importance with regard to the few-body problem, which is one of the major unsolved problems in physics. Kinematically complete experiments, i.e. experiments in which the complete momentum vect
Document 1:::
In physics, a charge carrier is a particle or quasiparticle that is free to move, carrying an electric charge, especially the particles that carry electric charges in electrical conductors. Examples are electrons, ions and holes. The term is used most commonly in solid state physics. In a conducting medium, an electric field can exert force on these free particles, causing a net motion of the particles through the medium; this is what constitutes an electric current.
The electron and the proton are the elementary charge carriers, each carrying one elementary charge (e), of the same magnitude and opposite sign.
In conductors
In conducting media, particles serve to carry charge:
In many metals, the charge carriers are electrons. One or two of the valence electrons from each atom are able to move about freely within the crystal structure of the metal. The free electrons are referred to as conduction electrons, and the cloud of free electrons is called a Fermi gas. Many metals have electron and hole bands. In some, the majority carriers are holes.
In electrolytes, such as salt water, the charge carriers are ions, which are atoms or molecules that have gained or lost electrons so they are electrically charged. Atoms that have gained electrons so they are negatively charged are called anions, atoms that have lost electrons so they are positively charged are called cations. Cations and anions of the dissociated liquid also serve as charge carriers in melted ionic solids (see e.g. the Hall–Héroult process for an example of electrolysis of a melted ionic solid). Proton conductors are electrolytic conductors employing positive hydrogen ions as carriers.
In a plasma, an electrically charged gas which is found in electric arcs through air, neon signs, and the sun and stars, the electrons and cations of ionized gas act as charge carriers.
In a vacuum, free electrons can act as charge carriers. In the electronic component known as the vacuum tube (also called valve), the mobil
Document 2:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 3:::
An ion () is an atom or molecule with a net electrical charge. The charge of an electron is considered to be negative by convention and this charge is equal and opposite to the charge of a proton, which is considered to be positive by convention. The net charge of an ion is not zero because its total number of electrons is unequal to its total number of protons.
A cation is a positively charged ion with fewer electrons than protons while an anion is a negatively charged ion with more electrons than protons. Opposite electric charges are pulled towards one another by electrostatic force, so cations and anions attract each other and readily form ionic compounds.
Ions consisting of only a single atom are termed atomic or monatomic ions, while two or more atoms form molecular ions or polyatomic ions. In the case of physical ionization in a fluid (gas or liquid), "ion pairs" are created by spontaneous molecule collisions, where each generated pair consists of a free electron and a positive ion. Ions are also created by chemical interactions, such as the dissolution of a salt in liquids, or by other means, such as passing a direct current through a conducting solution, dissolving an anode via ionization.
History of discovery
The word ion was coined from Greek neuter present participle of ienai (), meaning "to go". A cation is something that moves down ( pronounced kato, meaning "down") and an anion is something that moves up (, meaning "up"). They are so called because ions move toward the electrode of opposite charge. This term was introduced (after a suggestion by the English polymath William Whewell) by English physicist and chemist Michael Faraday in 1834 for the then-unknown species that goes from one electrode to the other through an aqueous medium. Faraday did not know the nature of these species, but he knew that since metals dissolved into and entered a solution at one electrode and new metal came forth from a solution at the other electrode; that some kind of
Document 4:::
Atomic physics is the field of physics that studies atoms as an isolated system of electrons and an atomic nucleus. Atomic physics typically refers to the study of atomic structure and the interaction between atoms. It is primarily concerned with the way in which electrons are arranged around the nucleus and
the processes by which these arrangements change. This comprises ions, neutral atoms and, unless otherwise stated, it can be assumed that the term atom includes ions.
The term atomic physics can be associated with nuclear power and nuclear weapons, due to the synonymous use of atomic and nuclear in standard English. Physicists distinguish between atomic physics—which deals with the atom as a system consisting of a nucleus and electrons—and nuclear physics, which studies nuclear reactions and special properties of atomic nuclei.
As with many scientific fields, strict delineation can be highly contrived and atomic physics is often considered in the wider context of atomic, molecular, and optical physics. Physics research groups are usually so classified.
Isolated atoms
Atomic physics primarily considers atoms in isolation. Atomic models will consist of a single nucleus that may be surrounded by one or more bound electrons. It is not concerned with the formation of molecules (although much of the physics is identical), nor does it examine atoms in a solid state as condensed matter. It is concerned with processes such as ionization and excitation by photons or collisions with atomic particles.
While modelling atoms in isolation may not seem realistic, if one considers atoms in a gas or plasma then the time-scales for atom-atom interactions are huge in comparison to the atomic processes that are generally considered. This means that the individual atoms can be treated as if each were in isolation, as the vast majority of the time they are. By this consideration, atomic physics provides the underlying theory in plasma physics and atmospheric physics, even though
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
This knocks electrons from atoms and turns them into ions?
A. convection
B. chemical reactions
C. radiation
D. evaporation
Answer:
|
|
sciq-1119
|
multiple_choice
|
What is the hardest of all minerals?
|
[
"gold",
"diamonds",
"platinum",
"titanium"
] |
B
|
Relavent Documents:
Document 0:::
See also
List of minerals
Document 1:::
Molybdenite is a mineral of molybdenum disulfide, MoS2. Similar in appearance and feel to graphite, molybdenite has a lubricating effect that is a consequence of its layered structure. The atomic structure consists of a sheet of molybdenum atoms sandwiched between sheets of sulfur atoms. The Mo-S bonds are strong, but the interaction between the sulfur atoms at the top and bottom of separate sandwich-like tri-layers is weak, resulting in easy slippage as well as cleavage planes.
Molybdenite crystallizes in the hexagonal crystal system as the common polytype 2H and also in the trigonal system as the 3R polytype.
Description
Occurrence
Molybdenite occurs in high temperature hydrothermal ore deposits.
Its associated minerals include pyrite, chalcopyrite, quartz, anhydrite, fluorite, and scheelite. Important deposits include the disseminated porphyry molybdenum deposits at Questa, New Mexico and the Henderson and Climax mines in Colorado. Molybdenite also occurs in porphyry copper deposits of Arizona, Utah, and Mexico.
The element rhenium is always present in molybdenite as a substitute for molybdenum, usually in the parts per million (ppm ) range, but often up to 1–2%. High rhenium content results in a structural variety detectable by X-ray diffraction techniques. Molybdenite ores are essentially the only source for rhenium. The presence of the radioactive isotope rhenium-187 and its daughter isotope osmium-187 provides a useful geochronologic dating technique.
Features
Molybdenite is extremely soft with a metallic luster, and is superficially almost identical to graphite, to the point where it is not possible to positively distinguish between the two minerals without scientific equipment. It marks paper in much the same way as graphite. Its distinguishing feature from graphite is its higher specific gravity, as well as its tendency to occur in a matrix.
Uses
Molybdenite is an important ore of molybdenum, and is the most common source of the metal. While
Document 2:::
Materials science has shaped the development of civilizations since the dawn of mankind. Better materials for tools and weapons has allowed mankind to spread and conquer, and advancements in material processing like steel and aluminum production continue to impact society today. Historians have regarded materials as such an important aspect of civilizations such that entire periods of time have defined by the predominant material used (Stone Age, Bronze Age, Iron Age). For most of recorded history, control of materials had been through alchemy or empirical means at best. The study and development of chemistry and physics assisted the study of materials, and eventually the interdisciplinary study of materials science emerged from the fusion of these studies. The history of materials science is the study of how different materials were used and developed through the history of Earth and how those materials affected the culture of the peoples of the Earth. The term "Silicon Age" is sometimes used to refer to the modern period of history during the late 20th to early 21st centuries.
Prehistory
In many cases, different cultures leave their materials as the only records; which anthropologists can use to define the existence of such cultures. The progressive use of more sophisticated materials allows archeologists to characterize and distinguish between peoples. This is partially due to the major material of use in a culture and to its associated benefits and drawbacks. Stone-Age cultures were limited by which rocks they could find locally and by which they could acquire by trading. The use of flint around 300,000 BCE is sometimes considered the beginning of the use of ceramics. The use of polished stone axes marks a significant advance, because a much wider variety of rocks could serve as tools.
The innovation of smelting and casting metals in the Bronze Age started to change the way that cultures developed and interacted with each other. Starting around 5,500 BCE,
Document 3:::
In geology, rock (or stone) is any naturally occurring solid mass or aggregate of minerals or mineraloid matter. It is categorized by the minerals included, its chemical composition, and the way in which it is formed. Rocks form the Earth's outer solid layer, the crust, and most of its interior, except for the liquid outer core and pockets of magma in the asthenosphere. The study of rocks involves multiple subdisciplines of geology, including petrology and mineralogy. It may be limited to rocks found on Earth, or it may include planetary geology that studies the rocks of other celestial objects.
Rocks are usually grouped into three main groups: igneous rocks, sedimentary rocks and metamorphic rocks. Igneous rocks are formed when magma cools in the Earth's crust, or lava cools on the ground surface or the seabed. Sedimentary rocks are formed by diagenesis and lithification of sediments, which in turn are formed by the weathering, transport, and deposition of existing rocks. Metamorphic rocks are formed when existing rocks are subjected to such high pressures and temperatures that they are transformed without significant melting.
Humanity has made use of rocks since the earliest humans. This early period, called the Stone Age, saw the development of many stone tools. Stone was then used as a major component in the construction of buildings and early infrastructure. Mining developed to extract rocks from the Earth and obtain the minerals within them, including metals. Modern technology has allowed the development of new man-made rocks and rock-like substances, such as concrete.
Study
Geology is the study of Earth and its components, including the study of rock formations. Petrology is the study of the character and origin of rocks. Mineralogy is the study of the mineral components that create rocks. The study of rocks and their components has contributed to the geological understanding of Earth's history, the archaeological understanding of human history, and the
Document 4:::
Mineral tests are several methods which can help identify the mineral type. This is used widely in mineralogy, hydrocarbon exploration and general mapping. There are over 4000 types of minerals known with each one with different sub-classes. Elements make minerals and minerals make rocks so actually testing minerals in the lab and in the field is essential to understand the history of the rock which aids data, zonation, metamorphic history, processes involved and other minerals.
The following tests are used on specimen and thin sections through polarizing microscope.
Color
Color of the mineral. This is not mineral specific. For example quartz can be almost any color, shape and within many rock types.
Streak
Color of the mineral's powder. This can be found by rubbing the mineral onto a concrete. This is more accurate but not always mineral specific.
Lustre
This is the way light reflects from the mineral's surface. A mineral can be metallic (shiny) or non-metallic (not shiny).
Transparency
The way light travels through minerals. The mineral can be transparent (clear), translucent (cloudy) or opaque (none).
Specific gravity
Ratio between the weight of the mineral relative to an equal volume of water.
Mineral habitat
The shape of the crystal and habitat.
Magnetism
Magnetic or nonmagnetic. Can be tested by using a magnet or a compass. This does not apply to all ion minerals (for example, pyrite).
Cleavage
Number, behaviour, size and way cracks fracture in the mineral.
UV fluorescence
Many minerals glow when put under a UV light.
Radioactivity
Is the mineral radioactive or non-radioactive? This is measured by a Geiger counter.
Taste
This is not recommended. Is the mineral salty, bitter or does it have no taste?
Bite Test
This is not recommended. This involves biting a mineral to see if its generally soft or hard. This was used in early gold exploration to tell the difference between pyrite (fools gold, hard) and gold (soft).
Hardness
The Mohs Hardn
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is the hardest of all minerals?
A. gold
B. diamonds
C. platinum
D. titanium
Answer:
|
|
sciq-52
|
multiple_choice
|
In the absence of air resistance, all falling objects accelerate at the same rate due to what force?
|
[
"velocity",
"gravity",
"motion",
"weight"
] |
B
|
Relavent Documents:
Document 0:::
The gravity of Earth, denoted by , is the net acceleration that is imparted to objects due to the combined effect of gravitation (from mass distribution within Earth) and the centrifugal force (from the Earth's rotation).
It is a vector quantity, whose direction coincides with a plumb bob and strength or magnitude is given by the norm .
In SI units this acceleration is expressed in metres per second squared (in symbols, m/s2 or m·s−2) or equivalently in newtons per kilogram (N/kg or N·kg−1). Near Earth's surface, the acceleration due to gravity, accurate to 2 significant figures, is . This means that, ignoring the effects of air resistance, the speed of an object falling freely will increase by about per second every second. This quantity is sometimes referred to informally as little (in contrast, the gravitational constant is referred to as big ).
The precise strength of Earth's gravity varies with location. The agreed upon value for is by definition. This quantity is denoted variously as , (though this sometimes means the normal gravity at the equator, ), , or simply (which is also used for the variable local value).
The weight of an object on Earth's surface is the downwards force on that object, given by Newton's second law of motion, or (). Gravitational acceleration contributes to the total gravity acceleration, but other factors, such as the rotation of Earth, also contribute, and, therefore, affect the weight of the object. Gravity does not normally include the gravitational pull of the Moon and Sun, which are accounted for in terms of tidal effects.
Variation in magnitude
A non-rotating perfect sphere of uniform mass density, or whose density varies solely with distance from the centre (spherical symmetry), would produce a gravitational field of uniform magnitude at all points on its surface. The Earth is rotating and is also not spherically symmetric; rather, it is slightly flatter at the poles while bulging at the Equator: an oblate spheroid.
Document 1:::
Velocity is the speed in combination with the direction of motion of an object. Velocity is a fundamental concept in kinematics, the branch of classical mechanics that describes the motion of bodies.
Velocity is a physical vector quantity: both magnitude and direction are needed to define it. The scalar absolute value (magnitude) of velocity is called , being a coherent derived unit whose quantity is measured in the SI (metric system) as metres per second (m/s or m⋅s−1). For example, "5 metres per second" is a scalar, whereas "5 metres per second east" is a vector. If there is a change in speed, direction or both, then the object is said to be undergoing an acceleration.
Constant velocity vs acceleration
To have a constant velocity, an object must have a constant speed in a constant direction. Constant direction constrains the object to motion in a straight path thus, a constant velocity means motion in a straight line at a constant speed.
For example, a car moving at a constant 20 kilometres per hour in a circular path has a constant speed, but does not have a constant velocity because its direction changes. Hence, the car is considered to be undergoing an acceleration.
Difference between speed and velocity
While the terms speed and velocity are often colloquially used interchangeably to connote how fast an object is moving, in scientific terms they are different. Speed, the scalar magnitude of a velocity vector, denotes only how fast an object is moving, while velocity indicates both an objects speed and direction.
Equation of motion
Average velocity
Velocity is defined as the rate of change of position with respect to time, which may also be referred to as the instantaneous velocity to emphasize the distinction from the average velocity. In some applications the average velocity of an object might be needed, that is to say, the constant velocity that would provide the same resultant displacement as a variable velocity in the same time interval, , over some
Document 2:::
As described by the third of Newton's laws of motion of classical mechanics, all forces occur in pairs such that if one object exerts a force on another object, then the second object exerts an equal and opposite reaction force on the first. The third law is also more generally stated as: "To every action there is always opposed an equal reaction: or the mutual actions of two bodies upon each other are always equal, and directed to contrary parts." The attribution of which of the two forces is the action and which is the reaction is arbitrary. Either of the two can be considered the action, while the other is its associated reaction.
Examples
Interaction with ground
When something is exerting force on the ground, the ground will push back with equal force in the opposite direction. In certain fields of applied physics, such as biomechanics, this force by the ground is called 'ground reaction force'; the force by the object on the ground is viewed as the 'action'.
When someone wants to jump, he or she exerts additional downward force on the ground ('action'). Simultaneously, the ground exerts upward force on the person ('reaction'). If this upward force is greater than the person's weight, this will result in upward acceleration. When these forces are perpendicular to the ground, they are also called a normal force.
Likewise, the spinning wheels of a vehicle attempt to slide backward across the ground. If the ground is not too slippery, this results in a pair of friction forces: the 'action' by the wheel on the ground in backward direction, and the 'reaction' by the ground on the wheel in forward direction. This forward force propels the vehicle.
Gravitational forces
The Earth, among other planets, orbits the Sun because the Sun exerts a gravitational pull that acts as a centripetal force, holding the Earth to it, which would otherwise go shooting off into space. If the Sun's pull is considered an action, then Earth simultaneously exerts a reaction as a gravi
Document 3:::
Specific force (SF) is a mass-specific quantity defined as the quotient of force per unit mass.
It is a physical quantity of kind acceleration, with dimension of length per time squared and units of metre per second squared (m·s−2).
It is normally applied to forces other than gravity, to emulate the relationship between gravitational acceleration and gravitational force.
It can also be called mass-specific weight (weight per unit mass), as the weight of an object is equal to the magnitude of the gravity force acting on it.
The g-force is an instance of specific force measured in units of the standard gravity (g) instead of m/s², i.e., in multiples of g (e.g., "3 g").
Type of acceleration
The (mass-)specific force is not a coordinate acceleration, but rather a proper acceleration, which is the acceleration relative to free-fall. Forces, specific forces, and proper accelerations are the same in all reference frames, but coordinate accelerations are frame-dependent. For free bodies, the specific force is the cause of, and a measure of, the body's proper acceleration.
The acceleration of an object free falling towards the earth depends on the reference frame (it disappears in the free-fall frame, also called the inertial frame), but any g-force "acceleration" will be present in all frames. This specific force is zero for freely-falling objects, since gravity acting alone does not produce g-forces or specific forces.
Accelerometers on the surface of the Earth measure a constant 9.8 m/s^2 even when they are not accelerating (that is, when they do not undergo coordinate acceleration). This is because accelerometers measure the proper acceleration produced by the g-force exerted by the ground (gravity acting alone never produces g-force or specific force). Accelerometers measure specific force (proper acceleration), which is the acceleration relative to free-fall, not the "standard" acceleration that is relative to a coordinate system.
Hydraulics
In open channel hydr
Document 4:::
In common usage, the mass of an object is often referred to as its weight, though these are in fact different concepts and quantities. Nevertheless, one object will always weigh more than another with less mass if both are subject to the same gravity (i.e. the same gravitational field strength).
In scientific contexts, mass is the amount of "matter" in an object (though "matter" may be difficult to define), but weight is the force exerted on an object's matter by gravity. At the Earth's surface, an object whose mass is exactly one kilogram weighs approximately 9.81 newtons, the product of its mass and the gravitational field strength there. The object's weight is less on Mars, where gravity is weaker; more on Saturn, where gravity is stronger; and very small in space, far from significant sources of gravity, but it always has the same mass.
Material objects at the surface of the Earth have weight despite such sometimes being difficult to measure. An object floating freely on water, for example, does not appear to have weight since it is buoyed by the water. But its weight can be measured if it is added to water in a container which is entirely supported by and weighed on a scale. Thus, the "weightless object" floating in water actually transfers its weight to the bottom of the container (where the pressure increases). Similarly, a balloon has mass but may appear to have no weight or even negative weight, due to buoyancy in air. However the weight of the balloon and the gas inside it has merely been transferred to a large area of the Earth's surface, making the weight difficult to measure. The weight of a flying airplane is similarly distributed to the ground, but does not disappear. If the airplane is in level flight, the same weight-force is distributed to the surface of the Earth as when the plane was on the runway, but spread over a larger area.
A better scientific definition of mass is its description as being a measure of inertia, which is the tendency of an
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
In the absence of air resistance, all falling objects accelerate at the same rate due to what force?
A. velocity
B. gravity
C. motion
D. weight
Answer:
|
|
sciq-7670
|
multiple_choice
|
The structure of the human tail bone is called what?
|
[
"an artifact",
"a vital structure",
"a abnormal structure",
"a vestigial structure"
] |
D
|
Relavent Documents:
Document 0:::
The coccyx (: coccyges or coccyxes), commonly referred to as the tailbone, is the final segment of the vertebral column in all apes, and analogous structures in certain other mammals such as horses. In tailless primates (e.g. humans and other great apes) since Nacholapithecus (a Miocene hominoid), the coccyx is the remnant of a vestigial tail. In animals with bony tails, it is known as tailhead or dock, in bird anatomy as tailfan. It comprises three to five separate or fused coccygeal vertebrae below the sacrum, attached to the sacrum by a fibrocartilaginous joint, the sacrococcygeal symphysis, which permits limited movement between the sacrum and the coccyx.
Structure
The coccyx is formed of three, four or five rudimentary vertebrae. It articulates superiorly with the sacrum. In each of the first three segments may be traced a rudimentary body and articular and transverse processes; the last piece (sometimes the third) is a mere nodule of bone. The transverse processes are most prominent and noticeable on the first coccygeal segment. All the segments lack pedicles, laminae and spinous processes. The first segment is the largest; it resembles the lowest sacral vertebra, and often exists as a separate piece; the remaining ones diminish in size rostrally.
Most anatomy books incorrectly state that the coccyx is normally fused in adults. It has been shown that the coccyx may, in some people, consist of up to five separate bony segments, the most common configuration being two or three segments.
Surfaces
The anterior surface is slightly concave and marked with three transverse grooves which indicate the junctions of the different segments. It gives attachment to the anterior sacrococcygeal ligament and the levatores ani and supports part of the rectum. The posterior surface is convex, marked by transverse grooves similar to those on the anterior surface, and presents on either side a linear row of tubercles – the undeveloped articular processes of the coccygeal ve
Document 1:::
The human skeleton is the internal framework of the human body. It is composed of around 270 bones at birth – this total decreases to around 206 bones by adulthood after some bones get fused together. The bone mass in the skeleton makes up about 14% of the total body weight (ca. 10–11 kg for an average person) and reaches maximum mass between the ages of 25 and 30. The human skeleton can be divided into the axial skeleton and the appendicular skeleton. The axial skeleton is formed by the vertebral column, the rib cage, the skull and other associated bones. The appendicular skeleton, which is attached to the axial skeleton, is formed by the shoulder girdle, the pelvic girdle and the bones of the upper and lower limbs.
The human skeleton performs six major functions: support, movement, protection, production of blood cells, storage of minerals, and endocrine regulation.
The human skeleton is not as sexually dimorphic as that of many other primate species, but subtle differences between sexes in the morphology of the skull, dentition, long bones, and pelvis exist. In general, female skeletal elements tend to be smaller and less robust than corresponding male elements within a given population. The human female pelvis is also different from that of males in order to facilitate childbirth. Unlike most primates, human males do not have penile bones.
Divisions
Axial
The axial skeleton (80 bones) is formed by the vertebral column (32–34 bones; the number of the vertebrae differs from human to human as the lower 2 parts, sacral and coccygeal bone may vary in length), a part of the rib cage (12 pairs of ribs and the sternum), and the skull (22 bones and 7 associated bones).
The upright posture of humans is maintained by the axial skeleton, which transmits the weight from the head, the trunk, and the upper extremities down to the lower extremities at the hip joints. The bones of the spine are supported by many ligaments. The erector spinae muscles are also supporting an
Document 2:::
Work
He is an associate professor of anatomy, Department of Anatomy, Howard University College of Medicine (US). He was among the most cited/influential anatomists in 2019.
Books
Single author or co-author books
DIOGO, R. (2021). Meaning of Life, Human Nature and Delusions - How Tales about Love, Sex, Races, Gods and Progress Affect Us and Earth's Splendor. Springer (New York, US).
MONTERO, R., ADESOMO, A. & R. DIOGO (2021). On viruses, pandemics, and us: a developing story [De virus, pandemias y nosotros: una historia en desarollo]. Independently published, Tucuman, Argentina. 495 pages.
DIOGO, R., J. ZIERMANN, J. MOLNAR, N. SIOMAVA & V. ABDALA (2018). Muscles of Chordates: development, homologies and evolution. Taylor & Francis (Oxford, UK). 650 pages.
DIOGO, R., B. SHEARER, J. M. POTAU, J. F. PASTOR, F. J. DE PAZ, J. ARIAS-MARTORELL, C. TURCOTTE, A. HAMMOND, E. VEREECKE, M. VANHOOF, S. NAUWELAERTS & B. WOOD (2017). Photographic and descriptive musculoskeletal atlas of bonobos - with notes on the weight, attachments, variations, and innervation of the muscles and comparisons with common chimpanzees and humans. Springer (New York, US). 259 pages.
DIOGO, R. (2017). Evolution driven by organismal behavior: a unifying view of life, function, form, mismatches and trends. Springer
Document 3:::
The following outline is provided as an overview of and topical guide to human anatomy:
Human anatomy – scientific study of the morphology of the adult human. It is subdivided into gross anatomy and microscopic anatomy. Gross anatomy (also called topographical anatomy, regional anatomy, or anthropotomy) is the study of anatomical structures that can be seen by unaided vision. Microscopic anatomy is the study of minute anatomical structures assisted with microscopes, and includes histology (the study of the organization of tissues), and cytology (the study of cells).
Essence of human anatomy
Human body
Anatomy
Branches of human anatomy
Gross anatomy- systemic or region-wise study of human body parts and organs. Gross anatomy encompasses cadaveric anatomy and osteology
Microscopic anatomy/histology
Cell biology (Cytology) & cytogenetics
Surface anatomy
Radiological anatomy
Developmental anatomy/embryology
Anatomy of the human body
The following list of human anatomical structures is based on the Terminologia Anatomica, the international standard for anatomical nomenclature. While the order is standardized, the hierarchical relationships in the TA are somewhat vague, and thus are open to interpretation.
General anatomy
Parts of human body
Head
Ear
Face
Forehead
Cheek
Chin
Eye
Nose
Nostril
Mouth
Lip
Tongue
Tooth
Neck
Torso
Thorax
Abdomen
Pelvis
Back
Pectoral girdle
Shoulder
Arm
Axilla
Elbow
Forearm
Wrist
Hand
Finger
Thumb
Palm
Lower limb
Pelvic girdle
Leg
Buttocks
Hip
Thigh
Knee
Calf
Foot
Ankle
Heel
Toe
Big toe
Sole
Cavities
Cranial cavity
Spinal cavity
Thoracic cavity
Abdominopelvic cavity
Abdominal cavity
Pelvic cavity
Planes, lines, and regions
Regions of head
Regions of neck
Anterior and lateral thoracic regions
Abdominal regions
Regions of back
Perineal regions
Regions of upper limb
Regions of lower limb
Bones
General terms
Bony part
Cortical bone
Compact bone
Spongy bone
Cartilaginous part
Membranous part
Periosteum
Perichondrium
Axial skele
Document 4:::
The skull is a bone protective cavity for the brain. The skull is composed of four types of bone i.e., cranial bones, facial bones, ear ossicles and hyoid bone. However two parts are more prominent: the cranium (: craniums or crania) and the mandible. In humans, these two parts are the neurocranium (braincase) and the viscerocranium (facial skeleton) that includes the mandible as its largest bone. The skull forms the anterior-most portion of the skeleton and is a product of cephalisation—housing the brain, and several sensory structures such as the eyes, ears, nose, and mouth. In humans these sensory structures are part of the facial skeleton.
Functions of the skull include protection of the brain, fixing the distance between the eyes to allow stereoscopic vision, and fixing the position of the ears to enable sound localisation of the direction and distance of sounds. In some animals, such as horned ungulates (mammals with hooves), the skull also has a defensive function by providing the mount (on the frontal bone) for the horns.
The English word skull is probably derived from Old Norse , while the Latin word comes from the Greek root (). The human skull fully develops two years after birth.The junctions of the skull bones are joined by structures called sutures.
The skull is made up of a number of fused flat bones, and contains many foramina, fossae, processes, and several cavities or sinuses. In zoology there are openings in the skull called fenestrae.
Structure
Humans
The human skull is the bone structure that forms the head in the human skeleton. It supports the structures of the face and forms a cavity for the brain. Like the skulls of other vertebrates, it protects the brain from injury.
The skull consists of three parts, of different embryological origin—the neurocranium, the sutures, and the facial skeleton (also called the membraneous viscerocranium). The neurocranium (or braincase) forms the protective cranial cavity that surrounds and houses the
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
The structure of the human tail bone is called what?
A. an artifact
B. a vital structure
C. a abnormal structure
D. a vestigial structure
Answer:
|
|
sciq-4992
|
multiple_choice
|
Most plant cells have a large central what?
|
[
"membranes",
"loci",
"nuclei",
"vacuole"
] |
D
|
Relavent Documents:
Document 0:::
Plant stem cells
Plant stem cells are innately undifferentiated cells located in the meristems of plants. Plant stem cells serve as the origin of plant vitality, as they maintain themselves while providing a steady supply of precursor cells to form differentiated tissues and organs in plants. Two distinct areas of stem cells are recognised: the apical meristem and the lateral meristem.
Plant stem cells are characterized by two distinctive properties, which are: the ability to create all differentiated cell types and the ability to self-renew such that the number of stem cells is maintained. Plant stem cells never undergo aging process but immortally give rise to new specialized and unspecialized cells, and they have the potential to grow into any organ, tissue, or cell in the body. Thus they are totipotent cells equipped with regenerative powers that facilitate plant growth and production of new organs throughout lifetime.
Unlike animals, plants are immobile. As plants cannot escape from danger by taking motion, they need a special mechanism to withstand various and sometimes unforeseen environmental stress. Here, what empowers them to withstand harsh external influence and preserve life is stem cells. In fact, plants comprise the oldest and the largest living organisms on earth, including Bristlecone Pines in California, U.S. (4,842 years old), and the Giant Sequoia in mountainous regions of California, U.S. (87 meters in height and 2,000 tons in weight). This is possible because they have a modular body plan that enables them to survive substantial damage by initiating continuous and repetitive formation of new structures and organs such as leaves and flowers.
Plant stem cells are also characterized by their location in specialized structures called meristematic tissues, which are located in root apical meristem (RAM), shoot apical meristem (SAM), and vascular system ((pro)cambium or vascular meristem.)
Research and development
Traditionally, plant stem ce
Document 1:::
A stem is one of two main structural axes of a vascular plant, the other being the root. It supports leaves, flowers and fruits, transports water and dissolved substances between the roots and the shoots in the xylem and phloem, photosynthesis takes place here, stores nutrients, and produces new living tissue. The stem can also be called halm or haulm or culms.
The stem is normally divided into nodes and internodes:
The nodes are the points of attachment for leaves and can hold one or more leaves. There are sometimes axillary buds between the stem and leaf which can grow into branches (with leaves, conifer cones, or flowers). Adventitious roots may also be produced from the nodes. Vines may produce tendrils from nodes.
The internodes distance one node from another.
The term "shoots" is often confused with "stems"; "shoots" generally refers to new fresh plant growth, including both stems and other structures like leaves or flowers.
In most plants, stems are located above the soil surface, but some plants have underground stems.
Stems have several main functions:
Support for and the elevation of leaves, flowers, and fruits. The stems keep the leaves in the light and provide a place for the plant to keep its flowers and fruits.
Transport of fluids between the roots and the shoots in the xylem and phloem.
Storage of nutrients.
Production of new living tissue. The normal lifespan of plant cells is one to three years. Stems have cells called meristems that annually generate new living tissue.
Photosynthesis.
Stems have two pipe-like tissues called xylem and phloem. The xylem tissue arises from the cell facing inside and transports water by the action of transpiration pull, capillary action, and root pressure. The phloem tissue arises from the cell facing outside and consists of sieve tubes and their companion cells. The function of phloem tissue is to distribute food from photosynthetic tissue to other tissues. The two tissues are separated by cambium, a tis
Document 2:::
In botany, a cortex is an outer layer of a stem or root in a vascular plant, lying below the epidermis but outside of the vascular bundles. The cortex is composed mostly of large thin-walled parenchyma cells of the ground tissue system and shows little to no structural differentiation. The outer cortical cells often acquire irregularly thickened cell walls, and are called collenchyma cells.
Plants
Stems and branches
In the three dimensional structure of herbaceous stems, the epidermis, cortex and vascular cambium form concentric cylinders around the inner cylindrical core of pith. Some of the outer cortical cells may contain chloroplasts, giving them a green color. They can therefore produce simple carbohydrates through photosynthesis.
In woody plants, the cortex is located between the periderm (bark) and the vascular tissue (phloem, in particular). It is responsible for the transportation of materials into the central cylinder of the root through diffusion and may also be used for storage of food in the form of starch.
Roots
In the roots of vascular plants, the cortex occupies a larger portion of the organ's volume than in herbaceous stems. The loosely packed cells of root cortex allow movement of water and oxygen in the intercellular spaces.
One of the main functions of the root cortex is to serve as a storage area for reserve foods. The innermost layer of the cortex in the roots of vascular plants is the endodermis. The endodermis is responsible for storing starch as well as regulating the transport of water, ions and plant hormones.
Lichen
On a lichen, the cortex is also the surface layer or "skin" of the nonfruiting part of the body of some lichens. It is the "skin", or outer layer of tissue, that covers the undifferentiated cells of the . Fruticose lichens have one cortex encircling the branches, even flattened, leaf-like forms. Foliose lichens have different upper and lower cortices. Crustose, placodioid, and squamulose lichens have an upper cor
Document 3:::
The quiescent centre is a group of cells, up to 1,000 in number, in the form of a hemisphere, with the flat face toward the root tip of vascular plants. It is a region in the apical meristem of a root where cell division proceeds very slowly or not at all, but the cells are capable of resuming meristematic activity when the tissue surrounding them is damaged.
Cells of root apical meristems do not all divide at the same rate. Determinations of relative rates of DNA synthesis show that primary roots of Zea, Vicia and Allium have quiescent centres to the meristems, in which the cells divide rarely or never in the course of normal root growth (Clowes, 1958). Such a quiescent centre includes the cells at the apices of the histogens of both stele and cortex. Its presence can be deduced from the anatomy of the apex in Zea (Clowes, 1958), but not in the other species which lack discrete histogens.
History
In 1953, during the course of analysing the organization and function of the root apices, Frederick Albert Lionel Clowes (born 10 September 1921), at the School of Botany (now Department of Plant Sciences), University of Oxford, proposed the term ‘cytogenerative centre’ to denote ‘the region of an apical meristem from which all future cells are derived’. This term had been suggested to him by Mr Harold K. Pusey, a lecturer in embryology at the Department of Zoology and Comparative Anatomy at the same university. The 1953 paper of Clowes reported results of his experiments on Fagus sylvatica and Vicia faba, in which small oblique and wedge-shaped excisions were made at the tip of the primary root, at the most distal level of the root body, near the boundary with the root cap. The results of these experiments were striking and showed that: the root which grew on following the excision was normal at the undamaged meristem side; the nonexcised meristem portion contributed to the regeneration of the excised portion; the regenerated part of the root had abnormal patterning and
Document 4:::
The ground tissue of plants includes all tissues that are neither dermal nor vascular. It can be divided into three types based on the nature of the cell walls. This tissue system is present between the dermal tissue and forms the main bulk of the plant body.
Parenchyma cells have thin primary walls and usually remain alive after they become mature. Parenchyma forms the "filler" tissue in the soft parts of plants, and is usually present in cortex, pericycle, pith, and medullary rays in primary stem and root.
Collenchyma cells have thin primary walls with some areas of secondary thickening. Collenchyma provides extra mechanical and structural support, particularly in regions of new growth.
Sclerenchyma cells have thick lignified secondary walls and often die when mature. Sclerenchyma provides the main structural support to a plant.
Parenchyma
Parenchyma is a versatile ground tissue that generally constitutes the "filler" tissue in soft parts of plants. It forms, among other things, the cortex (outer region) and pith (central region) of stems, the cortex of roots, the mesophyll of leaves, the pulp of fruits, and the endosperm of seeds. Parenchyma cells are often living cells and may remain meristematic, meaning that they are capable of cell division if stimulated. They have thin and flexible cellulose cell walls and are generally polyhedral when close-packed, but can be roughly spherical when isolated from their neighbors. Parenchyma cells are generally large. They have large central vacuoles, which allow the cells to store and regulate ions, waste products, and water. Tissue specialised for food storage is commonly formed of parenchyma cells.
Parenchyma cells have a variety of functions:
In leaves, they form two layers of mesophyll cells immediately beneath the epidermis of the leaf, that are responsible for photosynthesis and the exchange of gases. These layers are called the palisade parenchyma and spongy mesophyll. Palisade parenchyma cells can be either cu
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Most plant cells have a large central what?
A. membranes
B. loci
C. nuclei
D. vacuole
Answer:
|
|
sciq-8095
|
multiple_choice
|
What do tadpoles turn into?
|
[
"insects",
"fish",
"snakes",
"frogs"
] |
D
|
Relavent Documents:
Document 0:::
A pre-STEM program is a course of study at any two-year college that prepares a student to transfer to a four-year school to earn a bachelor's degree in a STEM field.
Overview
The concept of a pre-STEM program is being developed to address America's need for more college-trained professionals in science, technology, engineering, and mathematics (STEM). It is an innovation meant to fill a gap at community colleges that do not have 'major' degree paths that students identify with on their way to earning an Associates degree. Students must complete a considerable amount of STEM coursework before transferring from a two-year school to a four-year school and earn a baccalaureate degree in a STEM field. Schools with a pre-STEM program are able to identify those students and support them with STEM-specific academic and career advising, increasing the student's chances of going on to earn a STEM baccalaureate degree in a timely fashion.
With over 50% of America's college-bound students starting their college career at public or private two-year school, and with a very small proportion of students who start college at a two-year school matriculating to and earning STEM degrees from four-year schools, pre-STEM programs have great potential for broadening participation in baccalaureate STEM studies.
Example programs
The effectiveness of pre-STEM programs is being investigated by a consortium of schools in Missouri: Moberly Area Community College, St. Charles Community College, Metropolitan Community College, and Truman State University.
A larger group of schools met at the Belknap Springs Meetings in October 2009 to discuss the challenges and opportunities presented by STEM-focused partnerships between 2-year and 4-year schools. Each program represented a two-year school and a four-year school that were trying to increase the number of people who earn a baccalaureate degree in a STEM area through various means, some of which were pre-STEM programs. Other methods includes
Document 1:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 2:::
Female education in STEM refers to child and adult female representation in the educational fields of science, technology, engineering, and mathematics (STEM). In 2017, 33% of students in STEM fields were women.
The organization UNESCO has stated that this gender disparity is due to discrimination, biases, social norms and expectations that influence the quality of education women receive and the subjects they study. UNESCO also believes that having more women in STEM fields is desirable because it would help bring about sustainable development.
Current status of girls and women in STEM education
Overall trends in STEM education
Gender differences in STEM education participation are already visible in early childhood care and education in science- and math-related play, and become more pronounced at higher levels of education. Girls appear to lose interest in STEM subjects with age, particularly between early and late adolescence. This decreased interest affects participation in advanced studies at the secondary level and in higher education. Female students represent 35% of all students enrolled in STEM-related fields of study at this level globally. Differences are also observed by disciplines, with female enrollment lowest in engineering, manufacturing and construction, natural science, mathematics and statistics and ICT fields. Significant regional and country differences in female representation in STEM studies can be observed, though, suggesting the presence of contextual factors affecting girls’ and women's engagement in these fields. Women leave STEM disciplines in disproportionate numbers during their higher education studies, in their transition to the world of work and even in their career cycle.
Learning achievement in STEM education
Data on gender differences in learning achievement present a complex picture, depending on what is measured (subject, knowledge acquisition against knowledge application), the level of education/age of students, and
Document 3:::
Tech City College (Formerly STEM Academy) is a free school sixth form located in the Islington area of the London Borough of Islington, England.
It originally opened in September 2013, as STEM Academy Tech City and specialised in Science, Technology, Engineering and Maths (STEM) and the Creative Application of Maths and Science. In September 2015, STEM Academy joined the Aspirations Academy Trust was renamed Tech City College. Tech City College offers A-levels and BTECs as programmes of study for students.
Document 4:::
The Bachelor of Science in Aquatic Resources and Technology (B.Sc. in AQT) (or Bachelor of Aquatic Resource) is an undergraduate degree that prepares students to pursue careers in the public, private, or non-profit sector in areas such as marine science, fisheries science, aquaculture, aquatic resource technology, food science, management, biotechnology and hydrography. Post-baccalaureate training is available in aquatic resource management and related areas.
The Department of Animal Science and Export Agriculture, at the Uva Wellassa University of Badulla, Sri Lanka, has the largest enrollment of undergraduate majors in Aquatic Resources and Technology, with about 200 students as of 2014.
The Council on Education for Aquatic Resources and Technology includes undergraduate AQT degrees in the accreditation review of Aquatic Resources and Technology programs and schools.
See also
Marine Science
Ministry of Fisheries and Aquatic Resources Development
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What do tadpoles turn into?
A. insects
B. fish
C. snakes
D. frogs
Answer:
|
|
sciq-9368
|
multiple_choice
|
What happens to neutral matter when electrons are transferred between objects?
|
[
"osmosis",
"it dissolves itself",
"it stays neutral",
"it becomes charged"
] |
D
|
Relavent Documents:
Document 0:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 1:::
Advanced Placement (AP) Physics C: Electricity and Magnetism (also known as AP Physics C: E&M or AP E&M) is an introductory physics course administered by the College Board as part of its Advanced Placement program. It is intended to proxy a second-semester calculus-based university course in electricity and magnetism. The content of Physics C: E&M overlaps with that of AP Physics 2, but Physics 2 is algebra-based and covers other topics outside of electromagnetism, while Physics C is calculus-based and only covers electromagnetism. Physics C: E&M may be combined with its mechanics counterpart to form a year-long course that prepares for both exams.
Course content
E&M is equivalent to an introductory college course in electricity and magnetism for physics or engineering majors. The course modules are:
Electrostatics
Conductors, capacitors, and dielectrics
Electric circuits
Magnetic fields
Electromagnetism.
Methods of calculus are used wherever appropriate in formulating physical principles and in applying them to physical problems. Therefore, students should have completed or be concurrently enrolled in a calculus class.
AP test
The course culminates in an optional exam for which high-performing students may receive some credit towards their college coursework, depending on the institution.
Registration
The AP examination for AP Physics C: Electricity and Magnetism is separate from the AP examination for AP Physics C: Mechanics. Before 2006, test-takers paid only once and were given the choice of taking either one or two parts of the Physics C test.
Format
The exam is typically administered on a Monday afternoon in May. The exam is configured in two categories: a 35-question multiple choice section and a 3-question free response section. Test takers are allowed to use an approved calculator during the entire exam. The test is weighted such that each section is worth half of the final score. This and AP Physics C: Mechanics are the shortest AP exams, with
Document 2:::
A solvated electron is a free electron in (solvated in) a solution, and is the smallest possible anion. Solvated electrons occur widely. Often, discussions of solvated electrons focus on their solutions in ammonia, which are stable for days, but solvated electrons also occur in water and other solvents in fact, in any solvent that mediates outer-sphere electron transfer. The solvated electron is responsible for a great deal of radiation chemistry.
Ammonia solutions
Liquid ammonia will dissolve all of the alkali metals and other electropositive metals such as Ca, Sr, Ba, Eu, and Yb (also Mg using an electrolytic process), giving characteristic blue solutions. For alkali metals in liquid ammonia, the solution is blue when dilute and copper-colored when more concentrated (> 3 molar). These solutions conduct electricity. The blue colour of the solution is due to ammoniated electrons, which absorb energy in the visible region of light. The diffusivity of the solvated electron in liquid ammonia can be determined using potential-step chronoamperometry.
Solvated electrons in ammonia are the anions of salts called electrides.
Na + 6 NH3 → [Na(NH3)6]+e−
The reaction is reversible: evaporation of the ammonia solution produces a film of metallic sodium.
Case study: Li in NH3
A lithium–ammonia solution at −60 °C is saturated at about 15 mol% metal (MPM). When the concentration is increased in this range electrical conductivity increases from 10−2 to 104 Ω−1cm−1 (larger than liquid mercury). At around 8 MPM, a "transition to the metallic state" (TMS) takes place (also called a "metal-to-nonmetal transition" (MNMT)). At 4 MPM a liquid-liquid phase separation takes place: the less dense gold-colored phase becomes immiscible from a denser blue phase. Above 8 MPM the solution is bronze/gold-colored. In the same concentration range the overall density decreases by 30%.
Other solvents
Alkali metals also dissolve in some small primary amines, such as methylamine and ethylami
Document 3:::
In mathematical psychology and education theory, a knowledge space is a combinatorial structure used to formulate mathematical models describing the progression of a human learner. Knowledge spaces were introduced in 1985 by Jean-Paul Doignon and Jean-Claude Falmagne, and remain in extensive use in the education theory. Modern applications include two computerized tutoring systems, ALEKS and the defunct RATH.
Formally, a knowledge space assumes that a domain of knowledge is a collection of concepts or skills, each of which must be eventually mastered. Not all concepts are interchangeable; some require other concepts as prerequisites. Conversely, competency at one skill may ease the acquisition of another through similarity. A knowledge space marks out which collections of skills are feasible: they can be learned without mastering any other skills. Under reasonable assumptions, the collection of feasible competencies forms the mathematical structure known as an antimatroid.
Researchers and educators usually explore the structure of a discipline's knowledge space as a latent class model.
Motivation
Knowledge Space Theory attempts to address shortcomings of standardized testing when used in educational psychometry. Common tests, such as the SAT and ACT, compress a student's knowledge into a very small range of ordinal ranks, in the process effacing the conceptual dependencies between questions. Consequently, the tests cannot distinguish between true understanding and guesses, nor can they identify a student's particular weaknesses, only the general proportion of skills mastered. The goal of knowledge space theory is to provide a language by which exams can communicate
What the student can do and
What the student is ready to learn.
Model structure
Knowledge Space Theory-based models presume that an educational subject can be modeled as a finite set of concepts, skills, or topics. Each feasible state of knowledge about is then a subset of ; the set of
Document 4:::
There are four Advanced Placement (AP) Physics courses administered by the College Board as part of its Advanced Placement program: the algebra-based Physics 1 and Physics 2 and the calculus-based Physics C: Mechanics and Physics C: Electricity and Magnetism. All are intended to be at the college level. Each AP Physics course has an exam for which high-performing students may receive credit toward their college coursework.
AP Physics 1 and 2
AP Physics 1 and AP Physics 2 were introduced in 2015, replacing AP Physics B. The courses were designed to emphasize critical thinking and reasoning as well as learning through inquiry. They are algebra-based and do not require any calculus knowledge.
AP Physics 1
AP Physics 1 covers Newtonian mechanics, including:
Unit 1: Kinematics
Unit 2: Dynamics
Unit 3: Circular Motion and Gravitation
Unit 4: Energy
Unit 5: Momentum
Unit 6: Simple Harmonic Motion
Unit 7: Torque and Rotational Motion
Until 2020, the course also covered topics in electricity (including Coulomb's Law and resistive DC circuits), mechanical waves, and sound. These units were removed because they are included in AP Physics 2.
AP Physics 2
AP Physics 2 covers the following topics:
Unit 1: Fluids
Unit 2: Thermodynamics
Unit 3: Electric Force, Field, and Potential
Unit 4: Electric Circuits
Unit 5: Magnetism and Electromagnetic Induction
Unit 6: Geometric and Physical Optics
Unit 7: Quantum, Atomic, and Nuclear Physics
AP Physics C
From 1969 to 1972, AP Physics C was a single course with a single exam that covered all standard introductory university physics topics, including mechanics, fluids, electricity and magnetism, optics, and modern physics. In 1973, the College Board split the course into AP Physics C: Mechanics and AP Physics C: Electricity and Magnetism. The exam was also split into two separate 90-minute tests, each equivalent to a semester-length calculus-based college course. Until 2006, both exams could be taken for a single
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What happens to neutral matter when electrons are transferred between objects?
A. osmosis
B. it dissolves itself
C. it stays neutral
D. it becomes charged
Answer:
|
|
sciq-3722
|
multiple_choice
|
Some atoms are more stable when they gain or lose an electron and form what?
|
[
"electrons",
"protons",
"ions",
"molecules"
] |
C
|
Relavent Documents:
Document 0:::
An atom is a particle that consists of a nucleus of protons and neutrons surrounded by an electromagnetically-bound cloud of electrons. The atom is the basic particle of the chemical elements, and the chemical elements are distinguished from each other by the number of protons that are in their atoms. For example, any atom that contains 11 protons is sodium, and any atom that contains 29 protons is copper. The number of neutrons defines the isotope of the element.
Atoms are extremely small, typically around 100 picometers across. A human hair is about a million carbon atoms wide. This is smaller than the shortest wavelength of visible light, which means humans cannot see atoms with conventional microscopes. Atoms are so small that accurately predicting their behavior using classical physics is not possible due to quantum effects.
More than 99.94% of an atom's mass is in the nucleus. Each proton has a positive electric charge, while each electron has a negative charge, and the neutrons, if any are present, have no electric charge. If the numbers of protons and electrons are equal, as they normally are, then the atom is electrically neutral. If an atom has more electrons than protons, then it has an overall negative charge, and is called a negative ion (or anion). Conversely, if it has more protons than electrons, it has a positive charge, and is called a positive ion (or cation).
The electrons of an atom are attracted to the protons in an atomic nucleus by the electromagnetic force. The protons and neutrons in the nucleus are attracted to each other by the nuclear force. This force is usually stronger than the electromagnetic force that repels the positively charged protons from one another. Under certain circumstances, the repelling electromagnetic force becomes stronger than the nuclear force. In this case, the nucleus splits and leaves behind different elements. This is a form of nuclear decay.
Atoms can attach to one or more other atoms by chemical bonds to
Document 1:::
Atomic physics is the field of physics that studies atoms as an isolated system of electrons and an atomic nucleus. Atomic physics typically refers to the study of atomic structure and the interaction between atoms. It is primarily concerned with the way in which electrons are arranged around the nucleus and
the processes by which these arrangements change. This comprises ions, neutral atoms and, unless otherwise stated, it can be assumed that the term atom includes ions.
The term atomic physics can be associated with nuclear power and nuclear weapons, due to the synonymous use of atomic and nuclear in standard English. Physicists distinguish between atomic physics—which deals with the atom as a system consisting of a nucleus and electrons—and nuclear physics, which studies nuclear reactions and special properties of atomic nuclei.
As with many scientific fields, strict delineation can be highly contrived and atomic physics is often considered in the wider context of atomic, molecular, and optical physics. Physics research groups are usually so classified.
Isolated atoms
Atomic physics primarily considers atoms in isolation. Atomic models will consist of a single nucleus that may be surrounded by one or more bound electrons. It is not concerned with the formation of molecules (although much of the physics is identical), nor does it examine atoms in a solid state as condensed matter. It is concerned with processes such as ionization and excitation by photons or collisions with atomic particles.
While modelling atoms in isolation may not seem realistic, if one considers atoms in a gas or plasma then the time-scales for atom-atom interactions are huge in comparison to the atomic processes that are generally considered. This means that the individual atoms can be treated as if each were in isolation, as the vast majority of the time they are. By this consideration, atomic physics provides the underlying theory in plasma physics and atmospheric physics, even though
Document 2:::
In chemistry and physics, the iron group refers to elements that are in some way related to iron; mostly in period (row) 4 of the periodic table. The term has different meanings in different contexts.
In chemistry, the term is largely obsolete, but it often means iron, cobalt, and nickel, also called the iron triad; or, sometimes, other elements that resemble iron in some chemical aspects.
In astrophysics and nuclear physics, the term is still quite common, and it typically means those three plus chromium and manganese—five elements that are exceptionally abundant, both on Earth and elsewhere in the universe, compared to their neighbors in the periodic table. Titanium and vanadium are also produced in Type Ia supernovae.
General chemistry
In chemistry, "iron group" used to refer to iron and the next two elements in the periodic table, namely cobalt and nickel. These three comprised the "iron triad". They are the top elements of groups 8, 9, and 10 of the periodic table; or the top row of "group VIII" in the old (pre-1990) IUPAC system, or of "group VIIIB" in the CAS system. These three metals (and the three of the platinum group, immediately below them) were set aside from the other elements because they have obvious similarities in their chemistry, but are not obviously related to any of the other groups. The iron group and its alloys exhibit ferromagnetism.
The similarities in chemistry were noted as one of Döbereiner's triads and by Adolph Strecker in 1859. Indeed, Newlands' "octaves" (1865) were harshly criticized for separating iron from cobalt and nickel. Mendeleev stressed that groups of "chemically analogous elements" could have similar atomic weights as well as atomic weights which increase by equal increments, both in his original 1869 paper and his 1889 Faraday Lecture.
Analytical chemistry
In the traditional methods of qualitative inorganic analysis, the iron group consists of those cations which
have soluble chlorides; and
are not precipitated
Document 3:::
In chemistry and physics, valence electrons are electrons in the outermost shell of an atom, and that can participate in the formation of a chemical bond if the outermost shell is not closed. In a single covalent bond, a shared pair forms with both atoms in the bond each contributing one valence electron.
The presence of valence electrons can determine the element's chemical properties, such as its valence—whether it may bond with other elements and, if so, how readily and with how many. In this way, a given element's reactivity is highly dependent upon its electronic configuration. For a main-group element, a valence electron can exist only in the outermost electron shell; for a transition metal, a valence electron can also be in an inner shell.
An atom with a closed shell of valence electrons (corresponding to a noble gas configuration) tends to be chemically inert. Atoms with one or two valence electrons more than a closed shell are highly reactive due to the relatively low energy to remove the extra valence electrons to form a positive ion. An atom with one or two electrons fewer than a closed shell is reactive due to its tendency either to gain the missing valence electrons and form a negative ion, or else to share valence electrons and form a covalent bond.
Similar to a core electron, a valence electron has the ability to absorb or release energy in the form of a photon. An energy gain can trigger the electron to move (jump) to an outer shell; this is known as atomic excitation. Or the electron can even break free from its associated atom's shell; this is ionization to form a positive ion. When an electron loses energy (thereby causing a photon to be emitted), then it can move to an inner shell which is not fully occupied.
Overview
Electron configuration
The electrons that determine valence – how an atom reacts chemically – are those with the highest energy.
For a main-group element, the valence electrons are defined as those electrons residing in the e
Document 4:::
A fixed orbit is the concept, in atomic physics, where an electron is considered to remain in a specific orbit, at a fixed distance from an atom's nucleus, for a particular energy level.
The concept was promoted by quantum physicist Niels Bohr c. 1913.
The idea of the fixed orbit is considered a major component of the Bohr model (or Bohr theory).
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Some atoms are more stable when they gain or lose an electron and form what?
A. electrons
B. protons
C. ions
D. molecules
Answer:
|
|
sciq-2788
|
multiple_choice
|
What is the rate of change of velocity called?
|
[
"acceleration",
"vibration",
"transmission",
"speed"
] |
A
|
Relavent Documents:
Document 0:::
Velocity is the speed in combination with the direction of motion of an object. Velocity is a fundamental concept in kinematics, the branch of classical mechanics that describes the motion of bodies.
Velocity is a physical vector quantity: both magnitude and direction are needed to define it. The scalar absolute value (magnitude) of velocity is called , being a coherent derived unit whose quantity is measured in the SI (metric system) as metres per second (m/s or m⋅s−1). For example, "5 metres per second" is a scalar, whereas "5 metres per second east" is a vector. If there is a change in speed, direction or both, then the object is said to be undergoing an acceleration.
Constant velocity vs acceleration
To have a constant velocity, an object must have a constant speed in a constant direction. Constant direction constrains the object to motion in a straight path thus, a constant velocity means motion in a straight line at a constant speed.
For example, a car moving at a constant 20 kilometres per hour in a circular path has a constant speed, but does not have a constant velocity because its direction changes. Hence, the car is considered to be undergoing an acceleration.
Difference between speed and velocity
While the terms speed and velocity are often colloquially used interchangeably to connote how fast an object is moving, in scientific terms they are different. Speed, the scalar magnitude of a velocity vector, denotes only how fast an object is moving, while velocity indicates both an objects speed and direction.
Equation of motion
Average velocity
Velocity is defined as the rate of change of position with respect to time, which may also be referred to as the instantaneous velocity to emphasize the distinction from the average velocity. In some applications the average velocity of an object might be needed, that is to say, the constant velocity that would provide the same resultant displacement as a variable velocity in the same time interval, , over some
Document 1:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 2:::
Linear motion, also called rectilinear motion, is one-dimensional motion along a straight line, and can therefore be described mathematically using only one spatial dimension. The linear motion can be of two types: uniform linear motion, with constant velocity (zero acceleration); and non-uniform linear motion, with variable velocity (non-zero acceleration). The motion of a particle (a point-like object) along a line can be described by its position , which varies with (time). An example of linear motion is an athlete running a 100-meter dash along a straight track.
Linear motion is the most basic of all motion. According to Newton's first law of motion, objects that do not experience any net force will continue to move in a straight line with a constant velocity until they are subjected to a net force. Under everyday circumstances, external forces such as gravity and friction can cause an object to change the direction of its motion, so that its motion cannot be described as linear.
One may compare linear motion to general motion. In general motion, a particle's position and velocity are described by vectors, which have a magnitude and direction. In linear motion, the directions of all the vectors describing the system are equal and constant which means the objects move along the same axis and do not change direction. The analysis of such systems may therefore be simplified by neglecting the direction components of the vectors involved and dealing only with the magnitude.
Background
Displacement
The motion in which all the particles of a body move through the same distance in the same time is called translatory motion. There are two types of translatory motions: rectilinear motion; curvilinear motion. Since linear motion is a motion in a single dimension, the distance traveled by an object in particular direction is the same as displacement. The SI unit of displacement is the metre. If is the initial position of an object and is the final position, then mat
Document 3:::
Most of the terms listed in Wikipedia glossaries are already defined and explained within Wikipedia itself. However, glossaries like this one are useful for looking up, comparing and reviewing large numbers of terms together. You can help enhance this page by adding new terms or writing definitions for existing ones.
This glossary of mechanical engineering terms pertains specifically to mechanical engineering and its sub-disciplines. For a broad overview of engineering, see glossary of engineering.
A
Abrasion – is the process of scuffing, scratching, wearing down, marring, or rubbing away. It can be intentionally imposed in a controlled process using an abrasive. Abrasion can be an undesirable effect of exposure to normal use or exposure to the elements.
Absolute zero – is the lowest possible temperature of a system, defined as zero kelvin or −273.15 °C. No experiment has yet measured a temperature of absolute zero.
Accelerated life testing – is the process of testing a product by subjecting it to conditions (stress, strain, temperatures, voltage, vibration rate, pressure etc.) in excess of its normal service parameters in an effort to uncover faults and potential modes of failure in a short amount of time. By analyzing the product's response to such tests, engineers can make predictions about the service life and maintenance intervals of a product.
Acceleration – In physics, acceleration is the rate of change of velocity of an object with respect to time. An object's acceleration is the net result of any and all forces acting on the object, as described by Newton's Second Law. The SI unit for acceleration is metre per second squared Accelerations are vector quantities (they have magnitude and direction) and add according to the parallelogram law. As a vector, the calculated net force is equal to the product of the object's mass (a scalar quantity) and its acceleration.
Accelerometer – is a device that measures proper acceleration. Proper acceleration, being
Document 4:::
Advanced Placement (AP) Physics C: Mechanics (also known as AP Mechanics) is an introductory physics course administered by the College Board as part of its Advanced Placement program. It is intended to proxy a one-semester calculus-based university course in mechanics. The content of Physics C: Mechanics overlaps with that of AP Physics 1, but Physics 1 is algebra-based, while Physics C is calculus-based. Physics C: Mechanics may be combined with its electricity and magnetism counterpart to form a year-long course that prepares for both exams.
Course content
Intended to be equivalent to an introductory college course in mechanics for physics or engineering majors, the course modules are:
Kinematics
Newton's laws of motion
Work, energy and power
Systems of particles and linear momentum
Circular motion and rotation
Oscillations and gravitation.
Methods of calculus are used wherever appropriate in formulating physical principles and in applying them to physical problems. Therefore, students should have completed or be concurrently enrolled in a Calculus I class.
This course is often compared to AP Physics 1: Algebra Based for its similar course material involving kinematics, work, motion, forces, rotation, and oscillations. However, AP Physics 1: Algebra Based lacks concepts found in Calculus I, like derivatives or integrals.
This course may be combined with AP Physics C: Electricity and Magnetism to make a unified Physics C course that prepares for both exams.
AP test
The course culminates in an optional exam for which high-performing students may receive some credit towards their college coursework, depending on the institution.
Registration
The AP examination for AP Physics C: Mechanics is separate from the AP examination for AP Physics C: Electricity and Magnetism. Before 2006, test-takers paid only once and were given the choice of taking either one or two parts of the Physics C test.
Format
The exam is typically administered on a Monday aftern
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is the rate of change of velocity called?
A. acceleration
B. vibration
C. transmission
D. speed
Answer:
|
|
sciq-3436
|
multiple_choice
|
Blood enters the kidney through the which artery?
|
[
"vas deferens",
"cardiac artery",
"renal artery",
"jugular"
] |
C
|
Relavent Documents:
Document 0:::
The renal arteries are paired arteries that supply the kidneys with blood. Each is directed across the crus of the diaphragm, so as to form nearly a right angle.
The renal arteries carry a large portion of total blood flow to the kidneys. Up to a third of total cardiac output can pass through the renal arteries to be filtered by the kidneys.
Structure
The renal arteries normally arise at a 90° angle off of the left interior side of the abdominal aorta, immediately below the superior mesenteric artery. They have a radius of approximately 0.25 cm, 0.26 cm at the root. The measured mean diameter can differ depending on the imaging method used. For example, the diameter was found to be 5.04 ± 0.74 mm using ultrasound but 5.68 ± 1.19 mm using angiography.
Due to the anatomical position of the aorta, the inferior vena cava, and the kidneys, the right renal artery is normally longer than the left renal artery.
The right passes behind the inferior vena cava, the right renal vein, the head of the pancreas, and the descending part of the duodenum. It’s somewhat lower than the left one.
Left artery lies behind the left renal vein, the body of the pancreas and the splenic vein, and is crossed by the inferior mesenteric vein.
Branches
Before reaching the hilus of the kidney, each artery divides into four or five branches. The anterior branches (the upper, middle, lower and apical segmental arteries) lie between the renal vein and ureter, the vein being in front, the ureter behind. The posterior branches, which are fewer in number and include the posterior segmental artery, are usually situated behind the ureter.
Each vessel gives off some small inferior suprarenal branches to the suprarenal gland, the ureter, and the surrounding cellular tissue and muscles.
One or two accessory renal arteries are frequently found, especially on the left side since they usually arise from the aorta, and may come off above (more common) or below the main artery. Instead of entering the ki
Document 1:::
The renal circulation supplies the blood to the kidneys via the renal arteries, left and right, which branch directly from the abdominal aorta. Despite their relatively small size, the kidneys receive approximately 20% of the cardiac output.
Each renal artery branches into segmental arteries, dividing further into interlobar arteries, which penetrate the renal capsule and extend through the renal columns between the renal pyramids. The interlobar arteries then supply blood to the arcuate arteries that run through the boundary of the cortex and the medulla. Each arcuate artery supplies several interlobular arteries that feed into the afferent arterioles that supply the glomeruli.
After filtration occurs, the blood moves through a small network of venules that converge into interlobular veins. As with the arteriole distribution, the veins follow the same pattern: the interlobular provide blood to the arcuate veins then back to the interlobar veins, which come to form the renal vein exiting the kidney for transfusion for blood.
Structure
Arterial system
The table below shows the path that blood takes when it travels through the glomerulus, traveling "down" the arteries and "up" the veins. However, this model is greatly simplified for clarity and symmetry. Some of the other paths and complications are described at the bottom of the table. The interlobar artery and vein (not to be confused with interlobular) are between two renal lobes, also known as the renal column (cortex region between two pyramids).
Note 1: The renal artery also provides a branch to the inferior suprarenal artery to supply the adrenal gland.
Note 2: Also called the cortical radiate arteries. The interlobular artery also supplies to the stellate veins.
Note 3: The efferent arterioles do not directly drain into the interlobular vein, but rather they go to the peritubular capillaries first. The efferent arterioles of the juxtamedullary nephron drain into the vasa recta.
Segmental arteries
The
Document 2:::
The vasa recta of the kidney, (vasa recta renis) are the straight arterioles, and the straight venules of the kidney, – a series of blood vessels in the blood supply of the kidney that enter the medulla as the straight arterioles, and leave the medulla to ascend to the cortex as the straight venules. (Latin: vās, "vessel"; rēctus, "straight"). They lie parallel to the loop of Henle.
These vessels branch off the efferent arterioles of juxtamedullary nephrons (those nephrons closest to the medulla). They enter the medulla, and surround the loop of Henle. Whereas the peritubular capillaries surround the cortical parts of the tubules, the vasa recta go into the medulla and are closer to the loop of Henle, and leave to ascend to the cortex.
Terminations of the vasa recta form the straight venules, branches from the plexuses at the apices of the medullary pyramids. They run outward in a straight course between the tubes of the medullary substance and join the interlobular veins to form venous arcades. These in turn unite and form veins which pass along the sides of the renal pyramids. The descending vasa recta have a non-fenestrated endothelium that contains a facilitated transport for urea; the ascending vasa recta have, on the other hand, a fenestrated endothelium.
Structure
Microanatomy
On a histological slide, the straight arterioles can be distinguished from the tubules of the loop of Henle by the presence of blood.
Function
Each straight arteriole has a hairpin turn in the medulla and carries blood at a very slow rate – two factors crucial in the maintenance of countercurrent exchange that prevent washout of the concentration gradients established in the renal medulla.
The maintenance of this concentration gradient is one of the components responsible for the kidney's ability to produce concentrated urine.
On the descending portion of the vasa recta, sodium, chloride and urea are reabsorbed into the blood, while water is secreted. On the ascending portion, so
Document 3:::
The uterine artery supplies branches to the cervix uteri and others which descend on the vagina; the latter anastomose with branches of the vaginal arteries and form with them two median longitudinal vessels—the vaginal branches of uterine artery (or azygos arteries of the vagina)—one of which runs down in front of and the other behind the vagina.
Document 4:::
Watershed area is the medical term referring to regions of the body, that receive dual blood supply from the most distal branches of two large arteries, such as the splenic flexure of the large intestine. The term refers metaphorically to a geological watershed, or drainage divide, which separates adjacent drainage basins.
During times of blockage of one of the arteries that supply the watershed area, such as in atherosclerosis, these regions are spared from ischemia by virtue of their dual supply. However, during times of systemic hypoperfusion, such as in disseminated intravascular coagulation or heart failure, these regions are particularly vulnerable to ischemia because they are supplied by the most distal branches of their arteries, and thus the least likely to receive sufficient blood.
Watershed areas are found in the brain, where areas are perfused by both the anterior and middle cerebral arteries, and in the intestines, where areas are perfused by both the superior and inferior mesenteric arteries (i.e., splenic flexure). Additionally, the sigmoid colon and rectum form a watershed zone with blood supply from inferior mesenteric, pudendal and iliac circulations. Hypoperfusion in watershed areas can lead to mural and mucosal infarction in the case of ischemic bowel disease. When watershed stroke occurs in the brain, it produces unique focal neurologic symptoms that aid clinicians in diagnosis and localization. For example, a cerebral watershed area is situated in the dorsal prefrontal cortex; when it is affected on the left side, this can lead to transcortical motor aphasia.
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Blood enters the kidney through the which artery?
A. vas deferens
B. cardiac artery
C. renal artery
D. jugular
Answer:
|
|
sciq-6161
|
multiple_choice
|
Endotherms are warmed mostly by heat generated by what?
|
[
"metabolism",
"electricity",
"movement",
"the sun"
] |
A
|
Relavent Documents:
Document 0:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 1:::
A pre-STEM program is a course of study at any two-year college that prepares a student to transfer to a four-year school to earn a bachelor's degree in a STEM field.
Overview
The concept of a pre-STEM program is being developed to address America's need for more college-trained professionals in science, technology, engineering, and mathematics (STEM). It is an innovation meant to fill a gap at community colleges that do not have 'major' degree paths that students identify with on their way to earning an Associates degree. Students must complete a considerable amount of STEM coursework before transferring from a two-year school to a four-year school and earn a baccalaureate degree in a STEM field. Schools with a pre-STEM program are able to identify those students and support them with STEM-specific academic and career advising, increasing the student's chances of going on to earn a STEM baccalaureate degree in a timely fashion.
With over 50% of America's college-bound students starting their college career at public or private two-year school, and with a very small proportion of students who start college at a two-year school matriculating to and earning STEM degrees from four-year schools, pre-STEM programs have great potential for broadening participation in baccalaureate STEM studies.
Example programs
The effectiveness of pre-STEM programs is being investigated by a consortium of schools in Missouri: Moberly Area Community College, St. Charles Community College, Metropolitan Community College, and Truman State University.
A larger group of schools met at the Belknap Springs Meetings in October 2009 to discuss the challenges and opportunities presented by STEM-focused partnerships between 2-year and 4-year schools. Each program represented a two-year school and a four-year school that were trying to increase the number of people who earn a baccalaureate degree in a STEM area through various means, some of which were pre-STEM programs. Other methods includes
Document 2:::
Endothermic organisms known as homeotherms maintain internal temperatures with minimal metabolic regulation within a range of ambient temperatures called the thermal neutral zone (TNZ). Within the TNZ the basal rate of heat production is equal to the rate of heat loss to the environment. Homeothermic organisms adjust to the temperatures within the TNZ through different responses requiring little energy.
Environmental temperatures can cause fluctuations in a homeothermic organism's metabolic rate. This response is due to the energy required to maintain a relatively constant body temperature above ambient temperature by controlling heat loss and heat gain. The degree of this response depends not only on the species, but also on the levels of insulative and metabolic adaptation. Environmental temperatures below the TNZ, the lower critical temperature (LCT), require an organism to increase its metabolic rate to meet the environmental demands for heat. The Regulation about the TNZ requires metabolic heat production when the LCT is reached, as heat is lost to the environment. The organism reaches the LCT when the Ta (ambient temp.) decreases.
When an organism reaches this stage the metabolic rate increases significantly and thermogenesis increases the Tb (body temp.) If the Ta continues to decrease far below the LCT hypothermia occurs. Alternatively, evaporative heat loss for cooling occurs when temperatures above the TNZ, the upper critical zone (UCT), are realized (Speakman and Keijer 2013). When the Ta reaches too far above the UCT, the rate of heat gain and rate of heat production become higher than the rate of heat dissipation (heat loss through evaporative cooling), resulting in hyperthermia.
It can show postural changes where it changes its body shape or moves and exposes different areas to the sun/shade, and through radiation, convection and conduction, heat exchange occurs. Vasomotor responses allow control of the flow of blood between the periphery and the c
Document 3:::
The Science, Technology, Engineering and Mathematics Network or STEMNET is an educational charity in the United Kingdom that seeks to encourage participation at school and college in science and engineering-related subjects (science, technology, engineering, and mathematics) and (eventually) work.
History
It is based at Woolgate Exchange near Moorgate tube station in London and was established in 1996. The chief executive is Kirsten Bodley. The STEMNET offices are housed within the Engineering Council.
Function
Its chief aim is to interest children in science, technology, engineering and mathematics. Primary school children can start to have an interest in these subjects, leading secondary school pupils to choose science A levels, which will lead to a science career. It supports the After School Science and Engineering Clubs at schools. There are also nine regional Science Learning Centres.
STEM ambassadors
To promote STEM subjects and encourage young people to take up jobs in these areas, STEMNET have around 30,000 ambassadors across the UK. these come from a wide selection of the STEM industries and include TV personalities like Rob Bell.
Funding
STEMNET used to receive funding from the Department for Education and Skills. Since June 2007, it receives funding from the Department for Children, Schools and Families and Department for Innovation, Universities and Skills, since STEMNET sits on the chronological dividing point (age 16) of both of the new departments.
See also
The WISE Campaign
Engineering and Physical Sciences Research Council
National Centre for Excellence in Teaching Mathematics
Association for Science Education
Glossary of areas of mathematics
Glossary of astronomy
Glossary of biology
Glossary of chemistry
Glossary of engineering
Glossary of physics
Document 4:::
Thermofluids is a branch of science and engineering encompassing four intersecting fields:
Heat transfer
Thermodynamics
Fluid mechanics
Combustion
The term is a combination of "thermo", referring to heat, and "fluids", which refers to liquids, gases and vapors. Temperature, pressure, equations of state, and transport laws all play an important role in thermofluid problems. Phase transition and chemical reactions may also be important in a thermofluid context. The subject is sometimes also referred to as "thermal fluids".
Heat transfer
Heat transfer is a discipline of thermal engineering that concerns the transfer of thermal energy from one physical system to another. Heat transfer is classified into various mechanisms, such as heat conduction, convection, thermal radiation, and phase-change transfer. Engineers also consider the transfer of mass of differing chemical species, either cold or hot, to achieve heat transfer.
Sections include :
Energy transfer by heat, work and mass
Laws of thermodynamics
Entropy
Refrigeration Techniques
Properties and nature of pure substances
Applications
Engineering : Predicting and analysing the performance of machines
Thermodynamics
Thermodynamics is the science of energy conversion involving heat and other forms of energy, most notably mechanical work. It studies and interrelates the macroscopic variables, such as temperature, volume and pressure, which describe physical, thermodynamic systems.
Fluid mechanics
Fluid Mechanics the study of the physical forces at work during fluid flow. Fluid mechanics can be divided into fluid kinematics, the study of fluid motion, and fluid kinetics, the study of the effect of forces on fluid motion. Fluid mechanics can further be divided into fluid statics, the study of fluids at rest, and fluid dynamics, the study of fluids in motion. Some of its more interesting concepts include momentum and reactive forces in fluid flow and fluid machinery theory and performance.
Sections include:
Flu
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Endotherms are warmed mostly by heat generated by what?
A. metabolism
B. electricity
C. movement
D. the sun
Answer:
|
|
sciq-3982
|
multiple_choice
|
What do you call a type of mixture that has the same composition throughout?
|
[
"transfusion",
"structure",
"transition",
"solution"
] |
D
|
Relavent Documents:
Document 0:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 1:::
In chemistry, a mixture is a material made up of two or more different chemical substances which are not chemically bonded. A mixture is the physical combination of two or more substances in which the identities are retained and are mixed in the form of solutions, suspensions and colloids.
Mixtures are one product of mechanically blending or mixing chemical substances such as elements and compounds, without chemical bonding or other chemical change, so that each ingredient substance retains its own chemical properties and makeup. Despite the fact that there are no chemical changes to its constituents, the physical properties of a mixture, such as its melting point, may differ from those of the components. Some mixtures can be separated into their components by using physical (mechanical or thermal) means. Azeotropes are one kind of mixture that usually poses considerable difficulties regarding the separation processes required to obtain their constituents (physical or chemical processes or, even a blend of them).
Characteristics of mixtures
All mixtures can be characterized as being separable by mechanical means (e.g. purification, distillation, electrolysis, chromatography, heat, filtration, gravitational sorting, centrifugation). Mixtures differ from chemical compounds in the following ways:
the substances in a mixture can be separated using physical methods such as filtration, freezing, and distillation.
there is little or no energy change when a mixture forms (see Enthalpy of mixing).
The substances in a mixture keep its separate properties.
In the example of sand and water, neither one of the two substances changed in any way when they are mixed. Although the sand is in the water it still keeps the same properties that it had when it was outside the water.
mixtures have variable compositions, while compounds have a fixed, definite formula.
when mixed, individual substances keep their properties in a mixture, while if they form a compound their properties
Document 2:::
GRE Subject Biochemistry, Cell and Molecular Biology was a standardized exam provided by ETS (Educational Testing Service) that was discontinued in December 2016. It is a paper-based exam and there are no computer-based versions of it. ETS places this exam three times per year: once in April, once in October and once in November. Some graduate programs in the United States recommend taking this exam, while others require this exam score as a part of the application to their graduate programs. ETS sends a bulletin with a sample practice test to each candidate after registration for the exam. There are 180 questions within the biochemistry subject test.
Scores are scaled and then reported as a number between 200 and 990; however, in recent versions of the test, the maximum and minimum reported scores have been 760 (corresponding to the 99 percentile) and 320 (1 percentile) respectively. The mean score for all test takers from July, 2009, to July, 2012, was 526 with a standard deviation of 95.
After learning that test content from editions of the GRE® Biochemistry, Cell and Molecular Biology (BCM) Test has been compromised in Israel, ETS made the decision not to administer this test worldwide in 2016–17.
Content specification
Since many students who apply to graduate programs in biochemistry do so during the first half of their fourth year, the scope of most questions is largely that of the first three years of a standard American undergraduate biochemistry curriculum. A sampling of test item content is given below:
Biochemistry (36%)
A Chemical and Physical Foundations
Thermodynamics and kinetics
Redox states
Water, pH, acid-base reactions and buffers
Solutions and equilibria
Solute-solvent interactions
Chemical interactions and bonding
Chemical reaction mechanisms
B Structural Biology: Structure, Assembly, Organization and Dynamics
Small molecules
Macromolecules (e.g., nucleic acids, polysaccharides, proteins and complex lipids)
Supramolecular complexes (e.g.
Document 3:::
In physics, a dynamical system is said to be mixing if the phase space of the system becomes strongly intertwined, according to at least one of several mathematical definitions. For example, a measure-preserving transformation T is said to be strong mixing if
whenever A and B are any measurable sets and μ is the associated measure. Other definitions are possible, including weak mixing and topological mixing.
The mathematical definition of mixing is meant to capture the notion of physical mixing. A canonical example is the Cuba libre: suppose one is adding rum (the set A) to a glass of cola. After stirring the glass, the bottom half of the glass (the set B) will contain rum, and it will be in equal proportion as it is elsewhere in the glass. The mixing is uniform: no matter which region B one looks at, some of A will be in that region. A far more detailed, but still informal description of mixing can be found in the article on mixing (mathematics).
Every mixing transformation is ergodic, but there are ergodic transformations which are not mixing.
Physical mixing
The mixing of gases or liquids is a complex physical process, governed by a convective diffusion equation that may involve non-Fickian diffusion as in spinodal decomposition. The convective portion of the governing equation contains fluid motion terms that are governed by the Navier–Stokes equations. When fluid properties such as viscosity depend on composition, the governing equations may be coupled. There may also be temperature effects. It is not clear that fluid mixing processes are mixing in the mathematical sense.
Small rigid objects (such as rocks) are sometimes mixed in a rotating drum or tumbler. The 1969 Selective Service draft lottery was carried out by mixing plastic capsules which contained a slip of paper (marked with a day of the year).
See also
Miscibility
Document 4:::
Further Mathematics is the title given to a number of advanced secondary mathematics courses. The term "Higher and Further Mathematics", and the term "Advanced Level Mathematics", may also refer to any of several advanced mathematics courses at many institutions.
In the United Kingdom, Further Mathematics describes a course studied in addition to the standard mathematics AS-Level and A-Level courses. In the state of Victoria in Australia, it describes a course delivered as part of the Victorian Certificate of Education (see § Australia (Victoria) for a more detailed explanation). Globally, it describes a course studied in addition to GCE AS-Level and A-Level Mathematics, or one which is delivered as part of the International Baccalaureate Diploma.
In other words, more mathematics can also be referred to as part of advanced mathematics, or advanced level math.
United Kingdom
Background
A qualification in Further Mathematics involves studying both pure and applied modules. Whilst the pure modules (formerly known as Pure 4–6 or Core 4–6, now known as Further Pure 1–3, where 4 exists for the AQA board) build on knowledge from the core mathematics modules, the applied modules may start from first principles.
The structure of the qualification varies between exam boards.
With regard to Mathematics degrees, most universities do not require Further Mathematics, and may incorporate foundation math modules or offer "catch-up" classes covering any additional content. Exceptions are the University of Warwick, the University of Cambridge which requires Further Mathematics to at least AS level; University College London requires or recommends an A2 in Further Maths for its maths courses; Imperial College requires an A in A level Further Maths, while other universities may recommend it or may promise lower offers in return. Some schools and colleges may not offer Further mathematics, but online resources are available
Although the subject has about 60% of its cohort obtainin
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What do you call a type of mixture that has the same composition throughout?
A. transfusion
B. structure
C. transition
D. solution
Answer:
|
|
sciq-3240
|
multiple_choice
|
A machine is any device that makes work easier by doing what?
|
[
"removing barriers",
"changing a force",
"changing molecules",
"moving things"
] |
B
|
Relavent Documents:
Document 0:::
A machine is a physical system using power to apply forces and control movement to perform an action. The term is commonly applied to artificial devices, such as those employing engines or motors, but also to natural biological macromolecules, such as molecular machines. Machines can be driven by animals and people, by natural forces such as wind and water, and by chemical, thermal, or electrical power, and include a system of mechanisms that shape the actuator input to achieve a specific application of output forces and movement. They can also include computers and sensors that monitor performance and plan movement, often called mechanical systems.
Renaissance natural philosophers identified six simple machines which were the elementary devices that put a load into motion, and calculated the ratio of output force to input force, known today as mechanical advantage.
Modern machines are complex systems that consist of structural elements, mechanisms and control components and include interfaces for convenient use. Examples include: a wide range of vehicles, such as trains, automobiles, boats and airplanes; appliances in the home and office, including computers, building air handling and water handling systems; as well as farm machinery, machine tools and factory automation systems and robots.
Etymology
The English word machine comes through Middle French from Latin , which in turn derives from the Greek (Doric , Ionic 'contrivance, machine, engine', a derivation from 'means, expedient, remedy'). The word mechanical (Greek: ) comes from the same Greek roots. A wider meaning of 'fabric, structure' is found in classical Latin, but not in Greek usage. This meaning is found in late medieval French, and is adopted from the French into English in the mid-16th century.
In the 17th century, the word machine could also mean a scheme or plot, a meaning now expressed by the derived machination. The modern meaning develops out of specialized application of the term to st
Document 1:::
In electrical engineering, electric machine is a general term for machines using electromagnetic forces, such as electric motors, electric generators, and others. They are electromechanical energy converters: an electric motor converts electricity to mechanical power while an electric generator converts mechanical power to electricity. The moving parts in a machine can be rotating (rotating machines) or linear (linear machines). Besides motors and generators, a third category often included is transformers, which although they do not have any moving parts are also energy converters, changing the voltage level of an alternating current.
Electric machines, in the form of synchronous and induction generators, produce about 95% of all electric power on Earth (as of early 2020s), and in the form of electric motors consume approximately 60% of all electric power produced. Electric machines were developed beginning in the mid 19th century and since that time have been a ubiquitous component of the infrastructure. Developing more efficient electric machine technology is crucial to any global conservation, green energy, or alternative energy strategy.
Generator
An electric generator is a device that converts mechanical energy to electrical energy. A generator forces electrons to flow through an external electrical circuit. It is somewhat analogous to a water pump, which creates a flow of water but does not create the water inside. The source of mechanical energy, the prime mover, may be a reciprocating or turbine steam engine, water falling through a turbine or waterwheel, an internal combustion engine, a wind turbine, a hand crank, compressed air or any other source of mechanical energy.
The two main parts of an electrical machine can be described in either mechanical or electrical terms. In mechanical terms, the rotor is the rotating part, and the stator is the stationary part of an electrical machine. In electrical terms, the armature is the power-producing compo
Document 2:::
The machine industry or machinery industry is a subsector of the industry, that produces and maintains machines for consumers, the industry, and most other companies in the economy.
This machine industry traditionally belongs to the heavy industry. Nowadays, many smaller companies in this branch are considered part of the light industry. Most manufacturers in the machinery industry are called machine factories.
Overview
The machine industry is a subsector of the industry that produces a range of products from power tools, different types of machines, and domestic technology to factory equipment etc. On the one hand the machine industry provides:
The means of production for businesses in the agriculture, mining, industry and construction.
The means of production for public utility, such as equipment for the production and distribution of gas, electricity and water.
A range of supporting equipment for all sectors of the economy, such as equipment for heating, ventilation, and air conditioning of buildings.
These means of production are called capital goods, because a certain amount of capital is invested. Much of those production machines require regular maintenance, which becomes supplied specialized companies in the machine industry.
On the other end the machinery industry supplies consumer goods, including kitchen appliances, refrigerators, washers, dryers and a like. Production of radio and television, however, is generally considered belonging to the electrical equipment industry. The machinery industry itself is a major customer of the steel industry.
The production of the machinery industry varies widely from single-unit production and series production to mass production. Single-unit production is about constructing unique products, which are specified in specific customer requirements. Due to modular design such devices and machines can often be manufactured in small series, which significantly reduces the costs. From a certain stage in the production
Document 3:::
Machine element or hardware refers to an elementary component of a machine. These elements consist of three basic types:
structural components such as frame members, bearings, axles, splines, fasteners, seals, and lubricants,
mechanisms that control movement in various ways such as gear trains, belt or chain drives, linkages, cam and follower systems, including brakes and clutches, and
control components such as buttons, switches, indicators, sensors, actuators and computer controllers.
While generally not considered to be a machine element, the shape, texture and color of covers are an important part of a machine that provide a styling and operational interface between the mechanical components of a machine and its users.
Machine elements are basic mechanical parts and features used as the building blocks of most machines. Most are standardized to common sizes, but customs are also common for specialized applications.
Machine elements may be features of a part (such as screw threads or integral plain bearings) or they may be discrete parts in and of themselves such as wheels, axles, pulleys, rolling-element bearings, or gears. All of the simple machines may be described as machine elements, and many machine elements incorporate concepts of one or more simple machines. For example, a leadscrew incorporates a screw thread, which is an inclined plane wrapped around a cylinder.
Many mechanical design, invention, and engineering tasks involve a knowledge of various machine elements and an intelligent and creative combining of these elements into a component or assembly that fills a need (serves an application).
Structural elements
Beams,
Struts,
Bearings,
Fasteners
Keys,
Splines,
Cotter pin,
Seals
Machine guardings
Mechanical elements
Engine,
Electric motor,
Actuator,
Shafts,
Couplings
Belt,
Chain,
Cable drives,
Gear train,
Clutch,
Brake,
Flywheel,
Cam,
follower systems,
Linkage,
Simple machine
Types
Shafts
Document 4:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
A machine is any device that makes work easier by doing what?
A. removing barriers
B. changing a force
C. changing molecules
D. moving things
Answer:
|
|
sciq-1124
|
multiple_choice
|
What are the building blocks of peptides?
|
[
"rocks",
"amino acids",
"alkali",
"magnets"
] |
B
|
Relavent Documents:
Document 0:::
This is a list of topics in molecular biology. See also index of biochemistry articles.
Document 1:::
Bioorganic chemistry is a scientific discipline that combines organic chemistry and biochemistry. It is that branch of life science that deals with the study of biological processes using chemical methods. Protein and enzyme function are examples of these processes.
Sometimes biochemistry is used interchangeably for bioorganic chemistry; the distinction being that bioorganic chemistry is organic chemistry that is focused on the biological aspects. While biochemistry aims at understanding biological processes using chemistry, bioorganic chemistry attempts to expand organic-chemical researches (that is, structures, synthesis, and kinetics) toward biology. When investigating metalloenzymes and cofactors, bioorganic chemistry overlaps bioinorganic chemistry.
Sub disciplines
Biophysical organic chemistry is a term used when attempting to describe intimate details of molecular recognition by bioorganic chemistry.
Natural product chemistry is the process of Identifying compounds found in nature to determine their properties. Compound discoveries have and often lead to medicinal uses, development of herbicides and insecticides.
Document 2:::
In organic chemistry, peptide synthesis is the production of peptides, compounds where multiple amino acids are linked via amide bonds, also known as peptide bonds. Peptides are chemically synthesized by the condensation reaction of the carboxyl group of one amino acid to the amino group of another. Protecting group strategies are usually necessary to prevent undesirable side reactions with the various amino acid side chains. Chemical peptide synthesis most commonly starts at the carboxyl end of the peptide (C-terminus), and proceeds toward the amino-terminus (N-terminus). Protein biosynthesis (long peptides) in living organisms occurs in the opposite direction.
The chemical synthesis of peptides can be carried out using classical solution-phase techniques, although these have been replaced in most research and development settings by solid-phase methods (see below). Solution-phase synthesis retains its usefulness in large-scale production of peptides for industrial purposes moreover.
Chemical synthesis facilitates the production of peptides that are difficult to express in bacteria, the incorporation of unnatural amino acids, peptide/protein backbone modification, and the synthesis of D-proteins, which consist of D-amino acids.
Solid-phase synthesis
The established method for the production of synthetic peptides in the lab is known as solid phase peptide synthesis (SPPS). Pioneered by Robert Bruce Merrifield, SPPS allows the rapid assembly of a peptide chain through successive reactions of amino acid derivatives on a macroscopically insoluble solvent-swollen beaded resin support.
The solid support consists of small, polymeric resin beads functionalized with reactive groups (such as amine or hydroxyl groups) that link to the nascent peptide chain. Since the peptide remains covalently attached to the support throughout the synthesis, excess reagents and side products can be removed by washing and filtration. This approach circumvents the comparatively time-consu
Document 3:::
Peptide computing is a form of computing which uses peptides, instead of traditional electronic components. The basis of this computational model is the affinity of antibodies towards peptide sequences. Similar to DNA computing, the parallel interactions of peptide sequences and antibodies have been used by this model to solve a few NP-complete problems. Specifically, the hamiltonian path problem (HPP) and some versions of the set cover problem are a few NP-complete problems which have been solved using this computational model so far. This model of computation has also been shown to be computationally universal (or Turing complete).
This model of computation has some critical advantages over DNA computing. For instance, while DNA is made of four building blocks, peptides are made of twenty building blocks. The peptide-antibody interactions are also more flexible with respect to recognition and affinity than an interaction between a DNA strand and its reverse complement. However, unlike DNA computing, this model is yet to be practically realized. The main limitation is the availability of specific monoclonal antibodies required by the model.
See also
Biocomputers
Computational gene
Computational complexity theory
DNA computing
Molecular electronics
Parallel computing
Unconventional computing
Molecular logic gate
Document 4:::
A biomolecule or biological molecule is a loosely used term for molecules present in organisms that are essential to one or more typically biological processes, such as cell division, morphogenesis, or development. Biomolecules include the primary metabolites which are large macromolecules (or polyelectrolytes) such as proteins, carbohydrates, lipids, and nucleic acids, as well as small molecules such as vitamins and hormones. A more general name for this class of material is biological materials. Biomolecules are an important element of living organisms, those biomolecules are often endogenous, produced within the organism but organisms usually need exogenous biomolecules, for example certain nutrients, to survive.
Biology and its subfields of biochemistry and molecular biology study biomolecules and their reactions. Most biomolecules are organic compounds, and just four elements—oxygen, carbon, hydrogen, and nitrogen—make up 96% of the human body's mass. But many other elements, such as the various biometals, are also present in small amounts.
The uniformity of both specific types of molecules (the biomolecules) and of certain metabolic pathways are invariant features among the wide diversity of life forms; thus these biomolecules and metabolic pathways are referred to as "biochemical universals" or "theory of material unity of the living beings", a unifying concept in biology, along with cell theory and evolution theory.
Types of biomolecules
A diverse range of biomolecules exist, including:
Small molecules:
Lipids, fatty acids, glycolipids, sterols, monosaccharides
Vitamins
Hormones, neurotransmitters
Metabolites
Monomers, oligomers and polymers:
Nucleosides and nucleotides
Nucleosides are molecules formed by attaching a nucleobase to a ribose or deoxyribose ring. Examples of these include cytidine (C), uridine (U), adenosine (A), guanosine (G), and thymidine (T).
Nucleosides can be phosphorylated by specific kinases in the cell, producing nucl
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What are the building blocks of peptides?
A. rocks
B. amino acids
C. alkali
D. magnets
Answer:
|
|
sciq-8504
|
multiple_choice
|
Energy from sunlight is absorbed by what pigment in the thylakoid membrane of chloroplasts?
|
[
"tannin",
"chlorophyll",
"melatonin",
"keratin"
] |
B
|
Relavent Documents:
Document 0:::
The photosynthetic efficiency is the fraction of light energy converted into chemical energy during photosynthesis in green plants and algae. Photosynthesis can be described by the simplified chemical reaction
6 H2O + 6 CO2 + energy → C6H12O6 + 6 O2
where C6H12O6 is glucose (which is subsequently transformed into other sugars, starches, cellulose, lignin, and so forth). The value of the photosynthetic efficiency is dependent on how light energy is defined – it depends on whether we count only the light that is absorbed, and on what kind of light is used (see Photosynthetically active radiation). It takes eight (or perhaps ten or more) photons to use one molecule of CO2. The Gibbs free energy for converting a mole of CO2 to glucose is 114 kcal, whereas eight moles of photons of wavelength 600 nm contains 381 kcal, giving a nominal efficiency of 30%. However, photosynthesis can occur with light up to wavelength 720 nm so long as there is also light at wavelengths below 680 nm to keep Photosystem II operating (see Chlorophyll). Using longer wavelengths means less light energy is needed for the same number of photons and therefore for the same amount of photosynthesis. For actual sunlight, where only 45% of the light is in the photosynthetically active wavelength range, the theoretical maximum efficiency of solar energy conversion is approximately 11%. In actuality, however, plants do not absorb all incoming sunlight (due to reflection, respiration requirements of photosynthesis and the need for optimal solar radiation levels) and do not convert all harvested energy into biomass, which results in a maximum overall photosynthetic efficiency of 3 to 6% of total solar radiation. If photosynthesis is inefficient, excess light energy must be dissipated to avoid damaging the photosynthetic apparatus. Energy can be dissipated as heat (non-photochemical quenching), or emitted as chlorophyll fluorescence.
Typical efficiencies
Plants
Quoted values sunlight-to-biomass efficien
Document 1:::
Retinal (also known as retinaldehyde) is a polyene chromophore. Retinal, bound to proteins called opsins, is the chemical basis of visual phototransduction, the light-detection stage of visual perception (vision).
Some microorganisms use retinal to convert light into metabolic energy. In fact, a recent study suggests most living organisms on our planet ~3 billion years ago used retinal to convert sunlight into energy rather than chlorophyll. Since retinal absorbs mostly green light and transmits purple light, this gave rise to the Purple Earth Hypothesis.
There are many forms of vitamin A — all of which are converted to retinal, which cannot be made without them. Retinal itself is considered to be a form of vitamin A when eaten by an animal. The number of different molecules that can be converted to retinal varies from species to species. Retinal was originally called retinene, and was renamed after it was discovered to be vitamin A aldehyde.
Vertebrate animals ingest retinal directly from meat, or they produce retinal from carotenoids — either from α-carotene or β-carotene — both of which are carotenes. They also produce it from β-cryptoxanthin, a type of xanthophyll. These carotenoids must be obtained from plants or other photosynthetic organisms. No other carotenoids can be converted by animals to retinal. Some carnivores cannot convert any carotenoids at all. The other main forms of vitamin A — retinol and a partially active form, retinoic acid — may both be produced from retinal.
Invertebrates such as insects and squid use hydroxylated forms of retinal in their visual systems, which derive from conversion from other xanthophylls.
Vitamin A metabolism
Living organisms produce retinal by irreversible oxidative cleavage of carotenoids.
For example:
beta-carotene + O2 → 2 retinal,
catalyzed by a beta-carotene 15,15'-monooxygenase or a beta-carotene 15,15'-dioxygenase.
Just as carotenoids are the precursors of retinal, retinal is the precursor of the other
Document 2:::
Thylakoids are membrane-bound compartments inside chloroplasts and cyanobacteria. They are the site of the light-dependent reactions of photosynthesis. Thylakoids consist of a thylakoid membrane surrounding a thylakoid lumen. Chloroplast thylakoids frequently form stacks of disks referred to as grana (singular: granum). Grana are connected by intergranal or stromal thylakoids, which join granum stacks together as a single functional compartment.
In thylakoid membranes, chlorophyll pigments are found in packets called quantasomes. Each quantasome contains 230 to 250 chlorophyll molecules.
Etymology
The word Thylakoid comes from the Greek word thylakos or θύλακος, meaning "sac" or "pouch". Thus, thylakoid means "sac-like" or "pouch-like".
Structure
Thylakoids are membrane-bound structures embedded in the chloroplast stroma. A stack of thylakoids is called a granum and resembles a stack of coins.
Membrane
The thylakoid membrane is the site of the light-dependent reactions of photosynthesis with the photosynthetic pigments embedded directly in the membrane. It is an alternating pattern of dark and light bands measuring each 1 nanometre. The thylakoid lipid bilayer shares characteristic features with prokaryotic membranes and the inner chloroplast membrane. For example, acidic lipids can be found in thylakoid membranes, cyanobacteria and other photosynthetic bacteria and are involved in the functional integrity of the photosystems. The thylakoid membranes of higher plants are composed primarily of phospholipids and galactolipids that are asymmetrically arranged along and across the membranes. Thylakoid membranes are richer in galactolipids rather than phospholipids; also they predominantly consist of hexagonal phase II forming monogalacotosyl diglyceride lipid. Despite this unique composition, plant thylakoid membranes have been shown to assume largely lipid-bilayer dynamic organization. Lipids forming the thylakoid membranes, richest in high-fluidity linolenic acid
Document 3:::
The light-harvesting complex (or antenna complex; LH or LHC) is an array of protein and chlorophyll molecules embedded in the thylakoid membrane of plants and cyanobacteria, which transfer light energy to one chlorophyll a molecule at the reaction center of a photosystem.
The antenna pigments are predominantly chlorophyll b, xanthophylls, and carotenes. Chlorophyll a is known as the core pigment. Their absorption spectra are non-overlapping and broaden the range of light that can be absorbed in photosynthesis. The carotenoids have another role as an antioxidant to prevent photo-oxidative damage of chlorophyll molecules. Each antenna complex has between 250 and 400 pigment molecules and the energy they absorb is shuttled by resonance energy transfer to a specialized chlorophyll-protein complex known as the reaction center of each photosystem. The reaction center initiates a complex series of chemical reactions that capture energy in the form of chemical bonds.
For photosystem II, when either of the two chlorophyll a molecules at the reaction center absorb energy, an electron is excited and transferred to an electron acceptor molecule, pheophytin, leaving the chlorophyll a in an oxidized state. The oxidised chlorophyll a replaces the electrons by photolysis that involves the oxidation of water molecules to oxygen, protons and electrons.
The N-terminus of the chlorophyll a-b binding protein extends into the stroma where it is involved with adhesion of granal membranes and photo-regulated by reversible phosphorylation of its threonine residues. Both these processes are believed to mediate the distribution of excitation energy between photosystems I and II.
This family also includes the photosystem II protein PsbS, which plays a role in energy-dependent quenching that increases thermal dissipation of excess absorbed light energy in the photosystem.
LH 1
Light-harvesting complex I is permanently bound to photosystem I via the plant-specific subunit PsaG. It is made u
Document 4:::
Disc shedding is the process by which photoreceptor cells in the retina are renewed. The disc formations in the outer segment of photoreceptors, which contain the photosensitive opsins, are completely renewed every ten days.
Photoreceptors
The retina contains two types of photoreceptor – rod cells and cone cells. There are about 6-7 million cones that mediate photopic vision, and they are concentrated in the macula at the center of the retina. There are about 120 million rods that are more sensitive than the cones and therefore mediate scotopic vision.
A vertebrate's photoreceptors are divided into three parts:
an outer segment that contains the photosensitive opsins
an inner segment that contains the cell's metabolic machinery (endoplasmic reticulum, Golgi complex, ribosomes, mitochondria)
a synaptic terminal at which contacts with second-order neurons of the retina are made
Discs
The photosensitive outer segment consists of a series of discrete membranous discs .
While in the rod, these discs lack any direct connection to the surface membrane (with the exception of a few recently formed basal discs that remain in continuity with the surface), the cone's photosensitive membrane is continuous with the surface membrane. The outer segment (OS) discs are densely packed with rhodopsin for high-sensitivity light detection. These discs are completely replaced once every ten days and this continuous renewal continues throughout the lifetime of the sighted animal.
After the opsins are synthesized, they fuse to the plasma membrane, which then invaginates with discs budding off internally, forming the tightly packed stacks of outer segment discs. From translation of opsin to formation of the discs takes just a couple of hours.
Shedding
Disc shedding was first described by RW Young in 1967. Discs mature along with their distal migration; aged discs shed at the distal tip and are engulfed by the neighboring retinal pigment epithelial (RPE) cells for degradation.
One e
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Energy from sunlight is absorbed by what pigment in the thylakoid membrane of chloroplasts?
A. tannin
B. chlorophyll
C. melatonin
D. keratin
Answer:
|
|
sciq-5149
|
multiple_choice
|
Called the powerhouses of the cell, what organelles are the site of cellular energy production?
|
[
"mitochondria",
"cell wall",
"nucleus",
"mitosis"
] |
A
|
Relavent Documents:
Document 0:::
Cell physiology is the biological study of the activities that take place in a cell to keep it alive. The term physiology refers to normal functions in a living organism. Animal cells, plant cells and microorganism cells show similarities in their functions even though they vary in structure.
General characteristics
There are two types of cells: prokaryotes and eukaryotes.
Prokaryotes were the first of the two to develop and do not have a self-contained nucleus. Their mechanisms are simpler than later-evolved eukaryotes, which contain a nucleus that envelops the cell's DNA and some organelles.
Prokaryotes
Prokaryotes have DNA located in an area called the nucleoid, which is not separated from other parts of the cell by a membrane. There are two domains of prokaryotes: bacteria and archaea. Prokaryotes have fewer organelles than eukaryotes. Both have plasma membranes and ribosomes (structures that synthesize proteins and float free in cytoplasm). Two unique characteristics of prokaryotes are fimbriae (finger-like projections on the surface of a cell) and flagella (threadlike structures that aid movement).
Eukaryotes
Eukaryotes have a nucleus where DNA is contained. They are usually larger than prokaryotes and contain many more organelles. The nucleus, the feature of a eukaryote that distinguishes it from a prokaryote, contains a nuclear envelope, nucleolus and chromatin. In cytoplasm, endoplasmic reticulum (ER) synthesizes membranes and performs other metabolic activities. There are two types, rough ER (containing ribosomes) and smooth ER (lacking ribosomes). The Golgi apparatus consists of multiple membranous sacs, responsible for manufacturing and shipping out materials such as proteins. Lysosomes are structures that use enzymes to break down substances through phagocytosis, a process that comprises endocytosis and exocytosis. In the mitochondria, metabolic processes such as cellular respiration occur. The cytoskeleton is made of fibers that support the str
Document 1:::
Cellular components are the complex biomolecules and structures of which cells, and thus living organisms, are composed. Cells are the structural and functional units of life. The smallest organisms are single cells, while the largest organisms are assemblages of trillions of cells. DNA, double stranded macromolecule that carries the hereditary information of the cell and found in all living cells; each cell carries chromosome(s) having a distinctive DNA sequence.
Examples include macromolecules such as proteins and nucleic acids, biomolecular complexes such as a ribosome, and structures such as membranes, and organelles. While the majority of cellular components are located within the cell itself, some may exist in extracellular areas of an organism.
Cellular components may also be called biological matter or biological material. Most biological matter has the characteristics of soft matter, being governed by relatively small energies. All known life is made of biological matter. To be differentiated from other theoretical or fictional life forms, such life may be called carbon-based, cellular, organic, biological, or even simply living – as some definitions of life exclude hypothetical types of biochemistry.
See also
Cell (biology)
Cell biology
Biomolecule
Organelle
Tissue (biology)
External links
https://web.archive.org/web/20130918033010/http://bioserv.fiu.edu/~walterm/FallSpring/review1_fall05_chap_cell3.htm
Document 2:::
This lecture, named in memory of Keith R. Porter, is presented to an eminent cell biologist each year at the ASCB Annual Meeting. The ASCB Program Committee and the ASCB President recommend the Porter Lecturer to the Porter Endowment each year.
Lecturers
Source: ASCB
See also
List of biology awards
Document 3:::
Organelle biogenesis is the biogenesis, or creation, of cellular organelles in cells. Organelle biogenesis includes the process by which cellular organelles are split between daughter cells during mitosis; this process is called organelle inheritance.
Discovery
Following the discovery of cellular organelles in the nineteenth century, little was known about their function and synthesis until the development of electron microscopy and subcellular fractionation in the twentieth century. This allowed experiments on the function, structure, and biogenesis of these organelles to commence.
Mechanisms of protein sorting and retrieval have been found to give organelles their characteristic composition. It is known that cellular organelles can come from preexisting organelles; however, it is a subject of controversy whether organelles can be created without a preexisting one.
Process
Several processes are known to have developed for organelle biogenesis. These can range from de novo synthesis to the copying of a template organelle; the formation of an organelle 'from scratch' and using a preexisting organelle as a template to manufacture an organelle, respectively. The distinct structures of each organelle are thought to be caused by the different mechanisms of the processes which create them and the proteins that they are made up of. Organelles may also be 'split' between two cells during the process of cellular division (known as organelle inheritance), where the organelle of the parent cell doubles in size and then splits with each half being delivered to their respective daughter cells.
The process of organelle biogenesis is known to be regulated by specialized transcription networks that modulate the expression of the genes that code for specific organellar proteins. In order for organelle biogenesis to be carried out properly, the specific genes coding for the organellar proteins must be transcribed properly and the translation of the resulting mRNA must be succes
Document 4:::
The cell is the basic structural and functional unit of all forms of life. Every cell consists of cytoplasm enclosed within a membrane, and contains many macromolecules such as proteins, DNA and RNA, as well as many small molecules of nutrients and metabolites. The term comes from the Latin word meaning 'small room'.
Cells can acquire specified function and carry out various tasks within the cell such as replication, DNA repair, protein synthesis, and motility. Cells are capable of specialization and mobility within the cell.
Most plant and animal cells are only visible under a light microscope, with dimensions between 1 and 100 micrometres. Electron microscopy gives a much higher resolution showing greatly detailed cell structure. Organisms can be classified as unicellular (consisting of a single cell such as bacteria) or multicellular (including plants and animals). Most unicellular organisms are classed as microorganisms.
The study of cells and how they work has led to many other studies in related areas of biology, including: discovery of DNA, cancer systems biology, aging and developmental biology.
Cell biology is the study of cells, which were discovered by Robert Hooke in 1665, who named them for their resemblance to cells inhabited by Christian monks in a monastery. Cell theory, first developed in 1839 by Matthias Jakob Schleiden and Theodor Schwann, states that all organisms are composed of one or more cells, that cells are the fundamental unit of structure and function in all living organisms, and that all cells come from pre-existing cells. Cells emerged on Earth about 4 billion years ago.
Discovery
With continual improvements made to microscopes over time, magnification technology became advanced enough to discover cells. This discovery is largely attributed to Robert Hooke, and began the scientific study of cells, known as cell biology. When observing a piece of cork under the scope, he was able to see pores. This was shocking at the time as i
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Called the powerhouses of the cell, what organelles are the site of cellular energy production?
A. mitochondria
B. cell wall
C. nucleus
D. mitosis
Answer:
|
|
sciq-3162
|
multiple_choice
|
What term is used to describe structures that are similar in unrelated organisms?
|
[
"isolated",
"analogous",
"reproductive",
"symbolic"
] |
B
|
Relavent Documents:
Document 0:::
In mathematics, the notion of a germ of an object in/on a topological space is an equivalence class of that object and others of the same kind that captures their shared local properties. In particular, the objects in question are mostly functions (or maps) and subsets. In specific implementations of this idea, the functions or subsets in question will have some property, such as being analytic or smooth, but in general this is not needed (the functions in question need not even be continuous); it is however necessary that the space on/in which the object is defined is a topological space, in order that the word local has some meaning.
Name
The name is derived from cereal germ in a continuation of the sheaf metaphor, as a germ is (locally) the "heart" of a function, as it is for a grain.
Formal definition
Basic definition
Given a point x of a topological space X, and two maps (where Y is any set), then and define the same germ at x if there is a neighbourhood U of x such that restricted to U, f and g are equal; meaning that for all u in U.
Similarly, if S and T are any two subsets of X, then they define the same germ at x if there is again a neighbourhood U of x such that
It is straightforward to see that defining the same germ at x is an equivalence relation (be it on maps or sets), and the equivalence classes are called germs (map-germs, or set-germs accordingly). The equivalence relation is usually written
Given a map f on X, then its germ at x is usually denoted [f ]x. Similarly, the germ at x of a set S is written [S]x. Thus,
A map germ at x in X that maps the point x in X to the point y in Y is denoted as
When using this notation, f is then intended as an entire equivalence class of maps, using the same letter f for any representative map.
Notice that two sets are germ-equivalent at x if and only if their characteristic functions are germ-equivalent at x:
More generally
Maps need not be defined on all of X, and in particular they don't need to
Document 1:::
A cell type is a classification used to identify cells that share morphological or phenotypical features. A multicellular organism may contain cells of a number of widely differing and specialized cell types, such as muscle cells and skin cells, that differ both in appearance and function yet have identical genomic sequences. Cells may have the same genotype, but belong to different cell types due to the differential regulation of the genes they contain. Classification of a specific cell type is often done through the use of microscopy (such as those from the cluster of differentiation family that are commonly used for this purpose in immunology). Recent developments in single cell RNA sequencing facilitated classification of cell types based on shared gene expression patterns. This has led to the discovery of many new cell types in e.g. mouse cortex, hippocampus, dorsal root ganglion and spinal cord.
Animals have evolved a greater diversity of cell types in a multicellular body (100–150 different cell types), compared
with 10–20 in plants, fungi, and protists. The exact number of cell types is, however, undefined, and the Cell Ontology, as of 2021, lists over 2,300 different cell types.
Multicellular organisms
All higher multicellular organisms contain cells specialised for different functions. Most distinct cell types arise from a single totipotent cell that differentiates into hundreds of different cell types during the course of development. Differentiation of cells is driven by different environmental cues (such as cell–cell interaction) and intrinsic differences (such as those caused by the uneven distribution of molecules during division). Multicellular organisms are composed of cells that fall into two fundamental types: germ cells and somatic cells. During development, somatic cells will become more specialized and form the three primary germ layers: ectoderm, mesoderm, and endoderm. After formation of the three germ layers, cells will continue to special
Document 2:::
An equivalence group is a set of unspecified cells that have the same developmental potential or ability to adopt various fates. Our current understanding suggests that equivalence groups are limited to cells of the same ancestry, also known as sibling cells. Often, cells of an equivalence group adopt different fates from one another.
Equivalence groups assume various potential fates in two general, non-mutually exclusive ways. One mechanism, induction, occurs when a signal originating from outside of the equivalence group specifies a subset of the naïve cells. Another mode, known as lateral inhibition, arises when a signal within an equivalence group causes one cell to adopt a dominant fate while others in the group are inhibited from doing so. In many examples of equivalence groups, both induction and lateral inhibition are used to define patterns of distinct cell types.
Cells of an equivalence group that do not receive a signal adopt a default fate. Alternatively, cells that receive a signal take on different fates. At a certain point, the fates of cells within an equivalence group become irreversibly determined, thus they lose their multipotent potential. The following provides examples of equivalence groups studied in nematodes and ascidians.
Vulva Precursor Cell Equivalence Group
Introduction
A classic example of an equivalence group is the vulva precursor cells (VPCs) of nematodes. In Caenorhabditis elegans self-fertilized eggs exit the body through the vulva. This organ develops from a subset of cell of an equivalence group consisting of six VPCs, P3.p-P8.p, which lie ventrally along the anterior-posterior axis. In this example a single overlying somatic cells, the anchor cell, induces nearby VPCs to take on vulva fates 1° (P6.p) and 2° (P5.p and P7.p). VPCs that are not induced form the 3° lineage (P3.p, P4.p and P8.p), which make epidermal cells that fuse to a large syncytial epidermis (see image).
The six VPCs form an equivalence group beca
Document 3:::
The branches of science known informally as omics are various disciplines in biology whose names end in the suffix -omics, such as genomics, proteomics, metabolomics, metagenomics, phenomics and transcriptomics. Omics aims at the collective characterization and quantification of pools of biological molecules that translate into the structure, function, and dynamics of an organism or organisms.
The related suffix -ome is used to address the objects of study of such fields, such as the genome, proteome or metabolome respectively. The suffix -ome as used in molecular biology refers to a totality of some sort; it is an example of a "neo-suffix" formed by abstraction from various Greek terms in , a sequence that does not form an identifiable suffix in Greek.
Functional genomics aims at identifying the functions of as many genes as possible of a given organism. It combines
different -omics techniques such as transcriptomics and proteomics with saturated mutant collections.
Origin
The Oxford English Dictionary (OED) distinguishes three different fields of application for the -ome suffix:
in medicine, forming nouns with the sense "swelling, tumour"
in botany or zoology, forming nouns in the sense "a part of an animal or plant with a specified structure"
in cellular and molecular biology, forming nouns with the sense "all constituents considered collectively"
The -ome suffix originated as a variant of -oma, and became productive in the last quarter of the 19th century. It originally appeared in terms like sclerome or rhizome. All of these terms derive from Greek words in , a sequence that is not a single suffix, but analyzable as , the belonging to the word stem (usually a verb) and the being a genuine Greek suffix forming abstract nouns.
The OED suggests that its third definition originated as a back-formation from mitome, Early attestations include biome (1916) and genome (first coined as German Genom in 1920).
The association with chromosome in molecular bio
Document 4:::
Serial homology is a special type of homology, defined by Owen as "representative or repetitive relation in the segments of the same organism." Ernst Haeckel preferred the term "homotypy" for the same phenomenon.
Classical examples of serial homologies are the development of forelimbs and hind limbs of tetrapods and the iterative structure of the vertebrae.
See also
Deep homology
Evolutionary developmental biology
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What term is used to describe structures that are similar in unrelated organisms?
A. isolated
B. analogous
C. reproductive
D. symbolic
Answer:
|
|
sciq-5068
|
multiple_choice
|
Some plants open their leaves during the day to collect what?
|
[
"moisture",
"energy",
"sunlight",
"precipitation"
] |
C
|
Relavent Documents:
Document 0:::
What a Plant Knows is a popular science book by Daniel Chamovitz, originally published in 2012, discussing the sensory system of plants. A revised edition was published in 2017.
Release details / Editions / Publication
Hardcover edition, 2012
Paperback version, 2013
Revised edition, 2017
What a Plant Knows has been translated and published in a number of languages.
Document 1:::
Maintenance respiration (or maintenance energy) refers to metabolism occurring in an organism that is needed to maintain that organism in a healthy, living state. Maintenance respiration contrasts with growth respiration, which is responsible for the synthesis of new structures in growth, nutrient uptake, nitrogen (N) reduction and phloem loading, whereas maintenance respiration is associated with protein and membrane turnover and maintenance of ion concentrations and gradients.
In plants
Maintenance respiration in plants refers to the amount of cellular respiration, measured by the carbon dioxide (CO2) released or oxygen (O2) consumed, during the generation of usable energy (mainly ATP, NADPH, and NADH) and metabolic intermediates used for (i) resynthesis of compounds that undergo renewal (turnover) in the normal process of metabolism (examples are enzymatic proteins, ribonucleic acids, and membrane lipids); (ii) maintenance of chemical gradients of ions and metabolites across cellular membranes that are necessary for cellular integrity and plant health; and (iii) operation of metabolic processes involved in physiological adjustment (i.e., acclimation) to a change in the plant's environment. The metabolic costs of the repair of injury from biotic or abiotic stress may also be considered a part of maintenance respiration.
Maintenance respiration is essential for biological health and growth of plants. It is estimated that about half of the respiration carried out by terrestrial plants during their lifetime is for the support of maintenance processes. Because typically more than half of global terrestrial plant photosynthesis (or gross primary production) is used for plant respiration, more than one quarter of global terrestrial plant photosynthesis is presumably consumed in maintenance respiration.
Maintenance respiration is a key component of most physiologically based mathematical models of plant growth, including models of crop growth and yield and models of
Document 2:::
The soil-plant-atmosphere continuum (SPAC) is the pathway for water moving from soil through plants to the atmosphere. Continuum in the description highlights the continuous nature of water connection through the pathway. The low water potential of the atmosphere, and relatively higher (i.e. less negative) water potential inside leaves, leads to a diffusion gradient across the stomatal pores of leaves, drawing water out of the leaves as vapour. As water vapour transpires out of the leaf, further water molecules evaporate off the surface of mesophyll cells to replace the lost molecules since water in the air inside leaves is maintained at saturation vapour pressure. Water lost at the surface of cells is replaced by water from the xylem, which due to the cohesion-tension properties of water in the xylem of plants pulls additional water molecules through the xylem from the roots toward the leaf.
Components
The transport of water along this pathway occurs in components, variously defined among scientific disciplines:
Soil physics characterizes water in soil in terms of tension,
Physiology of plants and animals characterizes water in organisms in terms of diffusion pressure deficit, and
Meteorology uses vapour pressure or relative humidity to characterize atmospheric water.
SPAC integrates these components and is defined as a:
...concept recognising that the field with all its components (soil, plant, animals and the ambient atmosphere taken together) constitutes a physically integrated, dynamic system in which the various flow processes involving energy and matter occur simultaneously and independently like links in the chain.
This characterises the state of water in different components of the SPAC as expressions of the energy level or water potential of each. Modelling of water transport between components relies on SPAC, as do studies of water potential gradients between segments.
See also
Ecohydrology
Evapotranspiration
Hydraulic redistribution; a p
Document 3:::
A leaf (: leaves) is a principal appendage of the stem of a vascular plant, usually borne laterally aboveground and specialized for photosynthesis. Leaves are collectively called foliage, as in "autumn foliage", while the leaves, stem, flower, and fruit collectively form the shoot system. In most leaves, the primary photosynthetic tissue is the palisade mesophyll and is located on the upper side of the blade or lamina of the leaf but in some species, including the mature foliage of Eucalyptus, palisade mesophyll is present on both sides and the leaves are said to be isobilateral. Most leaves are flattened and have distinct upper (adaxial) and lower (abaxial) surfaces that differ in color, hairiness, the number of stomata (pores that intake and output gases), the amount and structure of epicuticular wax and other features. Leaves are mostly green in color due to the presence of a compound called chlorophyll which is essential for photosynthesis as it absorbs light energy from the sun. A leaf with lighter-colored or white patches or edges is called a variegated leaf.
Leaves can have many different shapes, sizes, textures and colors. The broad, flat leaves with complex venation of flowering plants are known as megaphylls and the species that bear them, the majority, as broad-leaved or megaphyllous plants, which also include acrogymnosperms and ferns. In the lycopods, with different evolutionary origins, the leaves are simple (with only a single vein) and are known as microphylls. Some leaves, such as bulb scales, are not above ground. In many aquatic species, the leaves are submerged in water. Succulent plants often have thick juicy leaves, but some leaves are without major photosynthetic function and may be dead at maturity, as in some cataphylls and spines. Furthermore, several kinds of leaf-like structures found in vascular plants are not totally homologous with them. Examples include flattened plant stems called phylloclades and cladodes, and flattened leaf stems
Document 4:::
Photosynthetic capacity (Amax) is a measure of the maximum rate at which leaves are able to fix carbon during photosynthesis. It is typically measured as the amount of carbon dioxide that is fixed per metre squared per second, for example as μmol m−2 sec−1.
Limitations
Photosynthetic capacity is limited by carboxylation capacity and electron transport capacity. For example, in high carbon dioxide concentrations or in low light, the plant is not able to regenerate ribulose-1,5-bisphosphate fast enough (also known RUBP, the acceptor molecule in photosynthetic carbon reduction). So in this case, photosynthetic capacity is limited by electron transport of the light reaction, which generates the NADPH and ATP required for the PCR (Calvin) Cycle, and regeneration of RUBP. On the other hand, in low carbon dioxide concentrations, the capacity of the plant to perform carboxylation (adding carbon dioxide to Rubisco) is limited by the amount of available carbon dioxide, with plenty of Rubisco left over.¹ Light response, or photosynthesis-irradiance, curves display these relationships.
Current Research
Recent studies have shown that photosynthetic capacity in leaves can be increased with an increase in the number of stomata per leaf. This could be important in further crop development engineering to increase the photosynthetic efficiency through increasing diffusion of carbon dioxide into the plant.²
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Some plants open their leaves during the day to collect what?
A. moisture
B. energy
C. sunlight
D. precipitation
Answer:
|
|
sciq-3186
|
multiple_choice
|
Carbon, hydrogen, nitrogen, oxygen, phosphorus and sulfur are basic elements that form what type of compounds, which are vital to life?
|
[
"hydrocarbons",
"amino acids",
"inorganic compounds",
"organic compounds"
] |
D
|
Relavent Documents:
Document 0:::
Carbon is a primary component of all known life on Earth, representing approximately 45–50% of all dry biomass. Carbon compounds occur naturally in great abundance on Earth. Complex biological molecules consist of carbon atoms bonded with other elements, especially oxygen and hydrogen and frequently also nitrogen, phosphorus, and sulfur (collectively known as CHNOPS).
Because it is lightweight and relatively small in size, carbon molecules are easy for enzymes to manipulate. It is frequently assumed in astrobiology that if life exists elsewhere in the Universe, it will also be carbon-based. Critics refer to this assumption as carbon chauvinism.
Characteristics
Carbon is capable of forming a vast number of compounds, more than any other element, with almost ten million compounds described to date, and yet that number is but a fraction of the number of theoretically possible compounds under standard conditions. The enormous diversity of carbon-containing compounds, known as organic compounds, has led to a distinction between them and compounds that do not contain carbon, known as inorganic compounds. The branch of chemistry that studies organic compounds is known as organic chemistry.
Carbon is the 15th most abundant element in the Earth's crust, and the fourth most abundant element in the universe by mass, after hydrogen, helium, and oxygen. Carbon's widespread abundance, its ability to form stable bonds with numerous other elements, and its unusual ability to form polymers at the temperatures commonly encountered on Earth enables it to serve as a common element of all known living organisms. In a 2018 study, carbon was found to compose approximately 550 billion tons of all life on Earth. It is the second most abundant element in the human body by mass (about 18.5%) after oxygen.
The most important characteristics of carbon as a basis for the chemistry of life are that each carbon atom is capable of forming up to four valence bonds with other atoms simultaneously
Document 1:::
This is an index of lists of molecules (i.e. by year, number of atoms, etc.). Millions of molecules have existed in the universe since before the formation of Earth. Three of them, carbon dioxide, water and oxygen were necessary for the growth of life. Although humanity had always been surrounded by these substances, it has not always known what they were composed of.
By century
The following is an index of list of molecules organized by time of discovery of their molecular formula or their specific molecule in case of isomers:
List of compounds
By number of carbon atoms in the molecule
List of compounds with carbon number 1
List of compounds with carbon number 2
List of compounds with carbon number 3
List of compounds with carbon number 4
List of compounds with carbon number 5
List of compounds with carbon number 6
List of compounds with carbon number 7
List of compounds with carbon number 8
List of compounds with carbon number 9
List of compounds with carbon number 10
List of compounds with carbon number 11
List of compounds with carbon number 12
List of compounds with carbon number 13
List of compounds with carbon number 14
List of compounds with carbon number 15
List of compounds with carbon number 16
List of compounds with carbon number 17
List of compounds with carbon number 18
List of compounds with carbon number 19
List of compounds with carbon number 20
List of compounds with carbon number 21
List of compounds with carbon number 22
List of compounds with carbon number 23
List of compounds with carbon number 24
List of compounds with carbon numbers 25-29
List of compounds with carbon numbers 30-39
List of compounds with carbon numbers 40-49
List of compounds with carbon numbers 50+
Other lists
List of interstellar and circumstellar molecules
List of gases
List of molecules with unusual names
See also
Molecule
Empirical formula
Chemical formula
Chemical structure
Chemical compound
Chemical bond
Coordination complex
L
Document 2:::
The Seven Pillars of Life are the essential principles of life described by Daniel E. Koshland in 2002 in order to create a universal definition of life. One stated goal of this universal definition is to aid in understanding and identifying artificial and extraterrestrial life. The seven pillars are Program, Improvisation, Compartmentalization, Energy, Regeneration, Adaptability, and Seclusion. These can be abbreviated as PICERAS.
The Seven Pillars
Program
Koshland defines "Program" as an "organized plan that describes both the ingredients themselves and the kinetics of the interactions among ingredients as the living system persists through time." In natural life as it is known on Earth, the program operates through the mechanisms of nucleic acids and amino acids, but the concept of program can apply to other imagined or undiscovered mechanisms.
Improvisation
"Improvisation" refers to the living system's ability to change its program in response to the larger environment in which it exists. An example of improvisation on earth is natural selection.
Compartmentalization
"Compartmentalization" refers to the separation of spaces in the living system that allow for separate environments for necessary chemical processes. Compartmentalization is necessary to protect the concentration of the ingredients for a reaction from outside environments.
Energy
Because living systems involve net movement in terms of chemical movement or body movement, and lose energy in those movements through entropy, energy is required for a living system to exist. The main source of energy on Earth is the sun, but other sources of energy exist for life on Earth, such as hydrogen gas or methane, used in chemosynthesis.
Regeneration
"Regeneration" in a living system refers to the general compensation for losses and degradation in the various components and processes in the system. This covers the thermodynamic loss in chemical reactions, the wear and tear of larger parts, and the large
Document 3:::
A carbon–carbon bond is a covalent bond between two carbon atoms. The most common form is the single bond: a bond composed of two electrons, one from each of the two atoms. The carbon–carbon single bond is a sigma bond and is formed between one hybridized orbital from each of the carbon atoms. In ethane, the orbitals are sp3-hybridized orbitals, but single bonds formed between carbon atoms with other hybridizations do occur (e.g. sp2 to sp2). In fact, the carbon atoms in the single bond need not be of the same hybridization. Carbon atoms can also form double bonds in compounds called alkenes or triple bonds in compounds called alkynes. A double bond is formed with an sp2-hybridized orbital and a p-orbital that is not involved in the hybridization. A triple bond is formed with an sp-hybridized orbital and two p-orbitals from each atom. The use of the p-orbitals forms a pi bond.
Chains and branching
Carbon is one of the few elements that can form long chains of its own atoms, a property called catenation. This coupled with the strength of the carbon–carbon bond gives rise to an enormous number of molecular forms, many of which are important structural elements of life, so carbon compounds have their own field of study: organic chemistry.
Branching is also common in C−C skeletons. Carbon atoms in a molecule are categorized by the number of carbon neighbors they have:
A primary carbon has one carbon neighbor.
A secondary carbon has two carbon neighbors.
A tertiary carbon has three carbon neighbors.
A quaternary carbon has four carbon neighbors.
In "structurally complex organic molecules", it is the three-dimensional orientation of the carbon–carbon bonds at quaternary loci which dictates the shape of the molecule. Further, quaternary loci are found in many biologically active small molecules, such as cortisone and morphine.
Synthesis
Carbon–carbon bond-forming reactions are organic reactions in which a new carbon–carbon bond is formed. They are important in th
Document 4:::
Biotic material or biological derived material is any material that originates from living organisms. Most such materials contain carbon and are capable of decay.
The earliest life on Earth arose at least 3.5 billion years ago. Earlier physical evidences of life include graphite, a biogenic substance, in 3.7 billion-year-old metasedimentary rocks discovered in southwestern Greenland, as well as, "remains of biotic life" found in 4.1 billion-year-old rocks in Western Australia. Earth's biodiversity has expanded continually except when interrupted by mass extinctions. Although scholars estimate that over 99 percent of all species of life (over five billion) that ever lived on Earth are extinct, there are still an estimated 10–14 million extant species, of which about 1.2 million have been documented and over 86% have not yet been described.
Examples of biotic materials are wood, straw, humus, manure, bark, crude oil, cotton, spider silk, chitin, fibrin, and bone.
The use of biotic materials, and processed biotic materials (bio-based material) as alternative natural materials, over synthetics is popular with those who are environmentally conscious because such materials are usually biodegradable, renewable, and the processing is commonly understood and has minimal environmental impact. However, not all biotic materials are used in an environmentally friendly way, such as those that require high levels of processing, are harvested unsustainably, or are used to produce carbon emissions.
When the source of the recently living material has little importance to the product produced, such as in the production of biofuels, biotic material is simply called biomass. Many fuel sources may have biological sources, and may be divided roughly into fossil fuels, and biofuel.
In soil science, biotic material is often referred to as organic matter. Biotic materials in soil include glomalin, Dopplerite and humic acid. Some biotic material may not be considered to be organic matte
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Carbon, hydrogen, nitrogen, oxygen, phosphorus and sulfur are basic elements that form what type of compounds, which are vital to life?
A. hydrocarbons
B. amino acids
C. inorganic compounds
D. organic compounds
Answer:
|
|
sciq-9333
|
multiple_choice
|
Generators usually transform kinetic energy into what kind of energy?
|
[
"nuclear energy",
"subsequent energy",
"potential energy",
"electrical energy"
] |
D
|
Relavent Documents:
Document 0:::
In electrical engineering, electric machine is a general term for machines using electromagnetic forces, such as electric motors, electric generators, and others. They are electromechanical energy converters: an electric motor converts electricity to mechanical power while an electric generator converts mechanical power to electricity. The moving parts in a machine can be rotating (rotating machines) or linear (linear machines). Besides motors and generators, a third category often included is transformers, which although they do not have any moving parts are also energy converters, changing the voltage level of an alternating current.
Electric machines, in the form of synchronous and induction generators, produce about 95% of all electric power on Earth (as of early 2020s), and in the form of electric motors consume approximately 60% of all electric power produced. Electric machines were developed beginning in the mid 19th century and since that time have been a ubiquitous component of the infrastructure. Developing more efficient electric machine technology is crucial to any global conservation, green energy, or alternative energy strategy.
Generator
An electric generator is a device that converts mechanical energy to electrical energy. A generator forces electrons to flow through an external electrical circuit. It is somewhat analogous to a water pump, which creates a flow of water but does not create the water inside. The source of mechanical energy, the prime mover, may be a reciprocating or turbine steam engine, water falling through a turbine or waterwheel, an internal combustion engine, a wind turbine, a hand crank, compressed air or any other source of mechanical energy.
The two main parts of an electrical machine can be described in either mechanical or electrical terms. In mechanical terms, the rotor is the rotating part, and the stator is the stationary part of an electrical machine. In electrical terms, the armature is the power-producing compo
Document 1:::
In electricity generation, a generator is a device that converts motion-based power (potential and kinetic energy) or fuel-based power (chemical energy) into electric power for use in an external circuit. Sources of mechanical energy include steam turbines, gas turbines, water turbines, internal combustion engines, wind turbines and even hand cranks. The first electromagnetic generator, the Faraday disk, was invented in 1831 by British scientist Michael Faraday. Generators provide nearly all the power for electrical grids.
In addition to electricity- and motion-based designs, photovoltaic and fuel cell powered generators use solar power and hydrogen-based fuels, respectively, to generate electrical output.
The reverse conversion of electrical energy into mechanical energy is done by an electric motor, and motors and generators are very similar. Many motors can generate electricity from mechanical energy.
Terminology
Electromagnetic generators fall into one of two broad categories, dynamos and alternators.
Dynamos generate pulsing direct current through the use of a commutator.
Alternators generate alternating current.
Mechanically, a generator consists of a rotating part and a stationary part which together form a magnetic circuit:
Rotor: The rotating part of an electrical machine.
Stator: The stationary part of an electrical machine, which surrounds the rotor.
One of these parts generates a magnetic field, the other has a wire winding in which the changing field induces an electric current:
Field winding or field (permanent) magnets: The magnetic field-producing component of an electrical machine. The magnetic field of the dynamo or alternator can be provided by either wire windings called field coils or permanent magnets. Electrically-excited generators include an excitation system to produce the field flux. A generator using permanent magnets (PMs) is sometimes called a magneto, or a permanent magnet synchronous generator (PMSG).
Armature: The power-p
Document 2:::
Micropower describes the use of very small electric generators and prime movers or devices to convert heat or motion to electricity, for use close to the generator. The generator is typically integrated with microelectronic devices and produces "several watts of power or less." These devices offer the promise of a power source for portable electronic devices which is lighter weight and has a longer operating time than batteries.
Microturbine technology
The components of any turbine engine — the gas compressor, the combustion chamber, and the turbine rotor — are fabricated from etched silicon, much like integrated circuits. The technology holds the promise of ten times the operating time of a battery of the same weight as the micropower unit, and similar efficiency to large utility gas turbines. Researchers at Massachusetts Institute of Technology have thus far succeeded in fabricating the parts for such a micro turbine out of six etched and stacked silicon wafers, and are working toward combining them into a functioning engine about the size of a U.S. quarter coin.
Researchers at Georgia Tech have built a micro generator 10 mm wide, which spins a magnet above an array of coils fabricated on a silicon chip. The device spins at 100,000 revolutions per minute, and produces 1.1 watts of electrical power, sufficient to operate a cell phone. Their goal is to produce 20 to 50 watts, sufficient to power a laptop computer.
Scientists at Lehigh University are developing a hydrogen generator on a silicon chip that can convert methanol, diesel, or gasoline into fuel for a microengine or a miniature fuel cell.
Professor Sanjeev Mukerjee of Northeastern University's chemistry department is developing fuel cells for the military that will burn hydrogen to power portable electronic equipment, such as night vision goggles, computers, and communication equipment. In his system, a cartridge of methanol would be used to produce hydrogen to run a small fuel cell for up to 5,000 ho
Document 3:::
This page lists examples of the power in watts produced by various sources of energy. They are grouped by orders of magnitude from small to large.
Below 1 W
1 to 102 W
103 to 108 W
The productive capacity of electrical generators operated by utility companies is often measured in MW. Few things can sustain the transfer or consumption of energy on this scale; some of these events or entities include: lightning strikes, naval craft (such as aircraft carriers and submarines), engineering hardware, and some scientific research equipment (such as supercolliders and large lasers).
For reference, about 10,000 100-watt lightbulbs or 5,000 computer systems would be needed to draw 1 MW. Also, 1 MW is approximately 1360 horsepower. Modern high-power diesel-electric locomotives typically have a peak power of 3–5 MW, while a typical modern nuclear power plant produces on the order of 500–2000 MW peak output.
109 to 1014 W
1015 to 1026 W
Over 1027 W
See also
Orders of magnitude (energy)
Orders of magnitude (voltage)
World energy resources and consumption
International System of Units (SI)
SI prefix
Notes
Document 4:::
The tip-speed ratio, λ, or TSR for wind turbines is the ratio between the tangential speed of the tip of a blade and the actual speed of the wind, . The tip-speed ratio is related to efficiency, with the optimum varying with blade design. Higher tip speeds result in higher noise levels and require stronger blades due to larger centrifugal forces.
The tip speed of the blade can be calculated as times R, where is the rotational speed of the rotor in radians/second, and R is the rotor radius in metres. Therefore, we can also write:
where is the wind speed in metres/second at the height of the blade hub.
Cp–λ curves
The power coefficient, is a quantity that expresses what fraction of the power in the wind is being extracted by the wind turbine. It is generally assumed to be a function of both tip-speed ratio and pitch angle. Below is a plot of the variation of the power coefficient with variations in the tip-speed ratio when the pitch is held constant:
The case for variable speed wind turbines
Originally, wind turbines were fixed speed. This has the benefit that the rotor speed in the generator is constant, thus the frequency of the AC voltage is fixed. This allows the wind turbine to be directly connected to a transmission system. However, from the figure above, we can see that the power coefficient is a function of the tip-speed ratio. By extension, the efficiency of the wind turbine is a function of the tip-speed ratio.
Ideally, one would like to have a turbine operating at the maximum value of at all wind speeds. This means that as the wind speed changes, the rotor speed must change to such that . A wind turbine with a variable rotor speed is called a variable speed wind turbine. Whilst this does mean that the wind turbine operates at or close to for a range of wind speeds, the frequency of the AC voltage generator will not be constant. This can be seen in the following equation:
where is the rotor angular speed, is the frequency of the AC volta
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Generators usually transform kinetic energy into what kind of energy?
A. nuclear energy
B. subsequent energy
C. potential energy
D. electrical energy
Answer:
|
|
ai2_arc-420
|
multiple_choice
|
In whitetail deer, females seldom grow antlers. Which best explains why male whitetail deer grow antlers but females seldom grow antlers?
|
[
"Female deer have no need for antlers.",
"Male deer are older than female deer.",
"Antler growth is controlled by genes.",
"Antler growth depends on behavior."
] |
C
|
Relavent Documents:
Document 0:::
Animal science is described as "studying the biology of animals that are under the control of humankind". It can also be described as the production and management of farm animals. Historically, the degree was called animal husbandry and the animals studied were livestock species, like cattle, sheep, pigs, poultry, and horses. Today, courses available look at a broader area, including companion animals, like dogs and cats, and many exotic species. Degrees in Animal Science are offered at a number of colleges and universities. Animal science degrees are often offered at land-grant universities, which will often have on-campus farms to give students hands-on experience with livestock animals.
Education
Professional education in animal science prepares students for careers in areas such as animal breeding, food and fiber production, nutrition, animal agribusiness, animal behavior, and welfare. Courses in a typical Animal Science program may include genetics, microbiology, animal behavior, nutrition, physiology, and reproduction. Courses in support areas, such as genetics, soils, agricultural economics and marketing, legal aspects, and the environment also are offered.
Bachelor degree
At many universities, a Bachelor of Science (BS) degree in Animal Science allows emphasis in certain areas. Typical areas are species-specific or career-specific. Species-specific areas of emphasis prepare students for a career in dairy management, beef management, swine management, sheep or small ruminant management, poultry production, or the horse industry. Other career-specific areas of study include pre-veterinary medicine studies, livestock business and marketing, animal welfare and behavior, animal nutrition science, animal reproduction science, or genetics. Youth programs are also an important part of animal science programs.
Pre-veterinary emphasis
Many schools that offer a degree option in Animal Science also offer a pre-veterinary emphasis such as Iowa State University, th
Document 1:::
Female education in STEM refers to child and adult female representation in the educational fields of science, technology, engineering, and mathematics (STEM). In 2017, 33% of students in STEM fields were women.
The organization UNESCO has stated that this gender disparity is due to discrimination, biases, social norms and expectations that influence the quality of education women receive and the subjects they study. UNESCO also believes that having more women in STEM fields is desirable because it would help bring about sustainable development.
Current status of girls and women in STEM education
Overall trends in STEM education
Gender differences in STEM education participation are already visible in early childhood care and education in science- and math-related play, and become more pronounced at higher levels of education. Girls appear to lose interest in STEM subjects with age, particularly between early and late adolescence. This decreased interest affects participation in advanced studies at the secondary level and in higher education. Female students represent 35% of all students enrolled in STEM-related fields of study at this level globally. Differences are also observed by disciplines, with female enrollment lowest in engineering, manufacturing and construction, natural science, mathematics and statistics and ICT fields. Significant regional and country differences in female representation in STEM studies can be observed, though, suggesting the presence of contextual factors affecting girls’ and women's engagement in these fields. Women leave STEM disciplines in disproportionate numbers during their higher education studies, in their transition to the world of work and even in their career cycle.
Learning achievement in STEM education
Data on gender differences in learning achievement present a complex picture, depending on what is measured (subject, knowledge acquisition against knowledge application), the level of education/age of students, and
Document 2:::
The SAT Subject Test in Biology was the name of a one-hour multiple choice test given on biology by the College Board. A student chose whether to take the test depending upon college entrance requirements for the schools in which the student is planning to apply. Until 1994, the SAT Subject Tests were known as Achievement Tests; and from 1995 until January 2005, they were known as SAT IIs. Of all SAT subject tests, the Biology E/M test was the only SAT II that allowed the test taker a choice between the ecological or molecular tests. A set of 60 questions was taken by all test takers for Biology and a choice of 20 questions was allowed between either the E or M tests. This test was graded on a scale between 200 and 800. The average for Molecular is 630 while Ecological is 591.
On January 19 2021, the College Board discontinued all SAT Subject tests, including the SAT Subject Test in Biology E/M. This was effective immediately in the United States, and the tests were to be phased out by the following summer for international students. This was done as a response to changes in college admissions due to the impact of the COVID-19 pandemic on education.
Format
This test had 80 multiple-choice questions that were to be answered in one hour. All questions had five answer choices. Students received one point for each correct answer, lost ¼ of a point for each incorrect answer, and received 0 points for questions left blank. The student's score was based entirely on his or her performance in answering the multiple-choice questions.
The questions covered a broad range of topics in general biology. There were more specific questions related respectively on ecological concepts (such as population studies and general Ecology) on the E test and molecular concepts such as DNA structure, translation, and biochemistry on the M test.
Preparation
The College Board suggested a year-long course in biology at the college preparatory level, as well as a one-year course in algebra, a
Document 3:::
The Force Concept Inventory is a test measuring mastery of concepts commonly taught in a first semester of physics developed by Hestenes, Halloun, Wells, and Swackhamer (1985). It was the first such "concept inventory" and several others have been developed since for a variety of topics. The FCI was designed to assess student understanding of the Newtonian concepts of force. Hestenes (1998) found that while "nearly 80% of the [students completing introductory college physics courses] could state Newton's Third Law at the beginning of the course, FCI data showed that less than 15% of them fully understood it at the end". These results have been replicated in a number of studies involving students at a range of institutions (see sources section below), and have led to greater recognition in the physics education research community of the importance of students' "active engagement" with the materials to be mastered.
The 1995 version has 30 five-way multiple choice questions.
Example question (question 4):
Gender differences
The FCI shows a gender difference in favor of males that has been the subject of some research in regard to gender equity in education. Men score on average about 10% higher.
Document 4:::
Sexual dimorphism is the condition where sexes of the same species exhibit different morphological characteristics, particularly characteristics not directly involved in reproduction. Sexual dimorphism in carnivorans, in which males are larger than females, is common among carnivorans. Sexual selection is frequently cited as the cause of the intraspecific divergence in body proportions and craniomandibular morphology between the sexes within the Carnivora order. It is anticipated that animals with polygynous mating systems and high levels of territoriality and solitary behavior will exhibit the highest levels of sexual size dimorphism. Pinnipeds offer an illustration for this.
Different types
Body size
Sexual size dimorphism is the difference in body size between the sexes within a group of some sort or another. Carnivorans exhibit high levels of sexual size dimorphism with males generally being larger than females.
Canine tooth
Males have larger, longer and more powerful canines than their female counterparts. A study of skull and tooth size in 45 species of Carnivorans showed that sexual dimorphism was most pronounced in the size of the canine tooth. Breeding systems seems to be the most reasonable explanation for the finding. Social species like lions, in which males have a canine teeth 25% larger than females are the most sexual dimorphic of felids.
Skeletal structure
In terms of skeletal structure,
Carnivorans are highly sexually dimorphic. Males have more robust and larger skulls which promotes a stronger biteforce, larger necks to permit more powerful neck muscles that work to prevent torsional loading of the neck and improve the ability to rend with the teeth and jerk the skull. Males also have larger scapulae that enable more muscle to transmit force from the trunk to the forelimbs and stabilize the shoulder joint and stronger limbs with better mechanical advantages due to anatomy.
Mechanisms
A secondary factor that propels the evolution of se
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
In whitetail deer, females seldom grow antlers. Which best explains why male whitetail deer grow antlers but females seldom grow antlers?
A. Female deer have no need for antlers.
B. Male deer are older than female deer.
C. Antler growth is controlled by genes.
D. Antler growth depends on behavior.
Answer:
|
|
sciq-5012
|
multiple_choice
|
Multiplying the linear momentum of a spinning object by the radius calculates what?
|
[
"applied momentum",
"angular torque",
"angular momentum",
"total momentum"
] |
C
|
Relavent Documents:
Document 0:::
The following table lists various orders of magnitude for angular momentum, in Joule-seconds.
Table
See also
Orders of magnitude (angular velocity)
Orders of magnitude (momentum)
Document 1:::
The angular momentum problem is a problem in astrophysics identified by Leon Mestel in 1965.
It was found that the angular momentum of a protoplanetary disk is misappropriated when compared to models during stellar birth. The Sun and other stars are predicted by models to be rotating considerably faster than they actually are. The Sun, for example, only accounts for about 0.3 percent of the total angular momentum of the Solar System while about 60% is attributed to Jupiter.
See also
History of Solar System formation and evolution hypotheses
Document 2:::
In quantum mechanics, the total angular momentum quantum number parametrises the total angular momentum of a given particle, by combining its orbital angular momentum and its intrinsic angular momentum (i.e., its spin).
If s is the particle's spin angular momentum and ℓ its orbital angular momentum vector, the total angular momentum j is
The associated quantum number is the main total angular momentum quantum number j. It can take the following range of values, jumping only in integer steps:
where ℓ is the azimuthal quantum number (parameterizing the orbital angular momentum) and s is the spin quantum number (parameterizing the spin).
The relation between the total angular momentum vector j and the total angular momentum quantum number j is given by the usual relation (see angular momentum quantum number)
The vector's z-projection is given by
where mj is the secondary total angular momentum quantum number, and the is the reduced Planck's constant. It ranges from −j to +j in steps of one. This generates 2j + 1 different values of mj.
The total angular momentum corresponds to the Casimir invariant of the Lie algebra so(3) of the three-dimensional rotation group.
See also
Principal quantum number
Orbital angular momentum quantum number
Magnetic quantum number
Spin quantum number
Angular momentum coupling
Clebsch–Gordan coefficients
Angular momentum diagrams (quantum mechanics)
Rotational spectroscopy
Document 3:::
The angular momentum of light is a vector quantity that expresses the amount of dynamical rotation present in the electromagnetic field of the light. While traveling approximately in a straight line, a beam of light can also be rotating (or "spinning, or "twisting) around its own axis. This rotation, while not visible to the naked eye, can be revealed by the interaction of the light beam with matter.
There are two distinct forms of rotation of a light beam, one involving its polarization and the other its wavefront shape. These two forms of rotation are therefore associated with two distinct forms of angular momentum, respectively named light spin angular momentum (SAM) and light orbital angular momentum (OAM).
The total angular momentum of light (or, more generally, of the electromagnetic field and the other force fields) and matter is conserved in time.
Introduction
Light, or more generally an electromagnetic wave, carries not only energy but also momentum, which is a characteristic property of all objects in translational motion. The existence of this momentum becomes apparent in the "radiation pressure phenomenon, in which a light beam transfers its momentum to an absorbing or scattering object, generating a mechanical pressure on it in the process.
Light may also carry angular momentum, which is a property of all objects in rotational motion. For example, a light beam can be rotating around its own axis while it propagates forward. Again, the existence of this angular momentum can be made evident by transferring it to small absorbing or scattering particles, which are thus subject to an optical torque.
For a light beam, one can usually distinguish two "forms of rotation, the first associated with the dynamical rotation of the electric and magnetic fields around the propagation direction, and the second with the dynamical rotation of light rays around the main beam axis. These two rotations are associated with two forms of angular momentum, namely SAM an
Document 4:::
The moment of inertia, otherwise known as the mass moment of inertia, angular mass, second moment of mass, or most accurately, rotational inertia, of a rigid body is a quantity that determines the torque needed for a desired angular acceleration about a rotational axis, akin to how mass determines the force needed for a desired acceleration. It depends on the body's mass distribution and the axis chosen, with larger moments requiring more torque to change the body's rate of rotation by a given amount.
It is an extensive (additive) property: for a point mass the moment of inertia is simply the mass times the square of the perpendicular distance to the axis of rotation. The moment of inertia of a rigid composite system is the sum of the moments of inertia of its component subsystems (all taken about the same axis). Its simplest definition is the second moment of mass with respect to distance from an axis.
For bodies constrained to rotate in a plane, only their moment of inertia about an axis perpendicular to the plane, a scalar value, matters. For bodies free to rotate in three dimensions, their moments can be described by a symmetric 3-by-3 matrix, with a set of mutually perpendicular principal axes for which this matrix is diagonal and torques around the axes act independently of each other.
In mechanical engineering, simply "inertia" is often used to refer to "inertial mass" or "moment of inertia".
Introduction
When a body is free to rotate around an axis, torque must be applied to change its angular momentum. The amount of torque needed to cause any given angular acceleration (the rate of change in angular velocity) is proportional to the moment of inertia of the body. Moments of inertia may be expressed in units of kilogram metre squared (kg·m2) in SI units and pound-foot-second squared (lbf·ft·s2) in imperial or US units.
The moment of inertia plays the role in rotational kinetics that mass (inertia) plays in linear kinetics—both characterize the resistance
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Multiplying the linear momentum of a spinning object by the radius calculates what?
A. applied momentum
B. angular torque
C. angular momentum
D. total momentum
Answer:
|
|
sciq-10449
|
multiple_choice
|
What type of earthquake creates a tsunami?
|
[
"Tectonic",
"underwater",
"collapse",
"volcanic"
] |
B
|
Relavent Documents:
Document 0:::
The Human-Induced Earthquake Database (HiQuake) is an online database that documents all reported cases of induced seismicity proposed on scientific grounds. It is the most complete compilation of its kind and is freely available to download via the associated website. The database is periodically updated to correct errors, revise existing entries, and add new entries reported in new scientific papers and reports. Suggestions for revisions and new entries can be made via the associated website.
History
In 2016, Nederlandse Aardolie Maatschappij funded a team of researchers from Durham University and Newcastle University to conduct a full review of induced seismicity. This review formed part of a scientific workshop aimed at estimating the maximum possible magnitude earthquake that might be induced by conventional gas production in the Groningen gas field.
The resulting database from the review was publicly released online on the 26 January 2017. The database was accompanied by the publication of two scientific papers, the more detailed of which is freely available online.
Document 1:::
A submarine, undersea, or underwater earthquake is an earthquake that occurs underwater at the bottom of a body of water, especially an ocean. They are the leading cause of tsunamis. The magnitude can be measured scientifically by the use of the moment magnitude scale and the intensity can be assigned using the Mercalli intensity scale.
Understanding plate tectonics helps to explain the cause of submarine earthquakes. The Earth's surface or lithosphere comprises tectonic plates which average approximately 50 miles in thickness, and are continuously moving very slowly upon a bed of magma in the asthenosphere and inner mantle. The plates converge upon one another, and one subducts below the other, or, where there is only shear stress, move horizontally past each other (see transform plate boundary below). Little movements called fault creep are minor and not measurable. The plates meet with each other, and if rough spots cause the movement to stop at the edges, the motion of the plates continue. When the rough spots can no longer hold, the sudden release of the built-up motion releases, and the sudden movement under the sea floor causes a submarine earthquake. This area of slippage both horizontally and vertically is called the epicenter, and has the highest magnitude, and causes the greatest damage.
As with a continental earthquake the severity of the damage is not often caused by the earthquake at the rift zone, but rather by events which are triggered by the earthquake. Where a continental earthquake will cause damage and loss of life on land from fires, damaged structures, and flying objects; a submarine earthquake alters the seabed, resulting in a series of waves, and depending on the length and magnitude of the earthquake, tsunami, which bear down on coastal cities causing property damage and loss of life.
Submarine earthquakes can also damage submarine communications cables, leading to widespread disruption of the Internet and international telephone networ
Document 2:::
The moment magnitude scale (MMS; denoted explicitly with or Mw, and generally implied with use of a single M for magnitude) is a measure of an earthquake's magnitude ("size" or strength) based on its seismic moment. It was defined in a 1979 paper by Thomas C. Hanks and Hiroo Kanamori. Similar to the local magnitude/Richter scale () defined by Charles Francis Richter in 1935, it uses a logarithmic scale; small earthquakes have approximately the same magnitudes on both scales. Despite the difference, news media often says "Richter scale" when referring to the moment magnitude scale.
Moment magnitude () is considered the authoritative magnitude scale for ranking earthquakes by size. It is more directly related to the energy of an earthquake than other scales, and does not saturatethat is, it does not underestimate magnitudes as other scales do in certain conditions. It has become the standard scale used by seismological authorities like the U.S. Geological Survey for reporting large earthquakes (typically M > 4), replacing the local magnitude () and surface wave magnitude () scales. Subtypes of the moment magnitude scale (, etc.) reflect different ways of estimating the seismic moment.
History
Richter scale: the original measure of earthquake magnitude
At the beginning of the twentieth century, very little was known about how earthquakes happen, how seismic waves are generated and propagate through the Earth's crust, and what information they carry about the earthquake rupture process; the first magnitude scales were therefore empirical. The initial step in determining earthquake magnitudes empirically came in 1931 when the Japanese seismologist Kiyoo Wadati showed that the maximum amplitude of an earthquake's seismic waves diminished with distance at a certain rate. Charles F. Richter then worked out how to adjust for epicentral distance (and some other factors) so that the logarithm of the amplitude of the seismograph trace could be used as a measure of "magnit
Document 3:::
Seismic moment is a quantity used by seismologists to measure the size of an earthquake. The scalar seismic moment is defined by the equation
, where
is the shear modulus of the rocks involved in the earthquake (in pascals (Pa), i.e. newtons per square meter)
is the area of the rupture along the geologic fault where the earthquake occurred (in square meters), and
is the average slip (displacement offset between the two sides of the fault) on (in meters).
thus has dimensions of torque, measured in newton meters. The connection between seismic moment and a torque is natural in the body-force equivalent representation of seismic sources as a double-couple (a pair of force couples with opposite torques): the seismic moment is the torque of each of the two couples. Despite having the same dimensions as energy, seismic moment is not a measure of energy. The relations between seismic moment, potential energy drop and radiated energy are indirect and approximative.
The seismic moment of an earthquake is typically estimated using whatever information is available to constrain its factors. For modern earthquakes, moment is usually estimated from ground motion recordings of earthquakes known as seismograms. For earthquakes that occurred in times before modern instruments were available, moment may be estimated from geologic estimates of the size of the fault rupture and the slip.
Seismic moment is the basis of the moment magnitude scale introduced by Hiroo Kanamori, which is often used to compare the size of different earthquakes and is especially useful for comparing the sizes of large (great) earthquakes.
The seismic moment is not restricted to earthquakes. For a more general seismic source described by a seismic moment tensor (a symmetric tensor, but not necessarily a double couple tensor), the seismic moment is
See also
Richter magnitude scale
Moment magnitude scale
Sources
.
.
.
.
Seismology measurement
Moment (physics)
Document 4:::
Sand boils or sand volcanoes occur when water under pressure wells up through a bed of sand. The water looks like it is boiling up from the bed of sand, hence the name.
Sand volcano
A sand volcano or sand blow is a cone of sand formed by the ejection of sand onto a surface from a central point. The sand builds up as a cone with slopes at the sand's angle of repose. A crater is commonly seen at the summit. The cone looks like a small volcanic cone and can range in size from millimetres to metres in diameter.
The process is often associated with soil liquefaction and the ejection of fluidized sand that can occur in water-saturated sediments during an earthquake. The New Madrid Seismic Zone exhibited many such features during the 1811–12 New Madrid earthquakes. Adjacent sand blows aligned in a row along a linear fracture within fine-grained surface sediments are just as common, and can still be seen in the New Madrid area.
In the past few years, much effort has gone into the mapping of liquefaction features to study ancient earthquakes. The basic idea is to map zones that are susceptible to the process and then go in for a closer look. The presence or absence of soil liquefaction features is strong evidence of past earthquake activity, or lack thereof.
These are to be contrasted with mud volcanoes, which occur in areas of geyser or subsurface gas venting.
Flood protection structures
Sand boils can be a mechanism contributing to liquefaction and levee failure during floods. This effect is caused by a difference in pressure on two sides of a levee or dike, most likely during a flood. This process can result in internal erosion, whereby the removal of soil particles results in a pipe through the embankment. The creation of the pipe will quickly pick up pace and will eventually result in failure of the embankment.
A sand boil is difficult to stop. The most effective method is by creating a body of water above the boil to create enough pressure to slow the flow of
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What type of earthquake creates a tsunami?
A. Tectonic
B. underwater
C. collapse
D. volcanic
Answer:
|
|
sciq-1652
|
multiple_choice
|
What kind of climates are found at or near the equator?
|
[
"tundra",
"arctic climates",
"dry grasslands",
"tropical climates"
] |
D
|
Relavent Documents:
Document 0:::
Climatic adaptation refers to adaptations of an organism that are triggered due to the patterns of variation of abiotic factors that determine a specific climate. Annual means, seasonal variation and daily patterns of abiotic factors are properties of a climate where organisms can be adapted to. Changes in behavior, physical structure, internal mechanisms and metabolism are forms of adaptation that is caused by climate properties. Organisms of the same species that occur in different climates can be compared to determine which adaptations are due to climate and which are influenced majorly by other factors. Climatic adaptations limits to adaptations that have been established, characterizing species that live within the specific climate. It is different from climate change adaptations which refers to the ability to adapt to gradual changes of a climate. Once a climate has changed, the climate change adaptation that led to the survival of the specific organisms as a species can be seen as a climatic adaptation. Climatic adaptation is constrained by the genetic variability of the species in question.
Climate patterns
The patterns of variation of abiotic factors determine a climate and thus climatic adaptation. There are many different climates around the world, each with its unique patterns. Because of this, the manner of climatic adaptation shows large differences between the climates. A subarctic climate, for instance, shows daylight time and temperature fluctuations as most important factors, while in rainforest climate, the most important factor is characterized by the stable high precipitation rate and high average temperature that doesn't fluctuate a lot. Humid continental climate is marked by seasonal temperature variances which commonly lead to seasonal climate adaptations. Because the variance of these abiotic factors differ depending on the type of climate, differences in the manner of climatic adaptation are expected.
Research
Research on climatic adaptat
Document 1:::
A climograph is a graphical representation of a location's basic climate. Climographs display data for two variables: (a) monthly average temperature and (b) monthly average precipitation. These are useful tools to quickly describe a location's climate.
Representation
While temperature is typically visualized using a line, some climographs opt to visualize the data using a bar. This method's advantage allows the climograph to display the average range in temperature (average minimum and average maximum temperatures) rather than a simple monthly average.
Use
The patterns in a climograph describe not just a location's climate but also provide evidence for that climate's relative geographical location. For example, a climograph with a narrow range in temperature over the year might represent a location close to the equator, or alternatively a location adjacent to a large body of water exerting a moderating effect on the temperature range. Meanwhile, a wide range in annual temperature might suggest the opposite. We could also derive information about a site's ecological conditions through a climograph. For example, if precipitation is consistently low year-round, we might suggest the location reflects a desert; if there is a noticeable seasonal pattern to the precipitation, we might suggest the location experiences a monsoon season. When combining the temperature and precipitation patterns together, we have even better clues as to the local conditions. Despite this, a number of local factors contribute to the patterns observed in a particular place; therefore, a climograph is not a foolproof tool that captures all the geographic variation that might exist.
Document 2:::
A temperate forest is a forest found between the tropical and boreal regions, located in the temperate zone. It is the second largest biome on our planet, covering 25% of the world's forest area, only behind the boreal forest, which covers about 33%. These forests cover both hemispheres at latitudes ranging from 25 to 50 degrees, wrapping the planet in a belt similar to that of the boreal forest. Due to its large size spanning several continents, there are several main types: deciduous, coniferous, mixed forest, and rainforest.
Climate
The climate of a temperate forest is highly variable depending on the location of the forest. For example, Los Angeles and Vancouver, Canada are both considered to be located in a temperate zone, however, Vancouver is located in a temperate rainforest, while Los Angeles is a relatively dry subtropical climate.
Types of temperate forest
Deciduous
They are found in Europe, East Asia, North America, and in some parts of South America.
Deciduous forests are composed mainly of broadleaf trees, such as maple and oak, that shed all their leaves during one season. They are typically found in three middle-latitude regions with temperate climates characterized by a winter season and year-round precipitation: eastern North America, western Eurasia and northeastern Asia.
Coniferous
Coniferous forests are composed of needle-leaved evergreen trees, such as pine or fir. Evergreen forests are typically found in regions with moderate climates. Boreal forests, however, are an exception as they are found in subarctic regions. Coniferous trees often have an advantage over broadleaf trees in harsher environments. Their leaves are typically hardier and longer lived but require more energy to grow.
Mixed
As the name implies, conifers and broadleaf trees grow in the same area. The main trees found in these forests in North America and Eurasia include fir, oak, ash, maple, birch, beech, poplar, elm and pine. Other plant species may include magnolia,
Document 3:::
A Centre of Endemism is an area in which the ranges of restricted-range species overlap, or a localised area which has a high occurrence of endemics. Centres of endemism may overlap with biodiversity hotspots which are biogeographic regions characterized both by high levels of plant endemism and by serious levels of habitat loss. The exact delineation of centres of endemism is difficult and some overlap with one another. Centres of endemism are high conservation priority areas.
Examples of Centres of Endemism
Tanzania
A local centre of endemism is focussed on an area of lowland forests around the plateaux inland of Lindi in SE Tanzania, with between 40 and 91 species of vascular plants which are not found elsewhere.
Southern Africa
There are at least 19 centres of plant endemism, including the following:
Albany Centre of Plant Endemism
Barberton Centre of Plant Endemism
Cape Floristic Region
Drakensberg Alpine Centre
Hantam–Roggeveld Centre of Plant Endemism
Kaokoveld Centre of Endemism
Maputaland Centre of Plant Endemism
Pondoland Centre of Plant Endemism
Sekhukhuneland Centre of Endemism
Soutpansberg Centre of Plant Endemism
See also
List of ecoregions with high endemism
Document 4:::
Hemiboreal means halfway between the temperate and subarctic (or boreal) zones. The term is most frequently used in the context of climates and ecosystems.
Botany
A hemiboreal forest has some characteristics of a boreal forest to the north, and also shares features with temperate-zone forests to the south. Coniferous trees predominate in the hemiboreal zone, but a significant number of deciduous species, such as aspens, oaks, maples, ash trees, birches, beeches, hazels, and hornbeams, also take root here.
Climate
The term sometimes denotes the form of climate characteristic of the zone of hemiboreal forests—specifically, the climates designated Dfb, Dwb and Dsb in the Köppen climate classification scheme. On occasion, it is applied to all areas that have long, cold winters and warm (but not hot) summers—which also including areas that are semiarid(BS) and arid(BW) based on average annual precipitation. It can also be applied to some areas with a subpolar oceanic climate (Cfc), particularly those with continental climate characteristics.
Examples
Examples of locations with hemiboreal climates or ecosystems include:
Much of southern Canada (all of southeastern Canada except for parts of southern Ontario as well as the central Prairie Provinces outside the grasslands)
Within the United States: most parts of Michigan, Wisconsin, and Minnesota, along with eastern North Dakota and the Adirondacks of New York State and Northern New England. Also, many mountain areas in the western United States.
The Southern Siberian rainforest in Russia includes hemiboreal forests.
Parts of northeast China bordering Russia
Northern areas of Japan including Hokkaido
Parts of southern Norway and Southern Sweden except the most southern municipalities.
Latvia, Lithuania, Belarus and Estonia.
Coastal zone and archipelago of Turku in Finland and region of Åland.
The Australian Alps in eastern Victoria and southeastern New South Wales, which makes up a small portion in the southeastern
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What kind of climates are found at or near the equator?
A. tundra
B. arctic climates
C. dry grasslands
D. tropical climates
Answer:
|
|
sciq-6304
|
multiple_choice
|
What scientist and monk is best known for his experiments with pea plants?
|
[
"aristotle",
"steiner mendel",
"gregor mendel",
"charles darwin"
] |
C
|
Relavent Documents:
Document 0:::
The Department of Plant Sciences is a department of the University of Cambridge that conducts research and teaching in plant sciences. It was established in 1904, although the university has had a professor of botany since 1724.
Research
, the department pursues three strategic targets of research
Global food security
Synthetic biology and biotechnology
Climate science and ecosystem conservation
See also the Sainsbury Laboratory Cambridge University
Notable academic staff
Sir David Baulcombe, FRS, Regius Professor of Botany
Beverley Glover, Professor of Plant systematics and evolution, director of the Cambridge University Botanic Garden
Howard Griffiths, Professor of Plant Ecology
Julian Hibberd, Professor of Photosynthesis
Alison Smith, Professor of Plant Biochemistry and Head of Department
, the department also has 66 members of faculty and postdoctoral researchers, 100 graduate students, 19 Biotechnology and Biological Sciences Research Council (BBSRC) Doctoral Training Program (DTP) PhD students, 20 part II Tripos undergraduate students and 44 support staff.
History
The University of Cambridge has a long and distinguished history in Botany including work by John Ray and Stephen Hales in the 17th century and 18th century, Charles Darwin’s mentor John Stevens Henslow in the 19th century, and Frederick Blackman, Arthur Tansley and Harry Godwin in the 20th century.
Emeritus and alumni
More recently, the department has been home to:
John C. Gray, Emeritus Professor of Plant Molecular Biology since 2011
Thomas ap Rees, Professor of Botany
F. Ian Woodward, Lecturer and Fellow of Trinity Hall, Cambridge before being appointed Professor of Plant Ecology at the University of Sheffield
Document 1:::
The Regius Chair of Botany at the University of Glasgow is a Regius Professorship established in 1818.
A lectureship in botany had been founded in 1704. From 1718 to 1818, the subject was combined with Anatomy. The chair was founded in 1818 by King George III.
Regius Professors of Botany
For 1718–1818, see: Regius Professor of Anatomy
Robert Graham MD (1818)
Sir William Jackson Hooker MA LLD DCL FRS (1820)
John Hutton Balfour MA MD (1841)
George Arnott Walker-Arnott MA LLD, Advocate (1845)
Alexander Dickson MA MD (1868)
Sir Isaac Bayley Balfour MA MD DSc FRS (1879)
Frederick Orpen Bower MA ScD LLD FRS (1885)
James Montague Frank Drummond MA (1925)
John Walton MA DSc ScD D-es-Sc LLD (1930)
Percy Wragg Brian BA PhD DPhil (1963)
John Harrison Burnett MA DPhil (1968)
Malcolm Barrett Wilkins PhD DSc FRSE (1970)
Michael Robert Blatt BSc PhD FRSE (2001)
Document 2:::
The history of model organisms began with the idea that certain organisms can be studied and used to gain knowledge of other organisms or as a control (ideal) for other organisms of the same species. Model organisms offer standards that serve as the authorized basis for comparison of other organisms. Model organisms are made standard by limiting genetic variance, creating, hopefully, this broad applicability to other organisms.
The idea of the model organism first took root in the middle of the 19th century with the work of men like Charles Darwin and Gregor Mendel and their respective work on natural selection and the genetics of heredity. As the first model organisms were introduced into labs in the 20th century, these early efforts to identify standards to measure organisms against persisted. Beginning in the early 1900s Drosophila entered the research laboratories and opened up the doors for other model organisms like tobacco mosaic virus, E. coli, C57BL/6 (lab mice), etc. These organisms have led to many advances in the past century.
Preliminary works on model organisms
Some of the first work with what would be considered model organisms started because Gregor Johann Mendel felt that the views of Darwin were insufficient in describing the formation of a new species and he began his work with the pea plants that are so famously known today. In his experimentation to find a method by which Darwin's ideas could be explained he hybridized and cross-bred the peas and found that in so doing he could isolate phenotypic characteristics of the peas. These discoveries made in the 1860s lay dormant for nearly forty years until they were rediscovered in 1900. Mendel's work was then correlated with what was being called chromosomes within the nucleus of each cell. Mendel created a practical guide to breeding and this method has successfully been applied to select for some of the first model organisms of other genus and species such as Guinea pigs, Drosophila (fruit
Document 3:::
Julius von Sachs (; 2 October 1832 – 29 May 1897) was a German botanist from Breslau, Prussian Silesia. He is considered the founder of experimental plant physiology and co-founder of modern water culture. Julius von Sachs and Wilhelm Knop are monumental figures in the history of botany by first demonstrating the importance of water culture for the study of plant nutrition and plant physiology in the 19th century.
Early life
Sachs was born at Breslau on 2 October 1832. His father, Graveur Sachs, was an engraver by trade, and father taught son delineation and accuracy of line and color. From earliest boyhood Julius was fascinated with plants, making collections of them on many field excursions with his father. He gave much of his time between the ages of thirteen and sixteen to drawing and painting the flowers, fungi, and other specimens which he collected. At the Gymnasium from 1845 to 1850 he was most interested in the natural sciences, collecting skulls, writing a monograph on the crayfish. His natural science teacher, one Krober, showed a singular lack of foresight when he solemnly warned young Sachs against devoting himself to the natural sciences.
When he was sixteen years old, his father died, and in the next year both his mother and a brother died of cholera. Suddenly without financial support, he was fortunate to be taken into the family of Jan Evangelista Purkyně who had accepted a professorship at the University of Prague. Sachs was admitted to the university in 1851.
Sachs famously labored long hours in the laboratory for Purkyně, and then long hours for himself each day after his work in the laboratory was finished. After the lab, he could devote himself entirely to establishing how plants grow.
Career
In 1856 Sachs graduated as doctor of philosophy, and then adopted a botanical career, establishing himself as Privatdozent for plant physiology. In 1859 he was appointed physiological assistant to the Agricultural Academy of Tharandt (now part of
Document 4:::
"Experiments on Plant Hybridization" (German: "Versuche über Pflanzen-Hybriden") is a seminal paper written in 1865 and published in 1866 by Gregor Mendel, an Augustinian friar considered to be the founder of modern genetics. The paper was the result after years spent studying genetic traits in Pisum sativum, the pea plant.
Content
In his paper, Mendel compared 7 pairs of discrete traits found in a pea plant:
Through experimentation, Mendel discovered that one inheritable trait would invariably be dominant to its recessive alternative. Mendel laid out the genetic model later known as Mendelian inheritance or Mendelian genetics. This model provided an alternative to blending inheritance, which was the prevailing theory at the time.
History
Mendel read his paper to the Natural History Society of Brünn. It was published in the Proceedings of the Natural History Society of Brünn the following year.
Mendel's work received little attention from the scientific community and was largely forgotten. It was not until the early 20th century that Mendel's work was rediscovered and his ideas used to help form the modern synthesis.
Analysis
In 1936, the statistician Ronald Fisher used a Pearson's chi-squared test to analyze Mendel's data and concluded that Mendel's results with the predicted ratios were far too perfect, suggesting that adjustments (intentional or unconscious) had been made to the data to make the observations fit the hypothesis.
Later authors have suggested Fisher's analysis was flawed, proposing various statistical and botanical explanations for Mendel's numbers. It is also possible that Mendel's results are "too good" merely because he reported the best subset of his data—Mendel mentioned in his paper that the data were from a subset of his experiments.
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What scientist and monk is best known for his experiments with pea plants?
A. aristotle
B. steiner mendel
C. gregor mendel
D. charles darwin
Answer:
|
|
sciq-794
|
multiple_choice
|
When energy is captured or transformed, it inevitably degrades and becomes what less useful form of energy?
|
[
"heat",
"temperature",
"motion",
"chemical"
] |
A
|
Relavent Documents:
Document 0:::
Energy transformation, also known as energy conversion, is the process of changing energy from one form to another. In physics, energy is a quantity that provides the capacity to perform work or moving (e.g. lifting an object) or provides heat. In addition to being converted, according to the law of conservation of energy, energy is transferable to a different location or object, but it cannot be created or destroyed.
The energy in many of its forms may be used in natural processes, or to provide some service to society such as heating, refrigeration, lighting or performing mechanical work to operate machines. For example, to heat a home, the furnace burns fuel, whose chemical potential energy is converted into thermal energy, which is then transferred to the home's air to raise its temperature.
Limitations in the conversion of thermal energy
Conversions to thermal energy from other forms of energy may occur with 100% efficiency. Conversion among non-thermal forms of energy may occur with fairly high efficiency, though there is always some energy dissipated thermally due to friction and similar processes. Sometimes the efficiency is close to 100%, such as when potential energy is converted to kinetic energy as an object falls in a vacuum. This also applies to the opposite case; for example, an object in an elliptical orbit around another body converts its kinetic energy (speed) into gravitational potential energy (distance from the other object) as it moves away from its parent body. When it reaches the furthest point, it will reverse the process, accelerating and converting potential energy into kinetic. Since space is a near-vacuum, this process has close to 100% efficiency.
Thermal energy is unique because it in most cases (willow) cannot be converted to other forms of energy. Only a difference in the density of thermal/heat energy (temperature) can be used to perform work, and the efficiency of this conversion will be (much) less than 100%. This is because t
Document 1:::
Energy flow is the flow of energy through living things within an ecosystem. All living organisms can be organized into producers and consumers, and those producers and consumers can further be organized into a food chain. Each of the levels within the food chain is a trophic level. In order to more efficiently show the quantity of organisms at each trophic level, these food chains are then organized into trophic pyramids. The arrows in the food chain show that the energy flow is unidirectional, with the head of an arrow indicating the direction of energy flow; energy is lost as heat at each step along the way.
The unidirectional flow of energy and the successive loss of energy as it travels up the food web are patterns in energy flow that are governed by thermodynamics, which is the theory of energy exchange between systems. Trophic dynamics relates to thermodynamics because it deals with the transfer and transformation of energy (originating externally from the sun via solar radiation) to and among organisms.
Energetics and the carbon cycle
The first step in energetics is photosynthesis, wherein water and carbon dioxide from the air are taken in with energy from the sun, and are converted into oxygen and glucose. Cellular respiration is the reverse reaction, wherein oxygen and sugar are taken in and release energy as they are converted back into carbon dioxide and water. The carbon dioxide and water produced by respiration can be recycled back into plants.
Energy loss can be measured either by efficiency (how much energy makes it to the next level), or by biomass (how much living material exists at those levels at one point in time, measured by standing crop). Of all the net primary productivity at the producer trophic level, in general only 10% goes to the next level, the primary consumers, then only 10% of that 10% goes on to the next trophic level, and so on up the food pyramid. Ecological efficiency may be anywhere from 5% to 20% depending on how efficient
Document 2:::
The energy systems language, also referred to as energese, or energy circuit language, or generic systems symbols, is a modelling language used for composing energy flow diagrams in the field of systems ecology. It was developed by Howard T. Odum and colleagues in the 1950s during studies of the tropical forests funded by the United States Atomic Energy Commission.
Design intent
The design intent of the energy systems language was to facilitate the generic depiction of energy flows through any scale system while encompassing the laws of physics, and in particular, the laws of thermodynamics (see energy transformation for an example).
In particular H.T. Odum aimed to produce a language which could facilitate the intellectual analysis, engineering synthesis and management of global systems such as the geobiosphere, and its many subsystems. Within this aim, H.T. Odum had a strong concern that many abstract mathematical models of such systems were not thermodynamically valid. Hence he used analog computers to make system models due to their intrinsic value; that is, the electronic circuits are of value for modelling natural systems which are assumed to obey the laws of energy flow, because, in themselves the circuits, like natural systems, also obey the known laws of energy flow, where the energy form is electrical. However Odum was interested not only in the electronic circuits themselves, but also in how they might be used as formal analogies for modeling other systems which also had energy flowing through them. As a result, Odum did not restrict his inquiry to the analysis and synthesis of any one system in isolation. The discipline that is most often associated with this kind of approach, together with the use of the energy systems language is known as systems ecology.
General characteristics
When applying the electronic circuits (and schematics) to modeling ecological and economic systems, Odum believed that generic categories, or characteristic modules, could
Document 3:::
Conceptual questions or conceptual problems in science, technology, engineering, and mathematics (STEM) education are questions that can be answered based only on the knowledge of relevant concepts, rather than performing extensive calculations. They contrast with most homework and exam problems in science and engineering that typically require plugging in numerical values into previously discussed formulas. Such "plug-and-chug" numerical problems can often be solved correctly by just matching the pattern of the problem to a previously discussed problem and changing the numerical inputs, which requires significant amounts of time to perform the calculations but does not test or deepen the understanding of how the concepts and formulas should work together. Conceptual questions, therefore, provide a good complement to conventional numerical problems because they need minimal or no calculations and instead encourage the students to engage more deeply with the underlying concepts and how they relate to formulas.
Conceptual problems are often formulated as multiple-choice questions, making them easy to use during in-class discussions, particularly when utilizing active learning, peer instruction, and audience response. An example of a conceptual question in undergraduate thermodynamics is provided below:
During adiabatic expansion of an ideal gas, its temperatureincreases
decreases
stays the same
Impossible to tell/need more information
The use of conceptual questions in physics was popularized by Eric Mazur, particularly in the form of multiple-choice tests that he called ConcepTests. In recent years, multiple websites that maintain lists of conceptual questions have been created by instructors for various disciplines. Some books on physics provide many examples of conceptual questions as well.
Multiple conceptual questions can be assembled into a concept inventory to test the working knowledge of students at the beginning of a course or to track the improvement in
Document 4:::
Energy quality is a measure of the ease with which a form of energy can be converted to useful work or to another form of energy: i.e. its content of thermodynamic free energy. A high quality form of energy has a high content of thermodynamic free energy, and therefore a high proportion of it can be converted to work; whereas with low quality forms of energy, only a small proportion can be converted to work, and the remainder is dissipated as heat. The concept of energy quality is also used in ecology, where it is used to track the flow of energy between different trophic levels in a food chain and in thermoeconomics, where it is used as a measure of economic output per unit of energy. Methods of evaluating energy quality often involve developing a ranking of energy qualities in hierarchical order.
Examples: Industrialization, Biology
The consideration of energy quality was a fundamental driver of industrialization from the 18th through 20th centuries. Consider for example the industrialization of New England in the 18th century. This refers to the construction of textile mills containing power looms for weaving cloth. The simplest, most economical and straightforward source of energy was provided by water wheels, extracting energy from a millpond behind a dam on a local creek. If another nearby landowner also decided to build a mill on the same creek, the construction of their dam would lower the overall hydraulic head to power the existing waterwheel, thus hurting power generation and efficiency. This eventually became an issue endemic to the entire region, reducing the overall profitability of older mills as newer ones were built. The search for higher quality energy was a major impetus throughout the 19th and 20th centuries. For example, burning coal to make steam to generate mechanical energy would not have been imaginable in the 18th century; by the end of the 19th century, the use of water wheels was long outmoded. Similarly, the quality of energy from elec
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
When energy is captured or transformed, it inevitably degrades and becomes what less useful form of energy?
A. heat
B. temperature
C. motion
D. chemical
Answer:
|
|
sciq-9448
|
multiple_choice
|
What are organisms that eat just one type of food?
|
[
"hedonists",
"gluttons",
"specialists",
"devotees"
] |
C
|
Relavent Documents:
Document 0:::
Consumer–resource interactions are the core motif of ecological food chains or food webs, and are an umbrella term for a variety of more specialized types of biological species interactions including prey-predator (see predation), host-parasite (see parasitism), plant-herbivore and victim-exploiter systems. These kinds of interactions have been studied and modeled by population ecologists for nearly a century. Species at the bottom of the food chain, such as algae and other autotrophs, consume non-biological resources, such as minerals and nutrients of various kinds, and they derive their energy from light (photons) or chemical sources. Species higher up in the food chain survive by consuming other species and can be classified by what they eat and how they obtain or find their food.
Classification of consumer types
The standard categorization
Various terms have arisen to define consumers by what they eat, such as meat-eating carnivores, fish-eating piscivores, insect-eating insectivores, plant-eating herbivores, seed-eating granivores, and fruit-eating frugivores and omnivores are meat eaters and plant eaters. An extensive classification of consumer categories based on a list of feeding behaviors exists.
The Getz categorization
Another way of categorizing consumers, proposed by South African American ecologist Wayne Getz, is based on a biomass transformation web (BTW) formulation that organizes resources into five components: live and dead animal, live and dead plant, and particulate (i.e. broken down plant and animal) matter. It also distinguishes between consumers that gather their resources by moving across landscapes from those that mine their resources by becoming sessile once they have located a stock of resources large enough for them to feed on during completion of a full life history stage.
In Getz's scheme, words for miners are of Greek etymology and words for gatherers are of Latin etymology. Thus a bestivore, such as a cat, preys on live animal
Document 1:::
A generalist species is able to thrive in a wide variety of environmental conditions and can make use of a variety of different resources (for example, a heterotroph with a varied diet). A specialist species can thrive only in a narrow range of environmental conditions or has a limited diet. Most organisms do not all fit neatly into either group, however. Some species are highly specialized (the most extreme case being monophagous, eating one specific type of food), others less so, and some can tolerate many different environments. In other words, there is a continuum from highly specialized to broadly generalist species.
Description
Omnivores are usually generalists. Herbivores are often specialists, but those that eat a variety of plants may be considered generalists. A well-known example of a specialist animal is the monophagous koala, which subsists almost entirely on eucalyptus leaves. The raccoon is a generalist, because it has a natural range that includes most of North and Central America, and it is omnivorous, eating berries, insects such as butterflies, eggs, and various small animals.
The distinction between generalists and specialists is not limited to animals. For example, some plants require a narrow range of temperatures, soil conditions and precipitation to survive while others can tolerate a broader range of conditions. A cactus could be considered a specialist species. It will die during winters at high latitudes or if it receives too much water.
When body weight is controlled for, specialist feeders such as insectivores and frugivores have larger home ranges than generalists like some folivores (leaf-eaters), whose food-source is less abundant; they need a bigger area for foraging. An example comes from the research of Tim Clutton-Brock, who found that the black-and-white colobus, a folivore generalist, needs a home range of only 15 ha. On the other hand, the more specialized red colobus monkey has a home range of 70 ha, which it requires to
Document 2:::
The trophic level of an organism is the position it occupies in a food web. A food chain is a succession of organisms that eat other organisms and may, in turn, be eaten themselves. The trophic level of an organism is the number of steps it is from the start of the chain. A food web starts at trophic level 1 with primary producers such as plants, can move to herbivores at level 2, carnivores at level 3 or higher, and typically finish with apex predators at level 4 or 5. The path along the chain can form either a one-way flow or a food "web". Ecological communities with higher biodiversity form more complex trophic paths.
The word trophic derives from the Greek τροφή (trophē) referring to food or nourishment.
History
The concept of trophic level was developed by Raymond Lindeman (1942), based on the terminology of August Thienemann (1926): "producers", "consumers", and "reducers" (modified to "decomposers" by Lindeman).
Overview
The three basic ways in which organisms get food are as producers, consumers, and decomposers.
Producers (autotrophs) are typically plants or algae. Plants and algae do not usually eat other organisms, but pull nutrients from the soil or the ocean and manufacture their own food using photosynthesis. For this reason, they are called primary producers. In this way, it is energy from the sun that usually powers the base of the food chain. An exception occurs in deep-sea hydrothermal ecosystems, where there is no sunlight. Here primary producers manufacture food through a process called chemosynthesis.
Consumers (heterotrophs) are species that cannot manufacture their own food and need to consume other organisms. Animals that eat primary producers (like plants) are called herbivores. Animals that eat other animals are called carnivores, and animals that eat both plants and other animals are called omnivores.
Decomposers (detritivores) break down dead plant and animal material and wastes and release it again as energy and nutrients into
Document 3:::
A monogastric organism has a simple single-chambered stomach (one stomach). Examples of monogastric herbivores are horses and rabbits. Examples of monogastric omnivores include humans, pigs, hamsters and rats. Furthermore, there are monogastric carnivores such as cats. A monogastric organism is comparable to ruminant organisms (which has a four-chambered complex stomach), such as cattle, goats, or sheep. Herbivores with monogastric digestion can digest cellulose in their diets by way of symbiotic gut bacteria. However, their ability to extract energy from cellulose digestion is less efficient than in ruminants.
Herbivores digest cellulose by microbial fermentation. Monogastric herbivores which can digest cellulose nearly as well as ruminants are called hindgut fermenters, while ruminants are called foregut fermenters. These are subdivided into two groups based on the relative size of various digestive organs in relationship to the rest of the system: colonic fermenters tend to be larger species such as horses and rhinos, and cecal fermenters are smaller animals such as rabbits and rodents. Great apes derive significant amounts of phytanic acid from the hindgut fermentation of plant materials.
Monogastrics cannot digest the fiber molecule cellulose as efficiently as ruminants, though the ability to digest cellulose varies amongst species.
A monogastric digestive system works as soon as the food enters the mouth. Saliva moistens the food and begins the digestive process. (Note that horses have no (or negligible amounts of) amylase in their saliva). After being swallowed, the food passes from the esophagus into the stomach, where stomach acid and enzymes help to break down the food. Once food leaves the stomach and enters the small intestine, the pancreas secretes enzymes and alkali to neutralize the stomach acid.
Document 4:::
Relatively speaking, the brain consumes an immense amount of energy in comparison to the rest of the body. The mechanisms involved in the transfer of energy from foods to neurons are likely to be fundamental to the control of brain function. Human bodily processes, including the brain, all require both macronutrients, as well as micronutrients.
Insufficient intake of selected vitamins, or certain metabolic disorders, may affect cognitive processes by disrupting the nutrient-dependent processes within the body that are associated with the management of energy in neurons, which can subsequently affect synaptic plasticity, or the ability to encode new memories.
Macronutrients
The human brain requires nutrients obtained from the diet to develop and sustain its physical structure and cognitive functions. Additionally, the brain requires caloric energy predominately derived from the primary macronutrients to operate. The three primary macronutrients include carbohydrates, proteins, and fats. Each macronutrient can impact cognition through multiple mechanisms, including glucose and insulin metabolism, neurotransmitter actions, oxidative stress and inflammation, and the gut-brain axis. Inadequate macronutrient consumption or proportion could impair optimal cognitive functioning and have long-term health implications.
Carbohydrates
Through digestion, dietary carbohydrates are broken down and converted into glucose, which is the sole energy source for the brain. Optimal brain function relies on adequate carbohydrate consumption, as carbohydrates provide the quickest source of glucose for the brain. Glucose deficiencies such as hypoglycaemia reduce available energy for the brain and impair all cognitive processes and performance. Additionally, situations with high cognitive demand, such as learning a new task, increase brain glucose utilization, depleting blood glucose stores and initiating the need for supplementation.
Complex carbohydrates, especially those with high d
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What are organisms that eat just one type of food?
A. hedonists
B. gluttons
C. specialists
D. devotees
Answer:
|
|
sciq-6544
|
multiple_choice
|
What is the enclosed, fluid-filled membrane that surrounds and protects the fetus and is attached to the placenta?
|
[
"womb",
"amniotic sac",
"umbilical sac",
"epithelial sac"
] |
B
|
Relavent Documents:
Document 0:::
The amniotic sac, also called the bag of waters or the membranes, is the sac in which the embryo and later fetus develops in amniotes. It is a thin but tough transparent pair of membranes that hold a developing embryo (and later fetus) until shortly before birth. The inner of these membranes, the amnion, encloses the amniotic cavity, containing the amniotic fluid and the embryo. The outer membrane, the chorion, contains the amnion and is part of the placenta. On the outer side, the amniotic sac is connected to the yolk sac, the allantois, and via the umbilical cord, the placenta.
The yolk sac, amnion, chorion, and allantois are the four extraembryonic membranes that lie outside of the embryo and are involved in providing nutrients and protection to the developing embryo. They form from the inner cell mass; the first to form is the yolk sac followed by the amnion which grows over the developing embryo. The amnion remains an important extraembryonic membrane throughout prenatal development. The third membrane is the allantois, and the fourth is the chorion which surrounds the embryo after about a month and eventually fuses with the amnion.
Amniocentesis is a medical procedure where fluid from the sac is sampled during fetal development, between 15 and 20 weeks of pregnancy, to be used in prenatal diagnosis of chromosomal abnormalities and fetal infections.
Structure
The amniotic cavity is the closed sac between the embryo and the amnion, containing the amniotic fluid. The amniotic cavity is formed by the fusion of the parts of the amniotic fold, which first makes its appearance at the cephalic extremity and subsequently at the caudal end and sides of the embryo. As the amniotic fold rises and fuses over the dorsal aspect of the embryo, the amniotic cavity is formed.
Development
At the beginning of the second week, a cavity appears within the inner cell mass, and when it enlarges, it becomes the amniotic cavity. The floor of the amniotic cavity is formed by the e
Document 1:::
The fetal membranes are the four extraembryonic membranes, associated with the developing embryo, and fetus in humans and other mammals. They are the amnion, chorion, allantois, and yolk sac. The amnion and the chorion are the chorioamniotic membranes that make up the amniotic sac which surrounds and protects the embryo. The fetal membranes are four of six accessory organs developed by the conceptus that are not part of the embryo itself, the other two are the placenta, and the umbilical cord.
Structure
The fetal membranes surround the developing embryo and form the fetal-maternal interface. The fetal membranes are derived from the trophoblast layer (outer layer of cells) of the implanting blastocyst. The trophoblast layer differentiates into amnion and the chorion, which then comprise the fetal membranes. The amnion is the innermost layer and, therefore, contacts the amniotic fluid, the fetus and the umbilical cord. The internal pressure of the amniotic fluid causes the amnion to be passively attached to the chorion. The chorion functions to separate the amnion from the maternal decidua and uterus. The placenta develops from the chorion of the embryo and the uterine tissue of the mother.
Development of the fetal membranes
Initially, the amnion is separated from the chorion by chorionic fluid. The fusion of the amnion and chorion is completed in the human at the 12th week of development.
Microanatomy
From inside to outside, the fetal membranes consist of amnion and chorion. In addition, parts of decidua are often attached to the outside of the chorion.
Amnion
The amnion is avascular, meaning it does not contain its own blood vessels. Therefore, it must obtain necessary nutrients and oxygen from nearby chorionic and amniotic fluid, and fetal surface vessels. The amnion is characterised by cuboidal and columnar epithelial layers. The columnar cells are located in the vicinity of the placenta, whereas the cuboidal cells are found in the periphery. During ea
Document 2:::
The gestational sac is the large cavity of fluid surrounding the embryo. During early embryogenesis it consists of the extraembryonic coelom, also called the chorionic cavity. The gestational sac is normally contained within the uterus. It is the only available structure that can be used to determine if an intrauterine pregnancy exists until the embryo can be identified.
On obstetric ultrasound, the gestational sac is a dark (anechoic) space surrounded by a white (hyperechoic) rim.
Structure
The gestational sac is spherical in shape, and is usually located in the upper part (fundus) of the uterus. By approximately nine weeks of gestational age, due to folding of the trilaminar germ disc, the amniotic sac expands and occupy the majority of the volume of the gestational sac, eventually reducing the extraembryonic coelom (the gestational sac or the chorionic cavity) to a thin layer between the parietal somatopleuric and visceral splanchnopleuric layer of extraembryonic mesoderm.
Development
During embryogenesis, the extraembryonic coelom (or chorionic cavity) that constitutes the gestational sac is a portion of the conceptus consisting of a cavity between Heuser's membrane and the trophoblast.
During formation of the primary yolk sac, some of the migrating hypoblast cells differentiate into mesenchymal cells that fill the space between Heuser's membrane and the trophoblast, forming the extraembryonic mesoderm. As development progresses, small lacunae begin to form within the extraembryonic mesoderm which enlarges to become the extraembryonic coelom.
The Heuser's membrane cells (hypoblast cells) that migrated along the inner cytotrophoblast lining of the blastocoel, secrete an extracellular matrix along the way. Cells of the hypoblast migrate along the outer edges of this reticulum and form the extraembryonic mesoderm; this disrupts the extraembryonic reticulum. Soon pockets form in the reticulum, which ultimately coalesce to form the extraembryonic coelom.
The e
Document 3:::
The placenta of humans, and certain other mammals contains structures known as cotyledons, which transmit fetal blood and allow exchange of oxygen and nutrients with the maternal blood.
Ruminants
The Artiodactyla have a cotyledonary placenta. In this form of placenta the chorionic villi form a number of separate circular structures (cotyledons) which are distributed over the surface of the chorionic sac. Sheep, goats and cattle have between 72 and 125 cotyledons whereas deer have 4-6 larger cotyledons.
Human
The form of the human placenta is generally classified as a discoid placenta. Within this the cotyledons are the approximately 15-25 separations of the decidua basalis of the placenta, separated by placental septa. Each cotyledon consists of a main stem of a chorionic villus as well as its branches and sub-branches.
Vasculature
The cotyledons receive fetal blood from chorionic vessels, which branch off cotyledon vessels into the cotyledons, which, in turn, branch into capillaries. The cotyledons are surrounded by maternal blood, which can exchange oxygen and nutrients with the fetal blood in the capillaries.
Document 4:::
Reichert's membrane is an extraembryonic membrane that forms during early mammalian embryonic development. It forms as a thickened basement membrane to cover the embryo immediately following implantation to give protection to the embryo from the uterine pressures exerted. Reichert's membrane is also important for the maternofetal exchange of nutrients. The membrane collapses once the placenta has fully developed.
Structure
Reichert's membrane is a multilayered, non-vascular, specialised thickened basement membrane that forms on the inner surface of the trophoblast around the time of implantation, and during the formation of the placenta.
It is composed of an extracellular matrix that includes laminin, type IV collagen, and nidogen, and is secreted by embryonic cells in the distal parietal endoderm.
The synthesis of laminin 111 in the embryo contributes to the formation of Reichert's membrane.
Function
Reichert's membrane functions as a buffer space between the embryo and the decidua. This space provides protection to the embryo from varying uterine pressures exerted by smooth muscle contractions of the myometrium.
During post gastrulation Reichert's membrane is necessary for the maternofetal exchange of nutrients. Reichert's membrane encloses the embryo until the amnion develops, and when the placenta is fully developed the membrane collapses.
A major difference in the early formation of the mouse embryo, and that of the human embryo is that in the mouse following implantation the epiblast takes on an egg or cylindrical shape; in the human the epiblast forms into a horizontal, disc-shape the bilaminar disc. A study that looked at this morphological difference between a human embryo initial development and a mouse embryo, concluded that it is likely that Reichert’s membrane is the key regulator of the epiblast’s horizontal growth.
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
What is the enclosed, fluid-filled membrane that surrounds and protects the fetus and is attached to the placenta?
A. womb
B. amniotic sac
C. umbilical sac
D. epithelial sac
Answer:
|
|
sciq-11551
|
multiple_choice
|
Protecting the surface of metal prevents what?
|
[
"corrosion",
"diffusion",
"evaporation",
"extraction"
] |
A
|
Relavent Documents:
Document 0:::
Rustproofing is the prevention or delay of rusting of iron and steel objects, or the permanent protection against corrosion. Typically, the protection is achieved by a process of surface finishing or treatment. Depending on mechanical wear or environmental conditions, the degradation may not be stopped completely, unless the process is periodically repeated. The term is particularly used in the automobile industry.
Vehicle rustproofing
Factory
In the factory, car bodies are protected with special chemical formulations.
Typically, phosphate conversion coatings were used. Some firms galvanized part or all of their car bodies before the primer coat of paint was applied. If a car is body-on-frame, then the frame (chassis) must also be rustproofed. In traditional automotive manufacturing of the early- and mid-20th century, paint was the final part of the rustproofing barrier between the body shell and the atmosphere, except on the underside. On the underside, an underseal rubberized or PVC-based coating was often sprayed on. These products will be breached eventually and can lead to unseen corrosion that spreads underneath the underseal. Old 1960s and 1970s rubberized underseal can become brittle on older cars and is particularly liable to this.
The first electrodeposition primer was developed in the 1950s, but were found to be impractical for widespread use. Revised cathodic automotive electrocoat primer systems were introduced in the 1970s that markedly reduced the problem of corrosion that had been experienced by a vast number of automobiles in the first seven decades of automobile manufacturing. Termed e-coat, "electrocoat automotive primers are applied by totally submerging the assembled car body in a large tank that contains the waterborne e-coat, and the coating is applied through cathodic electrodeposition. This assures nearly 100% coverage of all metal surfaces by the primer. The coating chemistry is waterborne enamel based on epoxy, an aminoalcohol adduct,
Document 1:::
The Zinagizado is an electrochemical process to provide a ferrous metal material with anti-corrosive properties. It involves the application of a constant electric current through a circuit to break the bonds and these are attached to the metal to be coated by forming a surface coating. The alloy used is called Zinag (Zn-Al-Ag); this alloy has excellent mechanical and corrosive properties, so the piece will have increased by 60% of life.
The deposition of Zinag provides environmental protection against corrosion and can be used in covering all kinds of steel metallic materials in contact with a corrosive medium. The anti-corrosive property has been obtained by the corrosion resistance of zinc achieved by the aluminium and silver addition, which is cathodically respect to the iron and steel. Cathodic protection
This process is an innovation by Said Robles Casolco and Adrianni Zanatta.
Patent called: Zinagizado as corrosion process for metals by electrolytic method. No. MX/a/2010/009200, IMPI-Mexico.
Document 2:::
Detectable tape or Underground warning tape is a conductive tape typically applied over buried utilities made of non-conductive materials such as plastic, fiberglass, or cement. It is used because most utility location methods work best on conductive objects, and hence may easily miss structures made of non-conductive materials.
The tape also serves as a physical warning. If uncovered during digging, it alerts the user to an underground object that might be damaged by further excavation. To aid in this, it is typically colored to reflect the nature of the buried object that it is protecting.
It is common for construction specifications to mandate the use of such tape.
The conductive material in detectable tapes is typically aluminium, but there have been studies investigating replacing this with a material which is both magnetic and conductive, to make it detectable to a wider variety of utility location techniques.
See also
Underground Service Alert, an organization that specializes in marking underground utilities.
Document 3:::
The Handle-o-Meter is a testing machine developed by Johnson & Johnson and now manufactured by Thwing-Albert that measures the "handle" of sheeted materials: a combination of its surface friction and flexibility. Originally, it was used to test the durability and flexibility of toilet paper and paper towels.
The test sample is placed over an adjustable slot. The resistance encountered by the penetrator blade as it is moved into the slot by a pivoting arm is measured by the machine.
Details
The data collected when such nonwovens, tissues, toweling, film and textiles are tested has been shown to correlate well with the actual performance of these specific material's performance as a finished product.
Materials are simply placed over the slot that extends across the instrument platform, and then the tester hits test. There are three different test modes which can be applied to the material: single, double, and quadruple. The average is automatically calculated for double or quadruple tests.
Features
Adjustable slot openings
Interchangeable beams
Auto-ranging
2 x 40 LCD display
Statistical Analysis
RS-232 Output and Serial Port
Industry Standards:
ASTM D2923, D6828-02
TAPPI T498
INDA IST 90.3
Document 4:::
In cooking several factors, including materials, techniques, and temperature, can influence the surface chemistry of the chemical reactions and interactions that create food. All of these factors depend on the chemical properties of the surfaces of the materials used. The material properties of cookware, such as hydrophobicity, surface roughness, and conductivity can impact the taste of a dish dramatically. The technique of food preparation alters food in fundamentally different ways, which produce unique textures and flavors. The temperature of food preparation must be considered when choosing the correct ingredients.
Materials in cooking
The interactions between food and pan are very dependent on the material that the pan is made of. Whether or not the pan is hydrophilic or hydrophobic, the heat conductivity and capacity, surface roughness, and more all determine how the food is cooked.
Stainless steel
Stainless steel is considered stainless because it has at least 11% chromium by mass. Chromium is a relatively inert metal and does not rust or react as easily as plain carbon steel. This is what makes it an exceptional material for cooking. It is also fairly inexpensive, but does not have a very high thermal conductivity. From a surface standpoint, this is because of the thin layer of chromium oxide that is formed on the surface. This thin layer protects the metal from rusting or corroding. While it is protective, the oxide layer is not very conductive, which makes cooking food less efficient than it could be. For most cooking applications, high thermal conductivity is desirable to create an evenly heated surface on which to cook. In this way, stainless steel is usually not considered high-grade cookware.
In terms of surface interactions, chromium oxide is polar. The oxygen atoms on the surface have a permanent dipole moment, and are therefore hydrophilic. This means that water will wet it, but oils or other lipids will not.
Cast iron
Cast-iron
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Protecting the surface of metal prevents what?
A. corrosion
B. diffusion
C. evaporation
D. extraction
Answer:
|
|
sciq-3717
|
multiple_choice
|
Photochemical smog consists mainly of what?
|
[
"acid",
"oxygen",
"ozone",
"carbon"
] |
C
|
Relavent Documents:
Document 0:::
A pre-STEM program is a course of study at any two-year college that prepares a student to transfer to a four-year school to earn a bachelor's degree in a STEM field.
Overview
The concept of a pre-STEM program is being developed to address America's need for more college-trained professionals in science, technology, engineering, and mathematics (STEM). It is an innovation meant to fill a gap at community colleges that do not have 'major' degree paths that students identify with on their way to earning an Associates degree. Students must complete a considerable amount of STEM coursework before transferring from a two-year school to a four-year school and earn a baccalaureate degree in a STEM field. Schools with a pre-STEM program are able to identify those students and support them with STEM-specific academic and career advising, increasing the student's chances of going on to earn a STEM baccalaureate degree in a timely fashion.
With over 50% of America's college-bound students starting their college career at public or private two-year school, and with a very small proportion of students who start college at a two-year school matriculating to and earning STEM degrees from four-year schools, pre-STEM programs have great potential for broadening participation in baccalaureate STEM studies.
Example programs
The effectiveness of pre-STEM programs is being investigated by a consortium of schools in Missouri: Moberly Area Community College, St. Charles Community College, Metropolitan Community College, and Truman State University.
A larger group of schools met at the Belknap Springs Meetings in October 2009 to discuss the challenges and opportunities presented by STEM-focused partnerships between 2-year and 4-year schools. Each program represented a two-year school and a four-year school that were trying to increase the number of people who earn a baccalaureate degree in a STEM area through various means, some of which were pre-STEM programs. Other methods includes
Document 1:::
The Inverness Campus is an area in Inverness, Scotland. 5.5 hectares of the site have been designated as an enterprise area for life sciences by the Scottish Government. This designation is intended to encourage research and development in the field of life sciences, by providing incentives to locate at the site.
The enterprise area is part of a larger site, over 200 acres, which will house Inverness College, Scotland's Rural College (SRUC), the University of the Highlands and Islands, a health science centre and sports and other community facilities. The purpose built research hub will provide space for up to 30 staff and researchers, allowing better collaboration.
The Highland Science Academy will be located on the site, a collaboration formed by Highland Council, employers and public bodies. The academy will be aimed towards assisting young people to gain the necessary skills to work in the energy, engineering and life sciences sectors.
History
The site was identified in 2006. Work started to develop the infrastructure on the site in early 2012. A virtual tour was made available in October 2013 to help mark Doors Open Day.
The construction had reached halfway stage in May 2014, meaning that it is on track to open doors to receive its first students in August 2015.
In May 2014, work was due to commence on a building designed to provide office space and laboratories as part of the campus's "life science" sector. Morrison Construction have been appointed to undertake the building work.
Scotland's Rural College (SRUC) will be able to relocate their Inverness-based activities to the Campus. SRUC's research centre for Comparative Epidemiology and Medicine, and Agricultural Business Consultancy services could co-locate with UHI where their activities have complementary themes.
By the start of 2017, there were more than 600 people working at the site.
In June 2021, a new bridge opened connecting Inverness Campus to Inverness Shopping Park. It crosses the Aberdeen
Document 2:::
The STEM (Science, Technology, Engineering, and Mathematics) pipeline is a critical infrastructure for fostering the development of future scientists, engineers, and problem solvers. It's the educational and career pathway that guides individuals from early childhood through to advanced research and innovation in STEM-related fields.
Description
The "pipeline" metaphor is based on the idea that having sufficient graduates requires both having sufficient input of students at the beginning of their studies, and retaining these students through completion of their academic program. The STEM pipeline is a key component of workplace diversity and of workforce development that ensures sufficient qualified candidates are available to fill scientific and technical positions.
The STEM pipeline was promoted in the United States from the 1970s onwards, as “the push for STEM (science, technology, engineering, and mathematics) education appears to have grown from a concern for the low number of future professionals to fill STEM jobs and careers and economic and educational competitiveness.”
Today, this metaphor is commonly used to describe retention problems in STEM fields, called “leaks” in the pipeline. For example, the White House reported in 2012 that 80% of minority groups and women who enroll in a STEM field switch to a non-STEM field or drop out during their undergraduate education. These leaks often vary by field, gender, ethnic and racial identity, socioeconomic background, and other factors, drawing attention to structural inequities involved in STEM education and careers.
Current efforts
The STEM pipeline concept is a useful tool for programs aiming at increasing the total number of graduates, and is especially important in efforts to increase the number of underrepresented minorities and women in STEM fields. Using STEM methodology, educational policymakers can examine the quantity and retention of students at all stages of the K–12 educational process and beyo
Document 3:::
In mathematical psychology and education theory, a knowledge space is a combinatorial structure used to formulate mathematical models describing the progression of a human learner. Knowledge spaces were introduced in 1985 by Jean-Paul Doignon and Jean-Claude Falmagne, and remain in extensive use in the education theory. Modern applications include two computerized tutoring systems, ALEKS and the defunct RATH.
Formally, a knowledge space assumes that a domain of knowledge is a collection of concepts or skills, each of which must be eventually mastered. Not all concepts are interchangeable; some require other concepts as prerequisites. Conversely, competency at one skill may ease the acquisition of another through similarity. A knowledge space marks out which collections of skills are feasible: they can be learned without mastering any other skills. Under reasonable assumptions, the collection of feasible competencies forms the mathematical structure known as an antimatroid.
Researchers and educators usually explore the structure of a discipline's knowledge space as a latent class model.
Motivation
Knowledge Space Theory attempts to address shortcomings of standardized testing when used in educational psychometry. Common tests, such as the SAT and ACT, compress a student's knowledge into a very small range of ordinal ranks, in the process effacing the conceptual dependencies between questions. Consequently, the tests cannot distinguish between true understanding and guesses, nor can they identify a student's particular weaknesses, only the general proportion of skills mastered. The goal of knowledge space theory is to provide a language by which exams can communicate
What the student can do and
What the student is ready to learn.
Model structure
Knowledge Space Theory-based models presume that an educational subject can be modeled as a finite set of concepts, skills, or topics. Each feasible state of knowledge about is then a subset of ; the set of
Document 4:::
Carbon is a primary component of all known life on Earth, representing approximately 45–50% of all dry biomass. Carbon compounds occur naturally in great abundance on Earth. Complex biological molecules consist of carbon atoms bonded with other elements, especially oxygen and hydrogen and frequently also nitrogen, phosphorus, and sulfur (collectively known as CHNOPS).
Because it is lightweight and relatively small in size, carbon molecules are easy for enzymes to manipulate. It is frequently assumed in astrobiology that if life exists elsewhere in the Universe, it will also be carbon-based. Critics refer to this assumption as carbon chauvinism.
Characteristics
Carbon is capable of forming a vast number of compounds, more than any other element, with almost ten million compounds described to date, and yet that number is but a fraction of the number of theoretically possible compounds under standard conditions. The enormous diversity of carbon-containing compounds, known as organic compounds, has led to a distinction between them and compounds that do not contain carbon, known as inorganic compounds. The branch of chemistry that studies organic compounds is known as organic chemistry.
Carbon is the 15th most abundant element in the Earth's crust, and the fourth most abundant element in the universe by mass, after hydrogen, helium, and oxygen. Carbon's widespread abundance, its ability to form stable bonds with numerous other elements, and its unusual ability to form polymers at the temperatures commonly encountered on Earth enables it to serve as a common element of all known living organisms. In a 2018 study, carbon was found to compose approximately 550 billion tons of all life on Earth. It is the second most abundant element in the human body by mass (about 18.5%) after oxygen.
The most important characteristics of carbon as a basis for the chemistry of life are that each carbon atom is capable of forming up to four valence bonds with other atoms simultaneously
The following are multiple choice questions (with answers) about knowledge and skills in advanced master-level STEM courses.
Photochemical smog consists mainly of what?
A. acid
B. oxygen
C. ozone
D. carbon
Answer:
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.