input
stringlengths 2.6k
28.8k
| output
stringlengths 4
150
|
---|---|
Context:
wounds or dead bodies should be examined, not avoided. the book became the first form of literature to help determine the cause of death. in one of song ci ' s accounts ( washing away of wrongs ), the case of a person murdered with a sickle was solved by an investigator who instructed each suspect to bring his sickle to one location. ( he realized it was a sickle by testing various blades on an animal carcass and comparing the wounds. ) flies, attracted by the smell of blood, eventually gathered on a single sickle. in light of this, the owner of that sickle confessed to the murder. the book also described how to distinguish between a drowning ( water in the lungs ) and strangulation ( broken neck cartilage ), and described evidence from examining corpses to determine if a death was caused by murder, suicide or accident. methods from around the world involved saliva and examination of the mouth and tongue to determine innocence or guilt, as a precursor to the polygraph test. in ancient india, some suspects were made to fill their mouths with dried rice and spit it back out. similarly, in ancient china, those accused of a crime would have rice powder placed in their mouths. in ancient middle - eastern cultures, the accused were made to lick hot metal rods briefly. it is thought that these tests had some validity since a guilty person would produce less saliva and thus have a drier mouth ; the accused would be considered guilty if rice was sticking to their mouths in abundance or if their tongues were severely burned due to lack of shielding from saliva. = = education and training = = initial glance, forensic intelligence may appear as a nascent facet of forensic science facilitated by advancements in information technologies such as computers, databases, and data - flow management software. however, a more profound examination reveals that forensic intelligence represents a genuine and emerging inclination among forensic practitioners to actively participate in investigative and policing strategies. in doing so, it elucidates existing practices within scientific literature, advocating for a paradigm shift from the prevailing conception of forensic science as a conglomerate of disciplines merely aiding the criminal justice system. instead, it urges a perspective that views forensic science as a discipline studying the informative potential of traces β remnants of criminal activity. embracing this transformative shift poses a significant challenge for education, necessitating a shift in learners ' mindset to accept concepts and methodologies in forensic intelligence. recent calls advocating for the integration of forensic scientists into the criminal justice system, as well as policing and intelligence missions, undersco
oil umbrella ) ; for calculating the time of death ( allowing for weather and insect activity ) ; described how to wash and examine the dead body to ascertain the reason for death. at that time the book had described methods for distinguishing between suicide and faked suicide. he wrote the book on forensics stating that all wounds or dead bodies should be examined, not avoided. the book became the first form of literature to help determine the cause of death. in one of song ci ' s accounts ( washing away of wrongs ), the case of a person murdered with a sickle was solved by an investigator who instructed each suspect to bring his sickle to one location. ( he realized it was a sickle by testing various blades on an animal carcass and comparing the wounds. ) flies, attracted by the smell of blood, eventually gathered on a single sickle. in light of this, the owner of that sickle confessed to the murder. the book also described how to distinguish between a drowning ( water in the lungs ) and strangulation ( broken neck cartilage ), and described evidence from examining corpses to determine if a death was caused by murder, suicide or accident. methods from around the world involved saliva and examination of the mouth and tongue to determine innocence or guilt, as a precursor to the polygraph test. in ancient india, some suspects were made to fill their mouths with dried rice and spit it back out. similarly, in ancient china, those accused of a crime would have rice powder placed in their mouths. in ancient middle - eastern cultures, the accused were made to lick hot metal rods briefly. it is thought that these tests had some validity since a guilty person would produce less saliva and thus have a drier mouth ; the accused would be considered guilty if rice was sticking to their mouths in abundance or if their tongues were severely burned due to lack of shielding from saliva. = = education and training = = initial glance, forensic intelligence may appear as a nascent facet of forensic science facilitated by advancements in information technologies such as computers, databases, and data - flow management software. however, a more profound examination reveals that forensic intelligence represents a genuine and emerging inclination among forensic practitioners to actively participate in investigative and policing strategies. in doing so, it elucidates existing practices within scientific literature, advocating for a paradigm shift from the prevailing conception of forensic science as a conglomerate of disciplines merely aiding the criminal justice system. instead, it urges a perspective that views forensic science as a discipline studying the informative potential of
, followed by a medical interview and a physical examination. basic diagnostic medical devices ( e. g., stethoscope, tongue depressor ) are typically used. after examining for signs and interviewing for symptoms, the doctor may order medical tests ( e. g., blood tests ), take a biopsy, or prescribe pharmaceutical drugs or other therapies. differential diagnosis methods help to rule out conditions based on the information provided. during the encounter, properly informing the patient of all relevant facts is an important part of the relationship and the development of trust. the medical encounter is then documented in the medical record, which is a legal document in many jurisdictions. follow - ups may be shorter but follow the same general procedure, and specialists follow a similar process. the diagnosis and treatment may take only a few minutes or a few weeks, depending on the complexity of the issue. the components of the medical interview and encounter are : chief complaint ( cc ) : the reason for the current medical visit. these are the symptoms. they are in the patient ' s own words and are recorded along with the duration of each one. also called chief concern or presenting complaint. current activity : occupation, hobbies, what the patient actually does. family history ( fh ) : listing of diseases in the family that may impact the patient. a family tree is sometimes used. history of present illness ( hpi ) : the chronological order of events of symptoms and further clarification of each symptom. distinguishable from history of previous illness, often called past medical history ( pmh ). medical history comprises hpi and pmh. medications ( rx ) : what drugs the patient takes including prescribed, over - the - counter, and home remedies, as well as alternative and herbal medicines or remedies. allergies are also recorded. past medical history ( pmh / pmhx ) : concurrent medical problems, past hospitalizations and operations, injuries, past infectious diseases or vaccinations, history of known allergies. review of systems ( ros ) or systems inquiry : a set of additional questions to ask, which may be missed on hpi : a general enquiry ( have you noticed any weight loss, change in sleep quality, fevers, lumps and bumps? etc. ), followed by questions on the body ' s main organ systems ( heart, lungs, digestive tract, urinary tract, etc. ). social history ( sh ) : birthplace, residences, marital history
the walls of a victim ' s stomach. toxicology, a subfield of forensic chemistry, focuses on detecting and identifying drugs, poisons, and other toxic substances in biological samples. forensic toxicologists work on cases involving drug overdoses, poisoning, and substance abuse. their work is critical in determining whether harmful substances play a role in a person β s death or impairment. read more james marsh was the first to apply this new science to the art of forensics. he was called by the prosecution in a murder trial to give evidence as a chemist in 1832. the defendant, john bodle, was accused of poisoning his grandfather with arsenic - laced coffee. marsh performed the standard test by mixing a suspected sample with hydrogen sulfide and hydrochloric acid. while he was able to detect arsenic as yellow arsenic trisulfide, when it was shown to the jury it had deteriorated, allowing the suspect to be acquitted due to reasonable doubt. annoyed by that, marsh developed a much better test. he combined a sample containing arsenic with sulfuric acid and arsenic - free zinc, resulting in arsine gas. the gas was ignited, and it decomposed to pure metallic arsenic, which, when passed to a cold surface, would appear as a silvery - black deposit. so sensitive was the test, known formally as the marsh test, that it could detect as little as one - fiftieth of a milligram of arsenic. he first described this test in the edinburgh philosophical journal in 1836. = = = ballistics and firearms = = = ballistics is " the science of the motion of projectiles in flight ". in forensic science, analysts examine the patterns left on bullets and cartridge casings after being ejected from a weapon. when fired, a bullet is left with indentations and markings that are unique to the barrel and firing pin of the firearm that ejected the bullet. this examination can help scientists identify possible makes and models of weapons connected to a crime. henry goddard at scotland yard pioneered the use of bullet comparison in 1835. he noticed a flaw in the bullet that killed the victim and was able to trace this back to the mold that was used in the manufacturing process. = = = anthropometry = = = the french police officer alphonse bertillon was the first to apply the anthropological technique of anthropometry to law enforcement, thereby creating an identification system based on physical measurements. before that time, criminals could be identified only by name or photograph. dissatisfied with the ad hoc methods used to identify captured
his sickle to one location. ( he realized it was a sickle by testing various blades on an animal carcass and comparing the wounds. ) flies, attracted by the smell of blood, eventually gathered on a single sickle. in light of this, the owner of that sickle confessed to the murder. the book also described how to distinguish between a drowning ( water in the lungs ) and strangulation ( broken neck cartilage ), and described evidence from examining corpses to determine if a death was caused by murder, suicide or accident. methods from around the world involved saliva and examination of the mouth and tongue to determine innocence or guilt, as a precursor to the polygraph test. in ancient india, some suspects were made to fill their mouths with dried rice and spit it back out. similarly, in ancient china, those accused of a crime would have rice powder placed in their mouths. in ancient middle - eastern cultures, the accused were made to lick hot metal rods briefly. it is thought that these tests had some validity since a guilty person would produce less saliva and thus have a drier mouth ; the accused would be considered guilty if rice was sticking to their mouths in abundance or if their tongues were severely burned due to lack of shielding from saliva. = = education and training = = initial glance, forensic intelligence may appear as a nascent facet of forensic science facilitated by advancements in information technologies such as computers, databases, and data - flow management software. however, a more profound examination reveals that forensic intelligence represents a genuine and emerging inclination among forensic practitioners to actively participate in investigative and policing strategies. in doing so, it elucidates existing practices within scientific literature, advocating for a paradigm shift from the prevailing conception of forensic science as a conglomerate of disciplines merely aiding the criminal justice system. instead, it urges a perspective that views forensic science as a discipline studying the informative potential of traces β remnants of criminal activity. embracing this transformative shift poses a significant challenge for education, necessitating a shift in learners ' mindset to accept concepts and methodologies in forensic intelligence. recent calls advocating for the integration of forensic scientists into the criminal justice system, as well as policing and intelligence missions, underscore the necessity for the establishment of educational and training initiatives in the field of forensic intelligence. this article contends that a discernible gap exists between the perceived and actual comprehension of forensic intelligence among law enforcement and forensic science managers, positing that this asymmetry can be rectified only through educational interventions.
within the military ranges from educational purposes, training exercises and sustainability technology. the technology used for educational purposes within the military are mainly wearables that tracks a soldier ' s vitals. by tracking a soldier ' s heart rate, blood pressure, emotional status, etc. helps the research and development team best help the soldiers. according to chemist, matt coppock, he has started to enhance a soldier ' s lethality by collecting different biorecognition receptors. by doing so it will eliminate emerging environmental threats to the soldiers. with the emergence of virtual reality it is only natural to start creating simulations using vr. this will better prepare the user for whatever situation they are training for. in the military there are combat simulations that soldiers will train on. the reason the military will use vr to train its soldiers is because it is the most interactive / immersive experience the user will feels without being put in a real situation. recent simulations include a soldier wearing a shock belt during a combat simulation. each time they are shot the belt will release a certain amount of electricity directly to the user ' s skin. this is to simulate a shot wound in the most humane way possible. there are many sustainability technologies that military personnel wear in the field. one of which is a boot insert. this insert gauges how soldiers are carrying the weight of their equipment and how daily terrain factors impact their mission panning optimization. these sensors will not only help the military plan the best timeline but will help keep the soldiers at best physical / mental health. = = fashion = = fashionable wearables are " designed garments and accessories that combines aesthetics and style with functional technology. " garments are the interface to the exterior mediated through digital technology. it allows endless possibilities for the dynamic customization of apparel. all clothes have social, psychological and physical functions. however, with the use of technology these functions can be amplified. there are some wearables that are called e - textiles. these are the combination of textiles ( fabric ) and electronic components to create wearable technology within clothing. they are also known as smart textile and digital textile. wearables are made from a functionality perspective or from an aesthetic perspective. when made from a functionality perspective, designers and engineers create wearables to provide convenience to the user. clothing and accessories are used as a tool to provide assistance to the user. designers and engineers are working together to incorporate technology in the manufacturing of garments in order to provide functionalities that can simplify the lives of the user. for example, through smartwatches
masculinity and warmth. the five phases β fire, earth, metal, wood, and water β described a cycle of transformations in nature. the water turned into wood, which turned into the fire when it burned. the ashes left by fire were earth. using these principles, chinese philosophers and doctors explored human anatomy, characterizing organs as predominantly yin or yang, and understood the relationship between the pulse, the heart, and the flow of blood in the body centuries before it became accepted in the west. little evidence survives of how ancient indian cultures around the indus river understood nature, but some of their perspectives may be reflected in the vedas, a set of sacred hindu texts. they reveal a conception of the universe as ever - expanding and constantly being recycled and reformed. surgeons in the ayurvedic tradition saw health and illness as a combination of three humors : wind, bile and phlegm. a healthy life resulted from a balance among these humors. in ayurvedic thought, the body consisted of five elements : earth, water, fire, wind, and space. ayurvedic surgeons performed complex surgeries and developed a detailed understanding of human anatomy. pre - socratic philosophers in ancient greek culture brought natural philosophy a step closer to direct inquiry about cause and effect in nature between 600 and 400 bc. however, an element of magic and mythology remained. natural phenomena such as earthquakes and eclipses were explained increasingly in the context of nature itself instead of being attributed to angry gods. thales of miletus, an early philosopher who lived from 625 to 546 bc, explained earthquakes by theorizing that the world floated on water and that water was the fundamental element in nature. in the 5th century bc, leucippus was an early exponent of atomism, the idea that the world is made up of fundamental indivisible particles. pythagoras applied greek innovations in mathematics to astronomy and suggested that the earth was spherical. = = = aristotelian natural philosophy ( 400 bc β 1100 ad ) = = = later socratic and platonic thought focused on ethics, morals, and art and did not attempt an investigation of the physical world ; plato criticized pre - socratic thinkers as materialists and anti - religionists. aristotle, however, a student of plato who lived from 384 to 322 bc, paid closer attention to the natural world in his philosophy. in his history of animals, he described the inner workings of 110 species, including the stingray, catfish and
equalizing the flow at these places, produces a lowering of the river above the rapids by facilitating the efflux, which may result in the appearance of fresh shoals at the low stage of the river. where, however, narrow rocky reefs or other hard shoals stretch across the bottom of a river and present obstacles to the erosion by the current of the soft materials forming the bed of the river above and below, their removal may result in permanent improvement by enabling the river to deepen its bed by natural scour. the capability of a river to provide a waterway for navigation during the summer or throughout the dry season depends on the depth that can be secured in the channel at the lowest stage. the problem in the dry season is the small discharge and deficiency in scour during this period. a typical solution is to restrict the width of the low - water channel, concentrate all of the flow in it, and also to fix its position so that it is scoured out every year by the floods which follow the deepest part of the bed along the line of the strongest current. this can be effected by closing subsidiary low - water channels with dikes across them, and narrowing the channel at the low stage by low - dipping cross dikes extending from the river banks down the slope and pointing slightly up - stream so as to direct the water flowing over them into a central channel. = = estuarine works = = the needs of navigation may also require that a stable, continuous, navigable channel is prolonged from the navigable river to deep water at the mouth of the estuary. the interaction of river flow and tide needs to be modeled by computer or using scale models, moulded to the configuration of the estuary under consideration and reproducing in miniature the tidal ebb and flow and fresh - water discharge over a bed of fine sand, in which various lines of training walls can be successively inserted. the models should be capable of furnishing valuable indications of the respective effects and comparative merits of the different schemes proposed for works. = = see also = = bridge scour flood control = = references = = = = external links = = u. s. army corps of engineers β civil works program river morphology and stream restoration references - wildland hydrology at the library of congress web archives ( archived 2002 - 08 - 13 )
becomes quite gentle. accordingly, in large basins, rivers in most cases begin as torrents with a variable flow, and end as gently flowing rivers with a comparatively regular discharge. the irregular flow of rivers throughout their course forms one of the main difficulties in devising works for mitigating inundations or for increasing the navigable capabilities of rivers. in tropical countries subject to periodical rains, the rivers are in flood during the rainy season and have hardly any flow during the rest of the year, while in temperate regions, where the rainfall is more evenly distributed throughout the year, evaporation causes the available rainfall to be much less in hot summer weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under
you noticed any weight loss, change in sleep quality, fevers, lumps and bumps? etc. ), followed by questions on the body ' s main organ systems ( heart, lungs, digestive tract, urinary tract, etc. ). social history ( sh ) : birthplace, residences, marital history, social and economic status, habits ( including diet, medications, tobacco, alcohol ). the physical examination is the examination of the patient for medical signs of disease that are objective and observable, in contrast to symptoms that are volunteered by the patient and are not necessarily objectively observable. the healthcare provider uses sight, hearing, touch, and sometimes smell ( e. g., in infection, uremia, diabetic ketoacidosis ). four actions are the basis of physical examination : inspection, palpation ( feel ), percussion ( tap to determine resonance characteristics ), and auscultation ( listen ), generally in that order, although auscultation occurs prior to percussion and palpation for abdominal assessments. the clinical examination involves the study of : abdomen and rectum cardiovascular ( heart and blood vessels ) general appearance of the patient and specific indicators of disease ( nutritional status, presence of jaundice, pallor or clubbing ) genitalia ( and pregnancy if the patient is or could be pregnant ) head, eye, ear, nose, and throat ( heent ) musculoskeletal ( including spine and extremities ) neurological ( consciousness, awareness, brain, vision, cranial nerves, spinal cord and peripheral nerves ) psychiatric ( orientation, mental state, mood, evidence of abnormal perception or thought ). respiratory ( large airways and lungs ) skin vital signs including height, weight, body temperature, blood pressure, pulse, respiration rate, and hemoglobin oxygen saturation it is to likely focus on areas of interest highlighted in the medical history and may not include everything listed above. the treatment plan may include ordering additional medical laboratory tests and medical imaging studies, starting therapy, referral to a specialist, or watchful observation. a follow - up may be advised. depending upon the health insurance plan and the managed care system, various forms of " utilization review ", such as prior authorization of tests, may place barriers on accessing expensive services. the medical decision - making ( mdm ) process includes the analysis and synthesis of all the above data to come up with a list of possible diagnoses ( the differential diagnoses ),
Question: How does a student best avoid becoming dehydrated during a field investigation on a hot day?
A) drink water from a bottle
B) wear a wide-brimmed hat
C) drink a lot before leaving school
D) drink from a clear running stream
|
A) drink water from a bottle
|
Context:
5 to 20 meters. north america and japan use a frequency of 315 mhz, while europe uses 433. 92 and 868 mhz. some models can also remotely start the engine, to warm up the car. a security concern with all keyless entry systems is a replay attack, in which a thief uses a special receiver ( " code grabber " ) to record the radio signal during opening, which can later be replayed to open the door. to prevent this, keyless systems use a rolling code system in which a pseudorandom number generator in the remote control generates a different random key each time it is used. to prevent thieves from simulating the pseudorandom generator to calculate the next key, the radio signal is also encrypted. garage door opener β a short - range handheld transmitter which can open or close a building ' s electrically operated garage door from outside, so the owner can open the door upon arrival, and close it after departure. when a button is pressed the control transmits a coded fsk radio signal to a receiver in the opener, raising or lowering the door. modern openers use 310, 315 or 390 mhz. to prevent a thief using a replay attack, modern openers use a rolling code system. radio - controlled models β a popular hobby is playing with radio - controlled model boats, cars, airplanes, and helicopters ( quadcopters ) which are controlled by radio signals from a handheld console with a joystick. most recent transmitters use the 2. 4 ghz ism band with multiple control channels modulated with pwm, pcm or fsk. wireless doorbell β a residential doorbell that uses wireless technology to eliminate the need to run wires through the building walls. it consists of a doorbell button beside the door containing a small battery powered transmitter. when the doorbell is pressed it sends a signal to a receiver inside the house with a speaker that sounds chimes to indicate someone is at the door. they usually use the 2. 4 ghz ism band. the frequency channel used can usually be changed by the owner in case another nearby doorbell is using the same channel. = = = = scientific research = = = = radio astronomy is the scientific study of radio waves emitted by astronomical objects. radio astronomers use radio telescopes, large radio antennas and receivers, to receive and study the radio waves from astronomical radio sources. since astronomical radio sources are so far away, the radio waves from them are extremely weak, requiring extremely sensitive receivers, and radio telescopes are the most sensitive radio receivers in existence. they use
it is stronger, then demodulates it, extracting the original modulation signal from the modulated carrier wave. the modulation signal is converted by a transducer back to a human - usable form : an audio signal is converted to sound waves by a loudspeaker or earphones, a video signal is converted to images by a display, while a digital signal is applied to a computer or microprocessor, which interacts with human users. the radio waves from many transmitters pass through the air simultaneously without interfering with each other because each transmitter ' s radio waves oscillate at a different frequency, measured in hertz ( hz ), kilohertz ( khz ), megahertz ( mhz ) or gigahertz ( ghz ). the receiving antenna typically picks up the radio signals of many transmitters. the receiver uses tuned circuits to select the radio signal desired out of all the signals picked up by the antenna and reject the others. a tuned circuit acts like a resonator, similar to a tuning fork. it has a natural resonant frequency at which it oscillates. the resonant frequency of the receiver ' s tuned circuit is adjusted by the user to the frequency of the desired radio station ; this is called tuning. the oscillating radio signal from the desired station causes the tuned circuit to oscillate in sympathy, and it passes the signal on to the rest of the receiver. radio signals at other frequencies are blocked by the tuned circuit and not passed on. = = = bandwidth = = = a modulated radio wave, carrying an information signal, occupies a range of frequencies. the information in a radio signal is usually concentrated in narrow frequency bands called sidebands ( sb ) just above and below the carrier frequency. the width in hertz of the frequency range that the radio signal occupies, the highest frequency minus the lowest frequency, is called its bandwidth ( bw ). for any given signal - to - noise ratio, a given bandwidth can carry the same amount of information regardless of where in the radio frequency spectrum it is located ; bandwidth is a measure of information - carrying capacity. the bandwidth required by a radio transmission depends on the data rate of the information being sent, and the spectral efficiency of the modulation method used ; how much data it can transmit in each unit of bandwidth. different types of information signals carried by radio have different data rates. for example, a television signal has a greater data rate than an audio signal. the radio spectrum, the total range of
in order to obtain the keys in this system, a key must be inserted and turned ( like the key at the bottom of the system of the picture ). once the key is turned, the operator may retrieve the remaining keys that will be used to open other doors. once all keys are returned, then the operator will be allowed to take out the original key from the beginning. the key will not turn unless the remaining keys are put back in place. another example is an electric kiln. to prevent access to the inside of an electric kiln, a trapped key system may be used to interlock a disconnecting switch and the kiln door. while the switch is turned on, the key is held by the interlock attached to the disconnecting switch. to open the kiln door, the switch is first opened, which releases the key. the key can then be used to unlock the kiln door. while the key is removed from the switch interlock, a plunger from the interlock mechanically prevents the switch from closing. power cannot be re - applied to the kiln until the kiln door is locked, releasing the key, and the key is then returned to the disconnecting switch interlock. a similar two - part interlock system can be used anywhere it is necessary to ensure the energy supply to a machine is interrupted before the machine is entered for adjustment or maintenance. = = mechanical = = interlocks may be strictly mechanical. an example of a mechanical interlock is a steering wheel of a car. in modern days, most cars have an anti - theft feature that restricts the turning of the steering wheel if the key is not inserted in the ignition. this prevents an individual from pushing the car since the mechanical interlock restricts the directional motion of the front wheels of the car. in the operation of a device such as a press or cutter that is hand fed or the workpiece hand removed, the use of two buttons to actuate the device, one for each hand, greatly reduces the possibility of operation endangering the operator. no such system is fool - proof, and such systems are often augmented by the use of cable β pulled gloves worn by the operator ; these are retracted away from the danger area by the stroke of the machine. a major problem in engineering operator safety is the tendency of operators to ignore safety precautions or even outright disabling forced interlocks due to work pressure and other factors. therefore, such safeties require and perhaps must facilitate operator cooperation. = = electrical =
radar signal transmit a return microwave signal. this causes the aircraft to show up more strongly on the radar screen. the radar which triggers the transponder and receives the return beam, usually mounted on top of the primary radar dish, is called the secondary surveillance radar. since radar cannot measure an aircraft ' s altitude with any accuracy, the transponder also transmits back the aircraft ' s altitude measured by its altimeter, and an id number identifying the aircraft, which is displayed on the radar screen. electronic countermeasures ( ecm ) β military defensive electronic systems designed to degrade enemy radar effectiveness, or deceive it with false information, to prevent enemies from locating local forces. it often consists of powerful microwave transmitters that can mimic enemy radar signals to create false target indications on the enemy radar screens. marine radar β an s or x band radar on ships used to detect nearby ships and obstructions like bridges. a rotating antenna sweeps a vertical fan - shaped beam of microwaves around the water surface surrounding the craft out to the horizon. weather radar β a doppler radar which maps weather precipitation intensities and wind speeds with the echoes returned from raindrops and their radial velocity by their doppler shift. phased - array radar β a radar set that uses a phased array, a computer - controlled antenna that can steer the radar beam quickly to point in different directions without moving the antenna. phased - array radars were developed by the military to track fast - moving missiles and aircraft. they are widely used in military equipment and are now spreading to civilian applications. synthetic aperture radar ( sar ) β a specialized airborne radar set that produces a high - resolution map of ground terrain. the radar is mounted on an aircraft or spacecraft and the radar antenna radiates a beam of radio waves sideways at right angles to the direction of motion, toward the ground. in processing the return radar signal, the motion of the vehicle is used to simulate a large antenna, giving the radar a higher resolution. ground - penetrating radar β a specialized radar instrument that is rolled along the ground surface in a cart and transmits a beam of radio waves into the ground, producing an image of subsurface objects. frequencies from 100 mhz to a few ghz are used. since radio waves cannot penetrate very far into earth, the depth of gpr is limited to about 50 feet. collision avoidance system β a short range radar or lidar system on an automobile or vehicle that detects if the vehicle is about to collide with an object and applies the brakes to
thieves from simulating the pseudorandom generator to calculate the next key, the radio signal is also encrypted. garage door opener β a short - range handheld transmitter which can open or close a building ' s electrically operated garage door from outside, so the owner can open the door upon arrival, and close it after departure. when a button is pressed the control transmits a coded fsk radio signal to a receiver in the opener, raising or lowering the door. modern openers use 310, 315 or 390 mhz. to prevent a thief using a replay attack, modern openers use a rolling code system. radio - controlled models β a popular hobby is playing with radio - controlled model boats, cars, airplanes, and helicopters ( quadcopters ) which are controlled by radio signals from a handheld console with a joystick. most recent transmitters use the 2. 4 ghz ism band with multiple control channels modulated with pwm, pcm or fsk. wireless doorbell β a residential doorbell that uses wireless technology to eliminate the need to run wires through the building walls. it consists of a doorbell button beside the door containing a small battery powered transmitter. when the doorbell is pressed it sends a signal to a receiver inside the house with a speaker that sounds chimes to indicate someone is at the door. they usually use the 2. 4 ghz ism band. the frequency channel used can usually be changed by the owner in case another nearby doorbell is using the same channel. = = = = scientific research = = = = radio astronomy is the scientific study of radio waves emitted by astronomical objects. radio astronomers use radio telescopes, large radio antennas and receivers, to receive and study the radio waves from astronomical radio sources. since astronomical radio sources are so far away, the radio waves from them are extremely weak, requiring extremely sensitive receivers, and radio telescopes are the most sensitive radio receivers in existence. they use large parabolic ( dish ) antennas up to 500 meters ( 2, 000 ft ) in diameter to collect enough radio wave energy to study. the rf front end electronics of the receiver is often cooled by liquid nitrogen to reduce thermal noise. multiple antennas are often linked together in arrays which function as a single antenna, to increase collecting power. in very long baseline interferometry ( vlbi ) radio telescopes on different continents are linked, which can achieve the resolution of an antenna thousands of miles in diameter. remote sensing β in radio, remote sensing is the reception of electromagnetic waves radiated by natural objects or the atmosphere for scientific research. all
and evaporative emissions. nvh engineering ( noise, vibration, and harshness ) : nvh involves customer feedback ( both tactile [ felt ] and audible [ heard ] ) concerning a vehicle. while sound can be interpreted as a rattle, squeal, or hot, a tactile response can be seat vibration or a buzz in the steering wheel. this feedback is generated by components either rubbing, vibrating, or rotating. nvh response can be classified in various ways : powertrain nvh, road noise, wind noise, component noise, and squeak and rattle. note, there are both good and bad nvh qualities. the nvh engineer works to either eliminate bad nvh or change the " bad nvh " to good ( i. e., exhaust tones ). vehicle electronics : automotive electronics is an increasingly important aspect of automotive engineering. modern vehicles employ dozens of electronic systems. these systems are responsible for operational controls such as the throttle, brake and steering controls ; as well as many comfort - and - convenience systems such as the hvac, infotainment, and lighting systems. it would not be possible for automobiles to meet modern safety and fuel - economy requirements without electronic controls. performance : performance is a measurable and testable value of a vehicle ' s ability to perform in various conditions. performance can be considered in a wide variety of tasks, but it generally considers how quickly a car can accelerate ( e. g. standing start 1 / 4 mile elapsed time, 0 β 60 mph, etc. ), its top speed, how short and quickly a car can come to a complete stop from a set speed ( e. g. 70 - 0 mph ), how much g - force a car can generate without losing grip, recorded lap - times, cornering speed, brake fade, etc. performance can also reflect the amount of control in inclement weather ( snow, ice, rain ). shift quality : shift quality is the driver ' s perception of the vehicle to an automatic transmission shift event. this is influenced by the powertrain ( internal combustion engine, transmission ), and the vehicle ( driveline, suspension, engine and powertrain mounts, etc. ) shift feel is both a tactile ( felt ) and audible ( heard ) response of the vehicle. shift quality is experienced as various events : transmission shifts are felt as an upshift at acceleration ( 1 β 2 ), or a downshift maneuver in passing ( 4 β 2
biotechnology to convert a food source into another form. before the time of charles darwin ' s work and life, animal and plant scientists had already used selective breeding. darwin added to that body of work with his scientific observations about the ability of science to change species. these accounts contributed to darwin ' s theory of natural selection. for thousands of years, humans have used selective breeding to improve the production of crops and livestock to use them for food. in selective breeding, organisms with desirable characteristics are mated to produce offspring with the same characteristics. for example, this technique was used with corn to produce the largest and sweetest crops. in the early twentieth century scientists gained a greater understanding of microbiology and explored ways of manufacturing specific products. in 1917, chaim weizmann first used a pure microbiological culture in an industrial process, that of manufacturing corn starch using clostridium acetobutylicum, to produce acetone, which the united kingdom desperately needed to manufacture explosives during world war i. biotechnology has also led to the development of antibiotics. in 1928, alexander fleming discovered the mold penicillium. his work led to the purification of the antibiotic formed by the mold by howard florey, ernst boris chain and norman heatley β to form what we today know as penicillin. in 1940, penicillin became available for medicinal use to treat bacterial infections in humans. the field of modern biotechnology is generally thought of as having been born in 1971 when paul berg ' s ( stanford ) experiments in gene splicing had early success. herbert w. boyer ( univ. calif. at san francisco ) and stanley n. cohen ( stanford ) significantly advanced the new technology in 1972 by transferring genetic material into a bacterium, such that the imported material would be reproduced. the commercial viability of a biotechnology industry was significantly expanded on june 16, 1980, when the united states supreme court ruled that a genetically modified microorganism could be patented in the case of diamond v. chakrabarty. indian - born ananda chakrabarty, working for general electric, had modified a bacterium ( of the genus pseudomonas ) capable of breaking down crude oil, which he proposed to use in treating oil spills. ( chakrabarty ' s work did not involve gene manipulation but rather the transfer of entire organelles between strains of the pseudomonas bacterium ). the mosfet invented at bell labs between 1955 and 1960, two years later, leland c. clark and champ lyons invented the
defective body parts. inside the body, artificial heart valves are in common use with artificial hearts and lungs seeing less common use but under active technology development. other medical devices and aids that can be considered prosthetics include hearing aids, artificial eyes, palatal obturator, gastric bands, and dentures. prostheses are specifically not orthoses, although given certain circumstances a prosthesis might end up performing some or all of the same functionary benefits as an orthosis. prostheses are technically the complete finished item. for instance, a c - leg knee alone is not a prosthesis, but only a prosthetic component. the complete prosthesis would consist of the attachment system to the residual limb β usually a " socket ", and all the attachment hardware components all the way down to and including the terminal device. despite the technical difference, the terms are often used interchangeably. the terms " prosthetic " and " orthotic " are adjectives used to describe devices such as a prosthetic knee. the terms " prosthetics " and " orthotics " are used to describe the respective allied health fields. an occupational therapist ' s role in prosthetics include therapy, training and evaluations. prosthetic training includes orientation to prosthetics components and terminology, donning and doffing, wearing schedule, and how to care for residual limb and the prosthesis. = = = exoskeletons = = = a powered exoskeleton is a wearable mobile machine that is powered by a system of electric motors, pneumatics, levers, hydraulics, or a combination of technologies that allow for limb movement with increased strength and endurance. its design aims to provide back support, sense the user ' s motion, and send a signal to motors which manage the gears. the exoskeleton supports the shoulder, waist and thigh, and assists movement for lifting and holding heavy items, while lowering back stress. = = = adaptive seating and positioning = = = people with balance and motor function challenges often need specialized equipment to sit or stand safely and securely. this equipment is frequently specialized for specific settings such as in a classroom or nursing home. positioning is often important in seating arrangements to ensure that user ' s body pressure is distributed equally without inhibiting movement in a desired way. positioning devices have been developed to aid in allowing people to stand and bear weight on their legs without risk of a fall.
in a voltaic cell, positive ( negative ) ions flow from the low ( high ) potential electrode to the high ( low ) potential electrode, driven by an ` electromotive force ' which points in opposite direction and overcomes the electric force. similarly in a superconductor charge flows in direction opposite to that dictated by the faraday electric field as the magnetic field is expelled in the meissner effect. the puzzle is the same in both cases : what drives electric charges against electromagnetic forces? i propose that the answer is also the same in both cases : kinetic energy lowering, or ` quantum pressure '.
we cut the volume of surface code s gates by 25 % by omitting a hadamard gate.
Question: The primary function of a doorbell circuit is to convert electrical energy into
A) sound.
B) motion.
C) radiant energy.
D) chemical energy.
|
A) sound.
|
Context:
the origin of the martian moons deimos and phobos is controversial. one hypothesis for their origin is that they are captured asteroids, but the mechanism requires an extremely dense martian atmosphere, and the mechanism by which an asteroid in solar orbit could shed sufficient orbital energy to be captured into mars orbit has not been well elucidated. since the discovery by the space probe galileo that the asteroid ida has a moon " dactyl ", a significant number of asteroids have been discovered to have smaller asteroids in orbit about them. the existence of asteroid moons provides a mechanism for the capture of the martian moons ( and the small moons of the outer planets ). when a binary asteroid makes a close approach to a planet, tidal forces can strip the moon from the asteroid. depending on the phasing, the asteroid can then be captured. clearly, the same process can be used to explain the origin of any of the small moons in the solar system.
and measuring radiation levels. the surveyor program conducted uncrewed lunar landings and takeoffs, as well as taking surface and regolith observations. despite the setback caused by the apollo 1 fire, which killed three astronauts, the program proceeded. apollo 8 was the first crewed spacecraft to leave low earth orbit and the first human spaceflight to reach the moon. the crew orbited the moon ten times on december 24 and 25, 1968, and then traveled safely back to earth. the three apollo 8 astronauts β frank borman, james lovell, and william anders β were the first humans to see the earth as a globe in space, the first to witness an earthrise, and the first to see and manually photograph the far side of the moon. the first lunar landing was conducted by apollo 11. commanded by neil armstrong with astronauts buzz aldrin and michael collins, apollo 11 was one of the most significant missions in nasa ' s history, marking the end of the space race when the soviet union gave up its lunar ambitions. as the first human to step on the surface of the moon, neil armstrong uttered the now famous words : that ' s one small step for man, one giant leap for mankind. nasa would conduct six total lunar landings as part of the apollo program, with apollo 17 concluding the program in 1972. = = = = end of apollo = = = = wernher von braun had advocated for nasa to develop a space station since the agency was created. in 1973, following the end of the apollo lunar missions, nasa launched its first space station, skylab, on the final launch of the saturn v. skylab reused a significant amount of apollo and saturn hardware, with a repurposed saturn v third stage serving as the primary module for the space station. damage to skylab during its launch required spacewalks to be performed by the first crew to make it habitable and operational. skylab hosted nine missions and was decommissioned in 1974 and deorbited in 1979, two years prior to the first launch of the space shuttle and any possibility of boosting its orbit. in 1975, the apollo β soyuz mission was the first ever international spaceflight and a major diplomatic accomplishment between the cold war rivals, which also marked the last flight of the apollo capsule. flown in 1975, a us apollo spacecraft docked with a soviet soyuz capsule. = = = interplanetary exploration and space science = = = during the 1960s, nasa started its space science and interplanetary probe program. the mariner program was its flagship
and taken up by the brain. by observing which areas of the brain take up the radioactive isotope, we can see which areas of the brain are more active than other areas. pet has similar spatial resolution to fmri, but it has extremely poor temporal resolution. electroencephalography. eeg measures the electrical fields generated by large populations of neurons in the cortex by placing a series of electrodes on the scalp of the subject. this technique has an extremely high temporal resolution, but a relatively poor spatial resolution. functional magnetic resonance imaging. fmri measures the relative amount of oxygenated blood flowing to different parts of the brain. more oxygenated blood in a particular region is assumed to correlate with an increase in neural activity in that part of the brain. this allows us to localize particular functions within different brain regions. fmri has moderate spatial and temporal resolution. optical imaging. this technique uses infrared transmitters and receivers to measure the amount of light reflectance by blood near different areas of the brain. since oxygenated and deoxygenated blood reflects light by different amounts, we can study which areas are more active ( i. e., those that have more oxygenated blood ). optical imaging has moderate temporal resolution, but poor spatial resolution. it also has the advantage that it is extremely safe and can be used to study infants ' brains. magnetoencephalography. meg measures magnetic fields resulting from cortical activity. it is similar to eeg, except that it has improved spatial resolution since the magnetic fields it measures are not as blurred or attenuated by the scalp, meninges and so forth as the electrical activity measured in eeg is. meg uses squid sensors to detect tiny magnetic fields. = = = computational modeling = = = computational models require a mathematically and logically formal representation of a problem. computer models are used in the simulation and experimental verification of different specific and general properties of intelligence. computational modeling can help us understand the functional organization of a particular cognitive phenomenon. approaches to cognitive modeling can be categorized as : ( 1 ) symbolic, on abstract mental functions of an intelligent mind by means of symbols ; ( 2 ) subsymbolic, on the neural and associative properties of the human brain ; and ( 3 ) across the symbolic β subsymbolic border, including hybrid. symbolic modeling evolved from the computer science paradigms using the technologies of knowledge - based systems, as well as a philosophical perspective ( e. g. " good old - fashioned artificial intelligence " ( gofa
of imaging techniques vary in their temporal ( time - based ) and spatial ( location - based ) resolution. brain imaging is often used in cognitive neuroscience. single - photon emission computed tomography and positron emission tomography. spect and pet use radioactive isotopes, which are injected into the subject ' s bloodstream and taken up by the brain. by observing which areas of the brain take up the radioactive isotope, we can see which areas of the brain are more active than other areas. pet has similar spatial resolution to fmri, but it has extremely poor temporal resolution. electroencephalography. eeg measures the electrical fields generated by large populations of neurons in the cortex by placing a series of electrodes on the scalp of the subject. this technique has an extremely high temporal resolution, but a relatively poor spatial resolution. functional magnetic resonance imaging. fmri measures the relative amount of oxygenated blood flowing to different parts of the brain. more oxygenated blood in a particular region is assumed to correlate with an increase in neural activity in that part of the brain. this allows us to localize particular functions within different brain regions. fmri has moderate spatial and temporal resolution. optical imaging. this technique uses infrared transmitters and receivers to measure the amount of light reflectance by blood near different areas of the brain. since oxygenated and deoxygenated blood reflects light by different amounts, we can study which areas are more active ( i. e., those that have more oxygenated blood ). optical imaging has moderate temporal resolution, but poor spatial resolution. it also has the advantage that it is extremely safe and can be used to study infants ' brains. magnetoencephalography. meg measures magnetic fields resulting from cortical activity. it is similar to eeg, except that it has improved spatial resolution since the magnetic fields it measures are not as blurred or attenuated by the scalp, meninges and so forth as the electrical activity measured in eeg is. meg uses squid sensors to detect tiny magnetic fields. = = = computational modeling = = = computational models require a mathematically and logically formal representation of a problem. computer models are used in the simulation and experimental verification of different specific and general properties of intelligence. computational modeling can help us understand the functional organization of a particular cognitive phenomenon. approaches to cognitive modeling can be categorized as : ( 1 ) symbolic, on abstract mental functions of an intelligent mind by means of symbols ; ( 2 ) subsymbolic, on the neural and associa
parts of australia have been privileged to see dazzling lights in the night sky as the aurora australis ( known as the southern lights ) puts on a show this year. aurorae are significant in australian indigenous astronomical traditions. aboriginal people associate aurorae with fire, death, blood, and omens, sharing many similarities with native american communities.
the location of a repeat plume detected at europa is found to be coincident with the strongest ionosphere detection made by galileo radio occultation in 1997.
fuel cells instead of batteries, and conducted the first american spacewalks and rendezvous operations. the ranger program was started in the 1950s as a response to soviet lunar exploration, however most missions ended in failure. the lunar orbiter program had greater success, mapping the surface in preparation for apollo landings, conducting meteoroid detection, and measuring radiation levels. the surveyor program conducted uncrewed lunar landings and takeoffs, as well as taking surface and regolith observations. despite the setback caused by the apollo 1 fire, which killed three astronauts, the program proceeded. apollo 8 was the first crewed spacecraft to leave low earth orbit and the first human spaceflight to reach the moon. the crew orbited the moon ten times on december 24 and 25, 1968, and then traveled safely back to earth. the three apollo 8 astronauts β frank borman, james lovell, and william anders β were the first humans to see the earth as a globe in space, the first to witness an earthrise, and the first to see and manually photograph the far side of the moon. the first lunar landing was conducted by apollo 11. commanded by neil armstrong with astronauts buzz aldrin and michael collins, apollo 11 was one of the most significant missions in nasa ' s history, marking the end of the space race when the soviet union gave up its lunar ambitions. as the first human to step on the surface of the moon, neil armstrong uttered the now famous words : that ' s one small step for man, one giant leap for mankind. nasa would conduct six total lunar landings as part of the apollo program, with apollo 17 concluding the program in 1972. = = = = end of apollo = = = = wernher von braun had advocated for nasa to develop a space station since the agency was created. in 1973, following the end of the apollo lunar missions, nasa launched its first space station, skylab, on the final launch of the saturn v. skylab reused a significant amount of apollo and saturn hardware, with a repurposed saturn v third stage serving as the primary module for the space station. damage to skylab during its launch required spacewalks to be performed by the first crew to make it habitable and operational. skylab hosted nine missions and was decommissioned in 1974 and deorbited in 1979, two years prior to the first launch of the space shuttle and any possibility of boosting its orbit. in 1975, the apollo β soyuz mission was the first ever international spaceflight and a major diplomatic accomplishment between the cold war
in 2023, 639, 300 people died in france, 35, 900 fewer than in 2022, a year of high mortality. over the last twenty years, from 2004 to 2023, january 3rd was the deadliest day, while august 15th was the least deadly one. elderly people die significantly less often in the summer. deaths are also less frequent on public holidays and sundays. finally, the risk of dying is higher on one ' s birthday, especially for young people.
three major planets, venus, earth, and mercury formed out of the solar nebula. a fourth planetesimal, theia, also formed near earth where it collided in a giant impact, rebounding as the planet mars. during this impact earth lost $ { \ approx } 4 $ \ % of its crust and mantle that is now is found on mars and the moon. at the antipode of the giant impact, $ \ approx $ 60 \ % of earth ' s crust, atmosphere, and a large amount of mantle were ejected into space forming the moon. the lost crust never reformed and became the earth ' s ocean basins. the theia impact site corresponds to indian ocean gravitational anomaly on earth and the hellas basin on mars. the dynamics of the giant impact are consistent with the rotational rates and axial tilts of both earth and mars. the giant impact removed sufficient co $ _ 2 $ from earth ' s atmosphere to avoid a runaway greenhouse effect, initiated plate tectonics, and gave life time to form near geothermal vents at the continental margins. mercury formed near venus where on a close approach it was slingshot into the sun ' s convective zone losing 94 \ % of its mass, much of which remains there today. black carbon, from co $ _ 2 $ decomposed by the intense heat, is still found on the surface of mercury. arriving at 616 km / s, mercury dramatically altered the sun ' s rotational energy, explaining both its anomalously slow rotation rate and axial tilt. these results are quantitatively supported by mass balances, the current locations of the terrestrial planets, and the orientations of their major orbital axes.
##hography is quite small, large area patterns must be created by stitching together the small fields. ion track technology is a deep cutting tool with a resolution limit around 8 nm applicable to radiation resistant minerals, glasses and polymers. it is capable of generating holes in thin films without any development process. structural depth can be defined either by ion range or by material thickness. aspect ratios up to several 104 can be reached. the technique can shape and texture materials at a defined inclination angle. random pattern, single - ion track structures and an aimed pattern consisting of individual single tracks can be generated. x - ray lithography is a process used in the electronic industry to selectively remove parts of a thin film. it uses x - rays to transfer a geometric pattern from a mask to a light - sensitive chemical photoresist, or simply " resist ", on the substrate. a series of chemical treatments then engraves the produced pattern into the material underneath the photoresist. diamond patterning is a method of forming diamond mems. it is achieved by the lithographic application of diamond films to a substrate such as silicon. the patterns can be formed by selective deposition through a silicon dioxide mask, or by deposition followed by micromachining or focused ion beam milling. = = = etching processes = = = there are two basic categories of etching processes : wet etching and dry etching. in the former, the material is dissolved when immersed in a chemical solution. in the latter, the material is sputtered or dissolved using reactive ions or a vapor phase etchant. = = = = wet etching = = = = wet chemical etching consists of the selective removal of material by dipping a substrate into a solution that dissolves it. the chemical nature of this etching process provides good selectivity, which means the etching rate of the target material is considerably higher than the mask material if selected carefully. wet etching can be performed using either isotropic wet etchants or anisotropic wet etchants. isotropic wet etchant etch in all directions of the crystalline silicon at approximately equal rates. anisotropic wet etchants preferably etch along certain crystal planes at faster rates than other planes, thereby allowing more complicated 3 - d microstructures to be implemented. wet anisotropic etchants are often used in conjunction with boron etch stops wherein the surface of the silicon is heavily doped with boron resulting in a silicon material layer that is
Question: A lunar eclipse occurs when the Moon passes through Earth's shadow. A lunar eclipse can only occur during a
A) full moon.
B) new moon.
C) first quarter moon.
D) last quarter moon.
|
A) full moon.
|
Context:
time - dependent distribution of the global extinction of megafauna is compared with the growth of human population. there is no correlation between the two processes. furthermore, the size of human population and its growth rate were far too small to have any significant impact on the environment and on the life of megafauna.
onset of electro - chemical corrosion. similar problems are encountered in coastal and offshore structures. = = = anti - fouling = = = anti - fouling is the process of eliminating obstructive organisms from essential components of seawater systems. depending on the nature and location of marine growth, this process is performed in a number of different ways : marine organisms may grow and attach to the surfaces of the outboard suction inlets used to obtain water for cooling systems. electro - chlorination involves running high electrical current through sea water, altering the water ' s chemical composition to create sodium hypochlorite, purging any bio - matter. an electrolytic method of anti - fouling involves running electrical current through two anodes ( scardino, 2009 ). these anodes typically consist of copper and aluminum ( or alternatively, iron ). the first metal, copper anode, releases its ion into the water, creating an environment that is too toxic for bio - matter. the second metal, aluminum, coats the inside of the pipes to prevent corrosion. other forms of marine growth such as mussels and algae may attach themselves to the bottom of a ship ' s hull. this growth interferes with the smoothness and uniformity of the ship ' s hull, causing the ship to have a less hydrodynamic shape that causes it to be slower and less fuel - efficient. marine growth on the hull can be remedied by using special paint that prevents the growth of such organisms. = = = pollution control = = = = = = = sulfur emission = = = = the burning of marine fuels releases harmful pollutants into the atmosphere. ships burn marine diesel in addition to heavy fuel oil. heavy fuel oil, being the heaviest of refined oils, releases sulfur dioxide when burned. sulfur dioxide emissions have the potential to raise atmospheric and ocean acidity causing harm to marine life. however, heavy fuel oil may only be burned in international waters due to the pollution created. it is commercially advantageous due to the cost effectiveness compared to other marine fuels. it is prospected that heavy fuel oil will be phased out of commercial use by the year 2020 ( smith, 2018 ). = = = = oil and water discharge = = = = water, oil, and other substances collect at the bottom of the ship in what is known as the bilge. bilge water is pumped overboard, but must pass a pollution threshold test of 15 ppm ( parts per million ) of oil to be discharged. water is tested
speculated that the church would fall apart, while others expected it to continue just as it had before. as it was, the movement continued to grow in the first few decades after 1910. the manual of the mother church prohibits the church from publishing membership figures, and it is not clear exactly when the height of the movement was. a 1936 census counted c. 268, 915 christian scientists in the united states ( 2, 098 per million ), and rodney stark believes this to be close to the height. however, the number of christian science churches continued to increase until around 1960, at which point there was a reversal and, since then, many churches have closed their doors. the number of christian science practitioners in the united states began to decline in the 1940s according to stark. according to j. gordon melton, in 1972 there were 3, 237 congregations worldwide, of which roughly 2, 400 were in the united states ; and, in the following ten years, about 200 congregations were closed. during the years after eddy ' s death, the church has gone through a number of hardships and controversies. this included attempts to make practicing christian science illegal in the united states and elsewhere ; a period known as the great litigation which involved two intertwined lawsuits regarding church governance ; persecution under the nazi and communist regimes in germany and the imperial regime in japan ; a series of lawsuits involving the deaths of members of the church, most notably some children ; and a controversial decision to publish a book by bliss knapp. in conjunction with the knapp book controversy, there was controversy within the church involving the monitor channel, part of the christian science monitor which had been losing money, and which eventually led to the channel shutting down. acknowledging their earlier mistake, of accepting a multi - million dollar publishing incentive to offset broadcasting losses, the christian science board of directors, with the concurrence of the trustees of the christian science publishing society, withdrew destiny of the mother church from publication in september 2023. in addition, it has since its beginning been branded as a cult by more fundamentalist strains of christianity, and attracted significant opposition as a result. a number of independent teachers and alternative movements of christian science have emerged since its founding, but none of these individuals or groups have achieved the prominence of the christian science church. despite the hardships and controversies, many christian science churches and reading rooms remain in existence around the world, and, in recent years, there have been reports of the religion growing in africa, though it remains significantly behind other evangelical groups. the christian science
the magellanic clouds were known before magellan ' s voyage exactly 500 years ago, and were not given that name by magellan himself or his chronicler antonio pigafetta. they were, of course, already known by local populations in south america, such as the mapuche and tupi - guaranis. the portuguese called them clouds of the cape, and scientific circles had long used the name of nubecula minor and major. we trace how and when the name magellanic clouds came into common usage by following the history of exploration of the southern hemisphere and the southern sky by european explorers. while the name of magellan was quickly associated to the strait he discovered ( within about 20 years only ), the clouds got their final scientific name only at the end of the 19th century, when scientists finally abandoned latin as their communication language.
generation of direct current in zigzag carbon nanotubes due to harmonic mixing of two coherent electromagnetic waves is being considered. the electromagnetic waves have commensurate frequencies of omega and two omega. the rectification of the waves at high frequencies is quite smooth whiles at low frequencies there are some fluctuations. the nonohmicity observed in the i - vcharacteristics is attributed to the nonparabolicity of the electron energy band which is very strong in carbon nanotubes because of high stark component. it is observed that the current falls off faster at lower electric field than the case in superlattice. for omega tau equal to two? the external electric field strength emax for the observation of negative differential conductivity occurs around 1. 03x10e6 v / m which is quite weak. it is interesting to note that the peak of the curve shifts to the left with increasing value of omega tau?
do - 178b was based on the consensus of the avionic software community as it existed in 1992. twenty two years after publication, we have no publically available experimental data as to its efficacy. it appears to work extremely well, since there have been no hull loss accidents in passenger service ascribed to software failure. this is a comforting and surprising result. however, if we don ' t know why do - 178b works so well, there is a danger that we could stop doing something that really matters, which could lead to an accident.
material is demonstrated by northwestern university, consisting of 100 trillion bonds per square centimetre, which its creators describe as having exceptional flexibility and strength. adding just 2. 5 % of the new material to ultem boosted the latter ' s tensile modulus by 45 %. the air monitoring station at mauna loa observatory in hawaii reports that co2 jumped by 3. 58 parts per million ( ppm ) in 2024, exceeding the previous record of 3. 36 ppm set in 2023. the global atmospheric concentration of co2 is now at 427 ppm, more than 50 % higher than the pre - industrial level. 21 january coral bleaching on the southern great barrier reef in early 2024 is reported to have struck 80 % of colonies, with some coral genera, such as acropora, experiencing a 95 % mortality rate. more than a third ( 34 % ) of the arctic - boreal zone is now reported to be a source of carbon emissions, rather than a carbon sink, a figure that rises to 40 % when including emissions from fires. the exoplanet wasp - 127b is discovered to have wind speeds of up to 33, 000 km / h, the fastest jetstream of its kind ever measured. 22 january β the second trump administration imposes an immediate freeze on scientific grants, communications, hiring, and meetings at the national institutes of health ( nih ) β by far the biggest supporter of biomedical research worldwide β impacting $ 47. 4 billion worth of activities. 23 january machine learning and 3d printing are used at the university of toronto to design nano - architected materials exhibiting the strength of carbon steel but the lightness of styrofoam. a study of adults with attention deficit hyperactivity disorder ( adhd ) finds that the condition may reduce life expectancy by 4. 5 to 9 years for men, and 6. 5 to 11 years for women. 24 january β a study by the university of birmingham finds that electric vehicles now have an average lifespan of 18. 4 years, outlasting the average diesel vehicle at 16. 8 years and almost matching the average petrol vehicle at 18. 7 years. 29 january β the european space agency ( esa ) announces that it has begun monitoring the asteroid 2024 yr4, which at the time had a 1 in 77 ( 1. 3 % ) chance of impacting earth on 22 december 2032. = = = february = = = 3 february β researchers in berkeley and cambridge attach copper nanoflower catalysts on per
a statistical study of the environment around polar ring galaxies is presented. two kinds of search are performed : 1 ) a study of the concentration and diameters of all the objects surrounding the polar rings, within a search field 5 times the ring diameter. new magnitudes for polar ring galaxies are presented. 2 ) a search, in a wider field, for galaxies of similar size that may have encountered the polar ring host galaxy in a time of the order of 1 gyr. differently from the results of similar searches in the fields of active galaxies, the environment of the polar ring galaxies seems to be similar to that of normal galaxies. this result may give support to the models suggesting long times for formation and evolution of the rings. if the rings are old ( and stable or in equilibrium ), no traces of the past interaction are expected in their surroundings. in addition, the formation of massive polar rings, too big to derive from the ingestion of a present - day dwarf galaxy, may be easily placed in epochs with a higher number of gas - rich galaxies.
the paper is withdrawn by the author because it is superseded by cond - mat / 0303357.
accelerating natural degradation. in 1969, environmental science once again became a household term after two striking disasters : ohio ' s cuyahoga river caught fire due to the amount of pollution in its waters and a santa barbara oil spill endangered thousands of marine animals, both receiving prolific media coverage. consequently, the united states passed an abundance of legislation, including the clean water act and the great lakes water quality agreement. the following year, in 1970, the first ever earth day was celebrated worldwide and the united states environmental protection agency ( epa ) was formed, legitimizing the study of environmental science in government policy. in the next two years, the united nations created the united nations environment programme ( unep ) in stockholm, sweden to address global environmental degradation. much of the interest in environmental science throughout the 1970s and the 1980s was characterized by major disasters and social movements. in 1978, hundreds of people were relocated from love canal, new york after carcinogenic pollutants were found to be buried underground near residential areas. the next year, in 1979, the nuclear power plant on three mile island in pennsylvania suffered a meltdown and raised concerns about the dangers of radioactive waste and the safety of nuclear energy. in response to landfills and toxic waste often disposed of near their homes, the official environmental justice movement was started by a black community in north carolina in 1982. two years later, the toxic methyl isocyanate gas was released to the public from a power plant disaster in bhopal, india, harming hundreds of thousands of people living near the disaster site, the effects of which are still felt today. in a groundbreaking discovery in 1985, a british team of researchers studying antarctica found evidence of a hole in the ozone layer, inspiring global agreements banning the use of chlorofluorocarbons ( cfcs ), which were previously used in nearly all aerosols and refrigerants. notably, in 1986, the meltdown at the chernobyl nuclear power plant in ukraine released radioactive waste to the public, leading to international studies on the ramifications of environmental disasters. over the next couple of years, the brundtland commission ( previously known as the world commission on environment and development ) published a report titled our common future and the montreal protocol formed the international panel on climate change ( ipcc ) as international communication focused on finding solutions for climate change and degradation. in the late 1980s, the exxon valdez company was fined for spilling large quantities of crude oil off the coast of alaska and the resulting cleanup, involving the work
Question: A recent study concluded that about 90% of all large predatory fish such as tuna have disappeared from the world's oceans in the past 50 years. Which of these is most likely the direct cause of this loss?
A) Urban development near the coast
B) Commercial harvesting from the ocean
C) Climate change on Earth
D) Acid precipitation over the oceans
|
B) Commercial harvesting from the ocean
|
Context:
. the phase of matter is defined by the phase transition, which is when energy put into or taken out of the system goes into rearranging the structure of the system, instead of changing the bulk conditions. sometimes the distinction between phases can be continuous instead of having a discrete boundary ; in this case the matter is considered to be in a supercritical state. when three states meet based on the conditions, it is known as a triple point and since this is invariant, it is a convenient way to define a set of conditions. the most familiar examples of phases are solids, liquids, and gases. many substances exhibit multiple solid phases. for example, there are three phases of solid iron ( alpha, gamma, and delta ) that vary based on temperature and pressure. a principal difference between solid phases is the crystal structure, or arrangement, of the atoms. another phase commonly encountered in the study of chemistry is the aqueous phase, which is the state of substances dissolved in aqueous solution ( that is, in water ). less familiar phases include plasmas, bose β einstein condensates and fermionic condensates and the paramagnetic and ferromagnetic phases of magnetic materials. while most familiar phases deal with three - dimensional systems, it is also possible to define analogs in two - dimensional systems, which has received attention for its relevance to systems in biology. = = = bonding = = = atoms sticking together in molecules or crystals are said to be bonded with one another. a chemical bond may be visualized as the multipole balance between the positive charges in the nuclei and the negative charges oscillating about them. more than simple attraction and repulsion, the energies and distributions characterize the availability of an electron to bond to another atom. the chemical bond can be a covalent bond, an ionic bond, a hydrogen bond or just because of van der waals force. each of these kinds of bonds is ascribed to some potential. these potentials create the interactions which hold atoms together in molecules or crystals. in many simple compounds, valence bond theory, the valence shell electron pair repulsion model ( vsepr ), and the concept of oxidation number can be used to explain molecular structure and composition. an ionic bond is formed when a metal loses one or more of its electrons, becoming a positively charged cation, and the electrons are then gained by the non - metal atom, becoming a negatively charged anion. the two oppositely charged ions attract one another, and the ionic bond
classifications ; however, some more exotic phases are incompatible with certain chemical properties. a phase is a set of states of a chemical system that have similar bulk structural properties, over a range of conditions, such as pressure or temperature. physical properties, such as density and refractive index tend to fall within values characteristic of the phase. the phase of matter is defined by the phase transition, which is when energy put into or taken out of the system goes into rearranging the structure of the system, instead of changing the bulk conditions. sometimes the distinction between phases can be continuous instead of having a discrete boundary ; in this case the matter is considered to be in a supercritical state. when three states meet based on the conditions, it is known as a triple point and since this is invariant, it is a convenient way to define a set of conditions. the most familiar examples of phases are solids, liquids, and gases. many substances exhibit multiple solid phases. for example, there are three phases of solid iron ( alpha, gamma, and delta ) that vary based on temperature and pressure. a principal difference between solid phases is the crystal structure, or arrangement, of the atoms. another phase commonly encountered in the study of chemistry is the aqueous phase, which is the state of substances dissolved in aqueous solution ( that is, in water ). less familiar phases include plasmas, bose β einstein condensates and fermionic condensates and the paramagnetic and ferromagnetic phases of magnetic materials. while most familiar phases deal with three - dimensional systems, it is also possible to define analogs in two - dimensional systems, which has received attention for its relevance to systems in biology. = = = bonding = = = atoms sticking together in molecules or crystals are said to be bonded with one another. a chemical bond may be visualized as the multipole balance between the positive charges in the nuclei and the negative charges oscillating about them. more than simple attraction and repulsion, the energies and distributions characterize the availability of an electron to bond to another atom. the chemical bond can be a covalent bond, an ionic bond, a hydrogen bond or just because of van der waals force. each of these kinds of bonds is ascribed to some potential. these potentials create the interactions which hold atoms together in molecules or crystals. in many simple compounds, valence bond theory, the valence shell electron pair repulsion model ( vsepr ), and the concept of oxidation number can be used
each packing of r ^ d by translates of the unit cube [ 0, 1 ) ^ d admits a decomposition into at most two parts such that if a translate of the unit cube is covered by one of them, then it also belongs to such a part.
##trahedron, cube, octahedron, dodecahedron, or icosahedron. " in logic, the extension of a predicate is the set of all objects for which the predicate is true. further, the logical principle of extensionality judges two objects to objects to be equal if they satisfy the same external properties. since, by the axiom, two sets are defined to be equal if they satisfy membership, sets are extentional. jose ferreiros credits richard dedekind for being the first to explicitly state the principle, although he does not assert it as a definition : it very frequently happens that different things a, b, c... considered for any reason under a common point of view, are collected together in the mind, and one then says that they form a system s ; one calls the things a, b, c... the elements of the system s, they are contained in s ; conversely, s consists of these elements. such a system s ( or a collection, a manifold, a totality ), as an object of our thought, is likewise a thing ; it is completely determined when, for every thing, it is determined whether it is an element of s or not. = = = background = = = around the turn of the 20th century, mathematics faced several paradoxes and counter - intuitive results. for example, russell ' s paradox showed a contradiction of naive set theory, it was shown that the parallel postulate cannot be proved, the existence of mathematical objects that cannot be computed or explicitly described, and the existence of theorems of arithmetic that cannot be proved with peano arithmetic. the result was a foundational crisis of mathematics. the resolution of this crisis involved the rise of a new mathematical discipline called mathematical logic, which studies formal logic within mathematics. subsequent discoveries in the 20th century then stabilized the foundations of mathematics into a coherent framework valid for all mathematics. this framework is based on a systematic use of axiomatic method and on set theory, specifically zermelo β fraenkel set theory, developed by ernst zermelo and abraham fraenkel. this set theory ( and set theory in general ) is now considered the most common foundation of mathematics. = = = set equality based on first - order logic with equality = = = in first - order logic with equality ( see Β§ axioms ), the axiom of extensionality states that two sets that contain the same elements are the same set. logic axiom : x = y [UNK] [UNK] z, ( z
necessary and sufficient conditions for a term to apply to an object. for example : " a platonic solid is a convex, regular polyhedron in three - dimensional euclidean space. " an extensional definition instead lists all objects where the term applies. for example : " a platonic solid is one of the following : tetrahedron, cube, octahedron, dodecahedron, or icosahedron. " in logic, the extension of a predicate is the set of all objects for which the predicate is true. further, the logical principle of extensionality judges two objects to objects to be equal if they satisfy the same external properties. since, by the axiom, two sets are defined to be equal if they satisfy membership, sets are extentional. jose ferreiros credits richard dedekind for being the first to explicitly state the principle, although he does not assert it as a definition : it very frequently happens that different things a, b, c... considered for any reason under a common point of view, are collected together in the mind, and one then says that they form a system s ; one calls the things a, b, c... the elements of the system s, they are contained in s ; conversely, s consists of these elements. such a system s ( or a collection, a manifold, a totality ), as an object of our thought, is likewise a thing ; it is completely determined when, for every thing, it is determined whether it is an element of s or not. = = = background = = = around the turn of the 20th century, mathematics faced several paradoxes and counter - intuitive results. for example, russell ' s paradox showed a contradiction of naive set theory, it was shown that the parallel postulate cannot be proved, the existence of mathematical objects that cannot be computed or explicitly described, and the existence of theorems of arithmetic that cannot be proved with peano arithmetic. the result was a foundational crisis of mathematics. the resolution of this crisis involved the rise of a new mathematical discipline called mathematical logic, which studies formal logic within mathematics. subsequent discoveries in the 20th century then stabilized the foundations of mathematics into a coherent framework valid for all mathematics. this framework is based on a systematic use of axiomatic method and on set theory, specifically zermelo β fraenkel set theory, developed by ernst zermelo and abraham fraenkel. this set theory ( and set theory in general ) is now considered the most common foundation of mathematics
building block. ceramics β not to be confused with raw, unfired clay β are usually seen in crystalline form. the vast majority of commercial glasses contain a metal oxide fused with silica. at the high temperatures used to prepare glass, the material is a viscous liquid which solidifies into a disordered state upon cooling. windowpanes and eyeglasses are important examples. fibers of glass are also used for long - range telecommunication and optical transmission. scratch resistant corning gorilla glass is a well - known example of the application of materials science to drastically improve the properties of common components. engineering ceramics are known for their stiffness and stability under high temperatures, compression and electrical stress. alumina, silicon carbide, and tungsten carbide are made from a fine powder of their constituents in a process of sintering with a binder. hot pressing provides higher density material. chemical vapor deposition can place a film of a ceramic on another material. cermets are ceramic particles containing some metals. the wear resistance of tools is derived from cemented carbides with the metal phase of cobalt and nickel typically added to modify properties. ceramics can be significantly strengthened for engineering applications using the principle of crack deflection. this process involves the strategic addition of second - phase particles within a ceramic matrix, optimizing their shape, size, and distribution to direct and control crack propagation. this approach enhances fracture toughness, paving the way for the creation of advanced, high - performance ceramics in various industries. = = = composites = = = another application of materials science in industry is making composite materials. these are structured materials composed of two or more macroscopic phases. applications range from structural elements such as steel - reinforced concrete, to the thermal insulating tiles, which play a key and integral role in nasa ' s space shuttle thermal protection system, which is used to protect the surface of the shuttle from the heat of re - entry into the earth ' s atmosphere. one example is reinforced carbon - carbon ( rcc ), the light gray material, which withstands re - entry temperatures up to 1, 510 Β°c ( 2, 750 Β°f ) and protects the space shuttle ' s wing leading edges and nose cap. rcc is a laminated composite material made from graphite rayon cloth and impregnated with a phenolic resin. after curing at high temperature in an autoclave, the laminate is pyrolized to convert the resin to carbon, impregnated with furfuryl alcohol in a
and is invariant under the process of counting. an identity is an equation that remains true for all values of its variables. there are also inequalities that remain true when the values of their variables change. the distance between two points on a number line is not changed by adding the same quantity to both numbers. on the other hand, multiplication does not have this same property, as distance is not invariant under multiplication. angles and ratios of distances are invariant under scalings, rotations, translations and reflections. these transformations produce similar shapes, which is the basis of trigonometry. in contrast, angles and ratios are not invariant under non - uniform scaling ( such as stretching ). the sum of a triangle ' s interior angles ( 180Β° ) is invariant under all the above operations. as another example, all circles are similar : they can be transformed into each other and the ratio of the circumference to the diameter is invariant ( denoted by the greek letter Ο ( pi ) ). some more complicated examples : the real part and the absolute value of a complex number are invariant under complex conjugation. the tricolorability of knots. the degree of a polynomial is invariant under a linear change of variables. the dimension and homology groups of a topological object are invariant under homeomorphism. the number of fixed points of a dynamical system is invariant under many mathematical operations. euclidean distance is invariant under orthogonal transformations. area is invariant under linear maps which have determinant Β±1 ( see equiareal map Β§ linear transformations ). some invariants of projective transformations include collinearity of three or more points, concurrency of three or more lines, conic sections, and the cross - ratio. the determinant, trace, eigenvectors, and eigenvalues of a linear endomorphism are invariant under a change of basis. in other words, the spectrum of a matrix is invariant under a change of basis. the principal invariants of tensors do not change with rotation of the coordinate system ( see invariants of tensors ). the singular values of a matrix are invariant under orthogonal transformations. lebesgue measure is invariant under translations. the variance of a probability distribution is invariant under translations of the real line. hence the variance of a random variable is unchanged after the addition of a constant. the fixed points of a transformation are the elements in the domain that are invariant under the transformation. they may, depending on the application, be called symmetric with respect to that transformation. for example,
##tering - based methods are simple ( " sinter " has roots in the english " cinder " ). the firing is done at a temperature below the melting point of the ceramic. once a roughly - held - together object called a " green body " is made, it is fired in a kiln, where atomic and molecular diffusion processes give rise to significant changes in the primary microstructural features. this includes the gradual elimination of porosity, which is typically accompanied by a net shrinkage and overall densification of the component. thus, the pores in the object may close up, resulting in a denser product of significantly greater strength and fracture toughness. another major change in the body during the firing or sintering process will be the establishment of the polycrystalline nature of the solid. significant grain growth tends to occur during sintering, with this growth depending on temperature and duration of the sintering process. the growth of grains will result in some form of grain size distribution, which will have a significant impact on the ultimate physical properties of the material. in particular, abnormal grain growth in which certain grains grow very large in a matrix of finer grains will significantly alter the physical and mechanical properties of the obtained ceramic. in the sintered body, grain sizes are a product of the thermal processing parameters as well as the initial particle size, or possibly the sizes of aggregates or particle clusters which arise during the initial stages of processing. the ultimate microstructure ( and thus the physical properties ) of the final product will be limited by and subject to the form of the structural template or precursor which is created in the initial stages of chemical synthesis and physical forming. hence the importance of chemical powder and polymer processing as it pertains to the synthesis of industrial ceramics, glasses and glass - ceramics. there are numerous possible refinements of the sintering process. some of the most common involve pressing the green body to give the densification a head start and reduce the sintering time needed. sometimes organic binders such as polyvinyl alcohol are added to hold the green body together ; these burn out during the firing ( at 200 β 350 Β°c ). sometimes organic lubricants are added during pressing to increase densification. it is common to combine these, and add binders and lubricants to a powder, then press. ( the formulation of these organic chemical additives is an art in itself. this is particularly important in the manufacture of high performance ceramics such as those used by the billions for
or homogeneous distribution of the dispersed particle or fiber phase. consider first the processing of particulate composites. the particulate phase of greatest interest is tetragonal zirconia because of the toughening that can be achieved from the phase transformation from the metastable tetragonal to the monoclinic crystalline phase, aka transformation toughening. there is also substantial interest in dispersion of hard, non - oxide phases such as sic, tib, tic, boron, carbon and especially oxide matrices like alumina and mullite. there is also interest too incorporating other ceramic particulates, especially those of highly anisotropic thermal expansion. examples include al2o3, tio2, graphite, and boron nitride. in processing particulate composites, the issue is not only homogeneity of the size and spatial distribution of the dispersed and matrix phases, but also control of the matrix grain size. however, there is some built - in self - control due to inhibition of matrix grain growth by the dispersed phase. particulate composites, though generally offer increased resistance to damage, failure, or both, are still quite sensitive to inhomogeneities of composition as well as other processing defects such as pores. thus they need good processing to be effective. particulate composites have been made on a commercial basis by simply mixing powders of the two constituents. although this approach is inherently limited in the homogeneity that can be achieved, it is the most readily adaptable for existing ceramic production technology. however, other approaches are of interest. from the technological standpoint, a particularly desirable approach to fabricating particulate composites is to coat the matrix or its precursor onto fine particles of the dispersed phase with good control of the starting dispersed particle size and the resultant matrix coating thickness. one should in principle be able to achieve the ultimate in homogeneity of distribution and thereby optimize composite performance. this can also have other ramifications, such as allowing more useful composite performance to be achieved in a body having porosity, which might be desired for other factors, such as limiting thermal conductivity. there are also some opportunities to utilize melt processing for fabrication of ceramic, particulate, whisker and short - fiber, and continuous - fiber composites. both particulate and whisker composites are conceivable by solid - state precipitation after solidification of the melt. this can also be obtained in some cases by sintering,
an electron inside liquid helium forms a bubble of 17 \ aa in radius. in an external magnetic field, the two - level system of a spin 1 / 2 electron is ideal for the implementation of a qubit for quantum computing. the electron spin is well isolated from other thermal reservoirs so that the qubit should have very long coherence time. by confining a chain of single electron bubbles in a linear rf quadrupole trap, a multi - bit quantum register can be implemented. all spins in the register can be initialized to the ground state either by establishing thermal equilibrium at a temperature around 0. 1 k and at a magnetic field of 1 t or by sorting the bubbles to be loaded into the trap with magnetic separation. schemes are designed to address individual spins and to do two - qubit cnot operations between the neighboring spins. the final readout can be carried out through a measurement similar to the stern - gerlach experiment.
Question: When an ice cube melts, its state of matter changes from a
A) gas to a liquid
B) solid to a liquid
C) liquid to a solid
D) solid to a gas
|
B) solid to a liquid
|
Context:
becomes quite gentle. accordingly, in large basins, rivers in most cases begin as torrents with a variable flow, and end as gently flowing rivers with a comparatively regular discharge. the irregular flow of rivers throughout their course forms one of the main difficulties in devising works for mitigating inundations or for increasing the navigable capabilities of rivers. in tropical countries subject to periodical rains, the rivers are in flood during the rainy season and have hardly any flow during the rest of the year, while in temperate regions, where the rainfall is more evenly distributed throughout the year, evaporation causes the available rainfall to be much less in hot summer weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under
gravity induced condensation takes the form of momentum alignment in an ensemble of identical particles. use is made of a one - dimensional ising model to calculate the alignment per particle and the correlation length as a function of the temperature. these parameters indicate that momentum alignment is possible in the proximity of some astrophysical objects and in earth, or near earth laboratories. momenta oscillations behave as known spin oscillations and obey identical dispersion relations.
approximately corresponds to the slope of the country it traverses ; as rivers rise close to the highest part of their basins, generally in hilly regions, their fall is rapid near their source and gradually diminishes, with occasional irregularities, until, in traversing plains along the latter part of their course, their fall usually becomes quite gentle. accordingly, in large basins, rivers in most cases begin as torrents with a variable flow, and end as gently flowing rivers with a comparatively regular discharge. the irregular flow of rivers throughout their course forms one of the main difficulties in devising works for mitigating inundations or for increasing the navigable capabilities of rivers. in tropical countries subject to periodical rains, the rivers are in flood during the rainy season and have hardly any flow during the rest of the year, while in temperate regions, where the rainfall is more evenly distributed throughout the year, evaporation causes the available rainfall to be much less in hot summer weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with
the gas giant planets in the solar system have a retinue of icy moons, and we expect giant exoplanets to have similar satellite systems. if a jupiter - like planet were to migrate toward its parent star the icy moons orbiting it would evaporate, creating atmospheres and possible habitable surface oceans. here, we examine how long the surface ice and possible oceans would last before being hydrodynamically lost to space. the hydrodynamic loss rate from the moons is determined, in large part, by the stellar flux available for absorption, which increases as the giant planet and icy moons migrate closer to the star. at some planet - star distance the stellar flux incident on the icy moons becomes so great that they enter a runaway greenhouse state. this runaway greenhouse state rapidly transfers all available surface water to the atmosphere as vapor, where it is easily lost from the small moons. however, for icy moons of ganymede ' s size around a sun - like star we found that surface water ( either ice or liquid ) can persist indefinitely outside the runaway greenhouse orbital distance. in contrast, the surface water on smaller moons of europa ' s size will only persist on timescales greater than 1 gyr at distances ranging 1. 49 to 0. 74 au around a sun - like star for bond albedos of 0. 2 and 0. 8, where the lower albedo becomes relevant if ice melts. consequently, small moons can lose their icy shells, which would create a torus of h atoms around their host planet that might be detectable in future observations.
navigable capabilities of rivers. in tropical countries subject to periodical rains, the rivers are in flood during the rainy season and have hardly any flow during the rest of the year, while in temperate regions, where the rainfall is more evenly distributed throughout the year, evaporation causes the available rainfall to be much less in hot summer weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under ordinary conditions, most of the materials brought down from the high lands by torrential water courses are carried forward by the main river to the sea, or partially strewn over flat alluvial plains during floods ; the size of the materials forming the bed of the river or borne along by the stream is gradually reduced on proceeding sea
basic properties of black holes are explained in terms of trapping horizons. it is shown that matter and information will escape from an evaporating black hole. a general scenario is outlined whereby a black hole evaporates completely without singularity, event horizon or loss of energy or information.
interaction between tannin and bovine serum albumin ( bsa ) was examined by the fluorescent quenching. the process of elimination between bsa and tannin was the one of a stationary state, and the coupling coefficient was one. the working strength between the tannin and the beef serum was hydrophobic one.
to investigate the affinity of acetylated wood for organic liquids, yezo spruce wood specimens were acetylated with acetic anhydride, and their swelling in various liquids were compared to those of untreated specimens. the acetylated wood was rapidly and remarkably swollen in aprotic organic liquids such as benzene and toluene in which the untreated wood was swollen only slightly and / or very slowly. on the other hand, the swelling of wood in water, ethylene glycol and alcohols remained unchanged or decreased by the acetylation. consequently the maximum volume of wood swollen in organic liquids was always larger than that in water. the effect of acetylation on the maximum swollen volume of wood was greater in liquids having smaller solubility parameters. the easier penetration of aprotic organic liquids into the acetylated wood was considered to be due to the scission of hydrogen bonds among the amorphous wood constituents by the substitution of hydroxyl groups with hydrophobic acetyl groups.
weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under ordinary conditions, most of the materials brought down from the high lands by torrential water courses are carried forward by the main river to the sea, or partially strewn over flat alluvial plains during floods ; the size of the materials forming the bed of the river or borne along by the stream is gradually reduced on proceeding seawards, so that in the po river in italy, for instance, pebbles and gravel are found for about 140 miles below turin, sand along the next 100 miles, and silt and mud in the last 110 miles ( 176 km ). = = channelization = = the removal of obstructions, natural or artificial
superheated droplets are proven to be excelent detectors for neutrons and could be used as a neutron dosimeter. to detect accurately the volume of the vapour formed upon nucleation and hence to observe the nucleation quantitatively an air displacement system has been developed.
Question: In which way are evaporation and condensation similar?
A) Both cause decreases in air temperature.
B) Both cause increases in air temperature.
C) Both are caused by the warming of the atmosphere.
D) Both are caused by changes in heat energy.
|
D) Both are caused by changes in heat energy.
|
Context:
joints. = = = metal alloys = = = the alloys of iron ( steel, stainless steel, cast iron, tool steel, alloy steels ) make up the largest proportion of metals today both by quantity and commercial value. iron alloyed with various proportions of carbon gives low, mid and high carbon steels. an iron - carbon alloy is only considered steel if the carbon level is between 0. 01 % and 2. 00 % by weight. for steels, the hardness and tensile strength of the steel is related to the amount of carbon present, with increasing carbon levels also leading to lower ductility and toughness. heat treatment processes such as quenching and tempering can significantly change these properties, however. in contrast, certain metal alloys exhibit unique properties where their size and density remain unchanged across a range of temperatures. cast iron is defined as an iron β carbon alloy with more than 2. 00 %, but less than 6. 67 % carbon. stainless steel is defined as a regular steel alloy with greater than 10 % by weight alloying content of chromium. nickel and molybdenum are typically also added in stainless steels. other significant metallic alloys are those of aluminium, titanium, copper and magnesium. copper alloys have been known for a long time ( since the bronze age ), while the alloys of the other three metals have been relatively recently developed. due to the chemical reactivity of these metals, the electrolytic extraction processes required were only developed relatively recently. the alloys of aluminium, titanium and magnesium are also known and valued for their high strength to weight ratios and, in the case of magnesium, their ability to provide electromagnetic shielding. these materials are ideal for situations where high strength to weight ratios are more important than bulk cost, such as in the aerospace industry and certain automotive engineering applications. = = = semiconductors = = = a semiconductor is a material that has a resistivity between a conductor and insulator. modern day electronics run on semiconductors, and the industry had an estimated us $ 530 billion market in 2021. its electronic properties can be greatly altered through intentionally introducing impurities in a process referred to as doping. semiconductor materials are used to build diodes, transistors, light - emitting diodes ( leds ), and analog and digital electric circuits, among their many uses. semiconductor devices have replaced thermionic devices like vacuum tubes in most applications. semiconductor devices are manufactured both as single discrete devices and as integrated circuits ( ics ), which consist of a number β from a
casting, also called the lost wax process, die casting, centrifugal casting, both vertical and horizontal, and continuous castings. each of these forms has advantages for certain metals and applications considering factors like magnetism and corrosion. forging β a red - hot billet is hammered into shape. rolling β a billet is passed through successively narrower rollers to create a sheet. extrusion β a hot and malleable metal is forced under pressure through a die, which shapes it before it cools. machining β lathes, milling machines and drills cut the cold metal to shape. sintering β a powdered metal is heated in a non - oxidizing environment after being compressed into a die. fabrication β sheets of metal are cut with guillotines or gas cutters and bent and welded into structural shape. laser cladding β metallic powder is blown through a movable laser beam ( e. g. mounted on a nc 5 - axis machine ). the resulting melted metal reaches a substrate to form a melt pool. by moving the laser head, it is possible to stack the tracks and build up a three - dimensional piece. 3d printing β sintering or melting amorphous powder metal in a 3d space to make any object to shape. cold - working processes, in which the product ' s shape is altered by rolling, fabrication or other processes, while the product is cold, can increase the strength of the product by a process called work hardening. work hardening creates microscopic defects in the metal, which resist further changes of shape. = = = heat treatment = = = metals can be heat - treated to alter the properties of strength, ductility, toughness, hardness and resistance to corrosion. common heat treatment processes include annealing, precipitation strengthening, quenching, and tempering : annealing process softens the metal by heating it and then allowing it to cool very slowly, which gets rid of stresses in the metal and makes the grain structure large and soft - edged so that, when the metal is hit or stressed it dents or perhaps bends, rather than breaking ; it is also easier to sand, grind, or cut annealed metal. quenching is the process of cooling metal very quickly after heating, thus " freezing " the metal ' s molecules in the very hard martensite form, which makes the metal harder. tempering relieves stresses in the metal that were caused by the hardening process ; tempering makes the metal less hard while making it better able to sustain
the valuable metals into individual constituents. = = metal and its alloys = = much effort has been placed on understanding iron β carbon alloy system, which includes steels and cast irons. plain carbon steels ( those that contain essentially only carbon as an alloying element ) are used in low - cost, high - strength applications, where neither weight nor corrosion are a major concern. cast irons, including ductile iron, are also part of the iron - carbon system. iron - manganese - chromium alloys ( hadfield - type steels ) are also used in non - magnetic applications such as directional drilling. other engineering metals include aluminium, chromium, copper, magnesium, nickel, titanium, zinc, and silicon. these metals are most often used as alloys with the noted exception of silicon, which is not a metal. other forms include : stainless steel, particularly austenitic stainless steels, galvanized steel, nickel alloys, titanium alloys, or occasionally copper alloys are used, where resistance to corrosion is important. aluminium alloys and magnesium alloys are commonly used, when a lightweight strong part is required such as in automotive and aerospace applications. copper - nickel alloys ( such as monel ) are used in highly corrosive environments and for non - magnetic applications. nickel - based superalloys like inconel are used in high - temperature applications such as gas turbines, turbochargers, pressure vessels, and heat exchangers. for extremely high temperatures, single crystal alloys are used to minimize creep. in modern electronics, high purity single crystal silicon is essential for metal - oxide - silicon transistors ( mos ) and integrated circuits. = = production = = in production engineering, metallurgy is concerned with the production of metallic components for use in consumer or engineering products. this involves production of alloys, shaping, heat treatment and surface treatment of product. the task of the metallurgist is to achieve balance between material properties, such as cost, weight, strength, toughness, hardness, corrosion, fatigue resistance and performance in temperature extremes. to achieve this goal, the operating environment must be carefully considered. determining the hardness of the metal using the rockwell, vickers, and brinell hardness scales is a commonly used practice that helps better understand the metal ' s elasticity and plasticity for different applications and production processes. in a saltwater environment, most ferrous metals and some non - ferrous alloys corrode quickly. metals exposed to cold or cryogenic conditions may undergo a ductile to brittle
applications, where neither weight nor corrosion are a major concern. cast irons, including ductile iron, are also part of the iron - carbon system. iron - manganese - chromium alloys ( hadfield - type steels ) are also used in non - magnetic applications such as directional drilling. other engineering metals include aluminium, chromium, copper, magnesium, nickel, titanium, zinc, and silicon. these metals are most often used as alloys with the noted exception of silicon, which is not a metal. other forms include : stainless steel, particularly austenitic stainless steels, galvanized steel, nickel alloys, titanium alloys, or occasionally copper alloys are used, where resistance to corrosion is important. aluminium alloys and magnesium alloys are commonly used, when a lightweight strong part is required such as in automotive and aerospace applications. copper - nickel alloys ( such as monel ) are used in highly corrosive environments and for non - magnetic applications. nickel - based superalloys like inconel are used in high - temperature applications such as gas turbines, turbochargers, pressure vessels, and heat exchangers. for extremely high temperatures, single crystal alloys are used to minimize creep. in modern electronics, high purity single crystal silicon is essential for metal - oxide - silicon transistors ( mos ) and integrated circuits. = = production = = in production engineering, metallurgy is concerned with the production of metallic components for use in consumer or engineering products. this involves production of alloys, shaping, heat treatment and surface treatment of product. the task of the metallurgist is to achieve balance between material properties, such as cost, weight, strength, toughness, hardness, corrosion, fatigue resistance and performance in temperature extremes. to achieve this goal, the operating environment must be carefully considered. determining the hardness of the metal using the rockwell, vickers, and brinell hardness scales is a commonly used practice that helps better understand the metal ' s elasticity and plasticity for different applications and production processes. in a saltwater environment, most ferrous metals and some non - ferrous alloys corrode quickly. metals exposed to cold or cryogenic conditions may undergo a ductile to brittle transition and lose their toughness, becoming more brittle and prone to cracking. metals under continual cyclic loading can suffer from metal fatigue. metals under constant stress at elevated temperatures can creep. = = = metalworking processes = = = casting β molten metal is poured into a shaped mold. variants of casting include sand casting, investment
temperature changes up to 1000 Β°c. = = processing steps = = the traditional ceramic process generally follows this sequence : milling β batching β mixing β forming β drying β firing β assembly. milling is the process by which materials are reduced from a large size to a smaller size. milling may involve breaking up cemented material ( in which case individual particles retain their shape ) or pulverization ( which involves grinding the particles themselves to a smaller size ). milling is generally done by mechanical means, including attrition ( which is particle - to - particle collision that results in agglomerate break up or particle shearing ), compression ( which applies a forces that results in fracturing ), and impact ( which employs a milling medium or the particles themselves to cause fracturing ). attrition milling equipment includes the wet scrubber ( also called the planetary mill or wet attrition mill ), which has paddles in water creating vortexes in which the material collides and break up. compression mills include the jaw crusher, roller crusher and cone crusher. impact mills include the ball mill, which has media that tumble and fracture the material, or the resonantacoustic mixer. shaft impactors cause particle - to particle attrition and compression. batching is the process of weighing the oxides according to recipes, and preparing them for mixing and drying. mixing occurs after batching and is performed with various machines, such as dry mixing ribbon mixers ( a type of cement mixer ), resonantacoustic mixers, mueller mixers, and pug mills. wet mixing generally involves the same equipment. forming is making the mixed material into shapes, ranging from toilet bowls to spark plug insulators. forming can involve : ( 1 ) extrusion, such as extruding " slugs " to make bricks, ( 2 ) pressing to make shaped parts, ( 3 ) slip casting, as in making toilet bowls, wash basins and ornamentals like ceramic statues. forming produces a " green " part, ready for drying. green parts are soft, pliable, and over time will lose shape. handling the green product will change its shape. for example, a green brick can be " squeezed ", and after squeezing it will stay that way. drying is removing the water or binder from the formed material. spray drying is widely used to prepare powder for pressing operations. other dryers are tunnel dryers and periodic dryers. controlled heat is applied in this two - stage process. first,
10 kgy most food, which is ( with regard to warming ) physically equivalent to water, would warm by only about 2. 5 Β°c ( 4. 5 Β°f ). the specialty of processing food by ionizing radiation is the fact, that the energy density per atomic transition is very high, it can cleave molecules and induce ionization ( hence the name ) which cannot be achieved by mere heating. this is the reason for new beneficial effects, however at the same time, for new concerns. the treatment of solid food by ionizing radiation can provide an effect similar to heat pasteurization of liquids, such as milk. however, the use of the term, cold pasteurization, to describe irradiated foods is controversial, because pasteurization and irradiation are fundamentally different processes, although the intended end results can in some cases be similar. detractors of food irradiation have concerns about the health hazards of induced radioactivity. a report for the industry advocacy group american council on science and health entitled " irradiated foods " states : " the types of radiation sources approved for the treatment of foods have specific energy levels well below that which would cause any element in food to become radioactive. food undergoing irradiation does not become any more radioactive than luggage passing through an airport x - ray scanner or teeth that have been x - rayed. " food irradiation is currently permitted by over 40 countries and volumes are estimated to exceed 500, 000 metric tons ( 490, 000 long tons ; 550, 000 short tons ) annually worldwide. food irradiation is essentially a non - nuclear technology ; it relies on the use of ionizing radiation which may be generated by accelerators for electrons and conversion into bremsstrahlung, but which may use also gamma - rays from nuclear decay. there is a worldwide industry for processing by ionizing radiation, the majority by number and by processing power using accelerators. food irradiation is only a niche application compared to medical supplies, plastic materials, raw materials, gemstones, cables and wires, etc. = = accidents = = nuclear accidents, because of the powerful forces involved, are often very dangerous. historically, the first incidents involved fatal radiation exposure. marie curie died from aplastic anemia which resulted from her high levels of exposure. two scientists, an american and canadian respectively, harry daghlian and louis slotin, died after mishandling the same plutonium mass. unlike conventional weapons, the intense light, heat, and explosive force is
we predict the upper bound on the dissociation temperatures of different quarkonium states.
intense research in the materials science community due to the unique properties that they exhibit. nanostructure deals with objects and structures that are in the 1 β 100 nm range. in many materials, atoms or molecules agglomerate to form objects at the nanoscale. this causes many interesting electrical, magnetic, optical, and mechanical properties. in describing nanostructures, it is necessary to differentiate between the number of dimensions on the nanoscale. nanotextured surfaces have one dimension on the nanoscale, i. e., only the thickness of the surface of an object is between 0. 1 and 100 nm. nanotubes have two dimensions on the nanoscale, i. e., the diameter of the tube is between 0. 1 and 100 nm ; its length could be much greater. finally, spherical nanoparticles have three dimensions on the nanoscale, i. e., the particle is between 0. 1 and 100 nm in each spatial dimension. the terms nanoparticles and ultrafine particles ( ufp ) often are used synonymously although ufp can reach into the micrometre range. the term ' nanostructure ' is often used, when referring to magnetic technology. nanoscale structure in biology is often called ultrastructure. = = = = microstructure = = = = microstructure is defined as the structure of a prepared surface or thin foil of material as revealed by a microscope above 25Γ magnification. it deals with objects from 100 nm to a few cm. the microstructure of a material ( which can be broadly classified into metallic, polymeric, ceramic and composite ) can strongly influence physical properties such as strength, toughness, ductility, hardness, corrosion resistance, high / low temperature behavior, wear resistance, and so on. most of the traditional materials ( such as metals and ceramics ) are microstructured. the manufacture of a perfect crystal of a material is physically impossible. for example, any crystalline material will contain defects such as precipitates, grain boundaries ( hall β petch relationship ), vacancies, interstitial atoms or substitutional atoms. the microstructure of materials reveals these larger defects and advances in simulation have allowed an increased understanding of how defects can be used to enhance material properties. = = = = macrostructure = = = = macrostructure is the appearance of a material in the scale millimeters to meters, it is the structure of
changed their sound to a more straightforward rock style. the penultimate track " deep inside " is another of the lighter songs on the album, drawing heavily from funk music. however, it also has a short heavy metal section beginning roughly 3 minutes into the song. the lyrics reference being high on drugs at 3 a. m. in the morning, and as the song goes on the time progresses to 4 a. m. and eventually 5 a. m. " deep inside " stopped being performed live when alex katunich left incubus at the beginning of april 2003, with the last known performance coming during december 2001. the closing track " calgone " lyrically revolves around an alien abduction. it ends with a sound clip of the band arguing with dj lyfe ( who is referred to by his real name of gavin ), for supposedly deleting a track they had been working on in the studio. on physical versions of the album, " calgone " is followed by a hidden track called " segue 1 ", which plays after 30 seconds of silence. the hidden track is also known as " jose loves kate moss, part 1 ", and has been treated as a separate track on streaming sites such as spotify. it begins with a sound recording at a morgue, and the pathologist is describing injuries a patient had sustained during a car crash on may 12, 1997. the hidden track goes on to feature several different sound samples and funk / electronic musical pieces. it samples sounds from the 1985 sorcerer pinball machine and the song " show me your titz ", from hoobastank ' s 1997 demo muffins. it also has an electronic piece in the style of mozart ' s toy symphony, a sound sample of an unknown woman saying " is this the shit, or what? ", and a short skit parodying the karate kid. = = = title and artwork = = = in 1997, einziger claimed that the title reflected the experimental nature of the album, and the creative freedom the band were given. he was quoted as saying, " our album is called s. c. i. e. n. c. e. because we were able to experiment. we were able to take our time and get everything to sound the way we wanted it to β weird science and energetic funk. " it has also been mentioned by various band members that the acronym s. c. i. e. n. c. e. stands for sailing catamarans is every nautical captain
billet is passed through successively narrower rollers to create a sheet. extrusion β a hot and malleable metal is forced under pressure through a die, which shapes it before it cools. machining β lathes, milling machines and drills cut the cold metal to shape. sintering β a powdered metal is heated in a non - oxidizing environment after being compressed into a die. fabrication β sheets of metal are cut with guillotines or gas cutters and bent and welded into structural shape. laser cladding β metallic powder is blown through a movable laser beam ( e. g. mounted on a nc 5 - axis machine ). the resulting melted metal reaches a substrate to form a melt pool. by moving the laser head, it is possible to stack the tracks and build up a three - dimensional piece. 3d printing β sintering or melting amorphous powder metal in a 3d space to make any object to shape. cold - working processes, in which the product ' s shape is altered by rolling, fabrication or other processes, while the product is cold, can increase the strength of the product by a process called work hardening. work hardening creates microscopic defects in the metal, which resist further changes of shape. = = = heat treatment = = = metals can be heat - treated to alter the properties of strength, ductility, toughness, hardness and resistance to corrosion. common heat treatment processes include annealing, precipitation strengthening, quenching, and tempering : annealing process softens the metal by heating it and then allowing it to cool very slowly, which gets rid of stresses in the metal and makes the grain structure large and soft - edged so that, when the metal is hit or stressed it dents or perhaps bends, rather than breaking ; it is also easier to sand, grind, or cut annealed metal. quenching is the process of cooling metal very quickly after heating, thus " freezing " the metal ' s molecules in the very hard martensite form, which makes the metal harder. tempering relieves stresses in the metal that were caused by the hardening process ; tempering makes the metal less hard while making it better able to sustain impacts without breaking. often, mechanical and thermal treatments are combined in what are known as thermo - mechanical treatments for better properties and more efficient processing of materials. these processes are common to high - alloy special steels, superalloys and titanium alloys. = = = plating = = = electroplating is
Question: The surface of a heated metal object measures 120Β°C. It is dropped into a bucket filled with water measuring 10Β°C. Which of the following is most likely to occur?
A) Both the water and metal will adjust to the same temperature below 10Β°C.
B) The water and metal will adjust to different temperatures above 120Β°C.
C) The water will remain the same temperature, but the metal will cool to 10Β°C.
D) The water and metal will adjust to the same temperature between 10Β°C and 120Β°C.
|
D) The water and metal will adjust to the same temperature between 10Β°C and 120Β°C.
|
Context:
missiles, ships, vehicles, and also to map weather patterns and terrain. a radar set consists of a transmitter and receiver. the transmitter emits a narrow beam of radio waves which is swept around the surrounding space. when the beam strikes a target object, radio waves are reflected back to the receiver. the direction of the beam reveals the object ' s location. since radio waves travel at a constant speed close to the speed of light, by measuring the brief time delay between the outgoing pulse and the received " echo ", the range to the target can be calculated. the targets are often displayed graphically on a map display called a radar screen. doppler radar can measure a moving object ' s velocity, by measuring the change in frequency of the return radio waves due to the doppler effect. radar sets mainly use high frequencies in the microwave bands, because these frequencies create strong reflections from objects the size of vehicles and can be focused into narrow beams with compact antennas. parabolic ( dish ) antennas are widely used. in most radars the transmitting antenna also serves as the receiving antenna ; this is called a monostatic radar. a radar which uses separate transmitting and receiving antennas is called a bistatic radar. airport surveillance radar β in aviation, radar is the main tool of air traffic control. a rotating dish antenna sweeps a vertical fan - shaped beam of microwaves around the airspace and the radar set shows the location of aircraft as " blips " of light on a display called a radar screen. airport radar operates at 2. 7 β 2. 9 ghz in the microwave s band. in large airports the radar image is displayed on multiple screens in an operations room called the tracon ( terminal radar approach control ), where air traffic controllers direct the aircraft by radio to maintain safe aircraft separation. secondary surveillance radar β aircraft carry radar transponders, transceivers which when triggered by the incoming radar signal transmit a return microwave signal. this causes the aircraft to show up more strongly on the radar screen. the radar which triggers the transponder and receives the return beam, usually mounted on top of the primary radar dish, is called the secondary surveillance radar. since radar cannot measure an aircraft ' s altitude with any accuracy, the transponder also transmits back the aircraft ' s altitude measured by its altimeter, and an id number identifying the aircraft, which is displayed on the radar screen. electronic countermeasures ( ecm ) β military defensive electronic systems designed to degrade enemy radar effectiveness, or deceive it
beam reveals the object ' s location. since radio waves travel at a constant speed close to the speed of light, by measuring the brief time delay between the outgoing pulse and the received " echo ", the range to the target can be calculated. the targets are often displayed graphically on a map display called a radar screen. doppler radar can measure a moving object ' s velocity, by measuring the change in frequency of the return radio waves due to the doppler effect. radar sets mainly use high frequencies in the microwave bands, because these frequencies create strong reflections from objects the size of vehicles and can be focused into narrow beams with compact antennas. parabolic ( dish ) antennas are widely used. in most radars the transmitting antenna also serves as the receiving antenna ; this is called a monostatic radar. a radar which uses separate transmitting and receiving antennas is called a bistatic radar. airport surveillance radar β in aviation, radar is the main tool of air traffic control. a rotating dish antenna sweeps a vertical fan - shaped beam of microwaves around the airspace and the radar set shows the location of aircraft as " blips " of light on a display called a radar screen. airport radar operates at 2. 7 β 2. 9 ghz in the microwave s band. in large airports the radar image is displayed on multiple screens in an operations room called the tracon ( terminal radar approach control ), where air traffic controllers direct the aircraft by radio to maintain safe aircraft separation. secondary surveillance radar β aircraft carry radar transponders, transceivers which when triggered by the incoming radar signal transmit a return microwave signal. this causes the aircraft to show up more strongly on the radar screen. the radar which triggers the transponder and receives the return beam, usually mounted on top of the primary radar dish, is called the secondary surveillance radar. since radar cannot measure an aircraft ' s altitude with any accuracy, the transponder also transmits back the aircraft ' s altitude measured by its altimeter, and an id number identifying the aircraft, which is displayed on the radar screen. electronic countermeasures ( ecm ) β military defensive electronic systems designed to degrade enemy radar effectiveness, or deceive it with false information, to prevent enemies from locating local forces. it often consists of powerful microwave transmitters that can mimic enemy radar signals to create false target indications on the enemy radar screens. marine radar β an s or x band radar on ships used to detect nearby ships and obstructions like bridges. a rotating antenna sweeps a vertical
a live / sound reinforcement engineer hears source material and tries to correlate that sonic experience with system performance. wireless microphone engineer, or a2. this position is responsible for wireless microphones during a theatre production, a sports event or a corporate event. foldback or monitor engineer β a person running foldback sound during a live event. the term foldback comes from the old practice of folding back audio signals from the front of house ( foh ) mixing console to the stage so musicians can hear themselves while performing. monitor engineers usually have a separate audio system from the foh engineer and manipulate audio signals independently from what the audience hears so they can satisfy the requirements of each performer on stage. in - ear systems, digital and analog mixing consoles, and a variety of speaker enclosures are typically used by monitor engineers. in addition, most monitor engineers must be familiar with wireless or rf ( radio - frequency ) equipment and often must communicate personally with the artist ( s ) during each performance. systems engineer β responsible for the design setup of modern pa systems, which are often very complex. a systems engineer is usually also referred to as a crew chief on tour and is responsible for the performance and day - to - day job requirements of the audio crew as a whole along with the foh audio system. this is a sound - only position concerned with implementation, not to be confused with the interdisciplinary field of system engineering, which typically requires a college degree. re - recording mixer β a person in post - production who mixes audio tracks for feature films or television programs. = = equipment = = an audio engineer is proficient with different types of recording media, such as analog tape, digital multi - track recorders and workstations, plug - ins and computer knowledge. with the advent of the digital age, it is increasingly important for the audio engineer to understand software and hardware integration, from synchronization to analog to digital transfers. in their daily work, audio engineers use many tools, including : tape machines analog - to - digital converters digital - to - analog converters digital audio workstations ( daws ) audio plug - ins dynamic range compressors audio data compressors equalization ( audio ) music sequencers signal processors headphones microphones preamplifiers mixing consoles amplifiers loudspeakers = = notable audio engineers = = = = = recording = = = = = = mastering = = = = = = live sound = = = = = see also = = = = references = = = = external links = = audio engineering society audio engineering
due to its location and climate, antarctica offers unique conditions for long - period observations across a broad wavelength regime, where important diagnostic lines for molecules and ions can be found, that are essential to understand the chemical properties of the interstellar medium. in addition to the natural benefits of the site, new technologies, resulting from astrophotonics, may allow miniaturised instruments, that are easier to winterise and advanced filters to further reduce the background in the infrared.
also launched missions to mercury in 2004, with the messenger probe demonstrating as the first use of a solar sail. nasa also launched probes to the outer solar system starting in the 1960s. pioneer 10 was the first probe to the outer planets, flying by jupiter, while pioneer 11 provided the first close up view of the planet. both probes became the first objects to leave the solar system. the voyager program launched in 1977, conducting flybys of jupiter and saturn, neptune, and uranus on a trajectory to leave the solar system. the galileo spacecraft, deployed from the space shuttle flight sts - 34, was the first spacecraft to orbit jupiter, discovering evidence of subsurface oceans on the europa and observed that the moon may hold ice or liquid water. a joint nasa - european space agency - italian space agency mission, cassini β huygens, was sent to saturn ' s moon titan, which, along with mars and europa, are the only celestial bodies in the solar system suspected of being capable of harboring life. cassini discovered three new moons of saturn and the huygens probe entered titan ' s atmosphere. the mission discovered evidence of liquid hydrocarbon lakes on titan and subsurface water oceans on the moon of enceladus, which could harbor life. finally launched in 2006, the new horizons mission was the first spacecraft to visit pluto and the kuiper belt. beyond interplanetary probes, nasa has launched many space telescopes. launched in the 1960s, the orbiting astronomical observatory were nasa ' s first orbital telescopes, providing ultraviolet, gamma - ray, x - ray, and infrared observations. nasa launched the orbiting geophysical observatory in the 1960s and 1970s to look down at earth and observe its interactions with the sun. the uhuru satellite was the first dedicated x - ray telescope, mapping 85 % of the sky and discovering a large number of black holes. launched in the 1990s and early 2000s, the great observatories program are among nasa ' s most powerful telescopes. the hubble space telescope was launched in 1990 on sts - 31 from the discovery and could view galaxies 15 billion light years away. a major defect in the telescope ' s mirror could have crippled the program, had nasa not used computer enhancement to compensate for the imperfection and launched five space shuttle servicing flights to replace the damaged components. the compton gamma ray observatory was launched from the atlantis on sts - 37 in 1991, discovering a possible source of antimatter at the center of the milky way and observing that the majority of gamma - ray bursts
compact antennas. parabolic ( dish ) antennas are widely used. in most radars the transmitting antenna also serves as the receiving antenna ; this is called a monostatic radar. a radar which uses separate transmitting and receiving antennas is called a bistatic radar. airport surveillance radar β in aviation, radar is the main tool of air traffic control. a rotating dish antenna sweeps a vertical fan - shaped beam of microwaves around the airspace and the radar set shows the location of aircraft as " blips " of light on a display called a radar screen. airport radar operates at 2. 7 β 2. 9 ghz in the microwave s band. in large airports the radar image is displayed on multiple screens in an operations room called the tracon ( terminal radar approach control ), where air traffic controllers direct the aircraft by radio to maintain safe aircraft separation. secondary surveillance radar β aircraft carry radar transponders, transceivers which when triggered by the incoming radar signal transmit a return microwave signal. this causes the aircraft to show up more strongly on the radar screen. the radar which triggers the transponder and receives the return beam, usually mounted on top of the primary radar dish, is called the secondary surveillance radar. since radar cannot measure an aircraft ' s altitude with any accuracy, the transponder also transmits back the aircraft ' s altitude measured by its altimeter, and an id number identifying the aircraft, which is displayed on the radar screen. electronic countermeasures ( ecm ) β military defensive electronic systems designed to degrade enemy radar effectiveness, or deceive it with false information, to prevent enemies from locating local forces. it often consists of powerful microwave transmitters that can mimic enemy radar signals to create false target indications on the enemy radar screens. marine radar β an s or x band radar on ships used to detect nearby ships and obstructions like bridges. a rotating antenna sweeps a vertical fan - shaped beam of microwaves around the water surface surrounding the craft out to the horizon. weather radar β a doppler radar which maps weather precipitation intensities and wind speeds with the echoes returned from raindrops and their radial velocity by their doppler shift. phased - array radar β a radar set that uses a phased array, a computer - controlled antenna that can steer the radar beam quickly to point in different directions without moving the antenna. phased - array radars were developed by the military to track fast - moving missiles and aircraft. they are widely used in military equipment and are now spreading to civilian applications. synthetic aperture
and measuring radiation levels. the surveyor program conducted uncrewed lunar landings and takeoffs, as well as taking surface and regolith observations. despite the setback caused by the apollo 1 fire, which killed three astronauts, the program proceeded. apollo 8 was the first crewed spacecraft to leave low earth orbit and the first human spaceflight to reach the moon. the crew orbited the moon ten times on december 24 and 25, 1968, and then traveled safely back to earth. the three apollo 8 astronauts β frank borman, james lovell, and william anders β were the first humans to see the earth as a globe in space, the first to witness an earthrise, and the first to see and manually photograph the far side of the moon. the first lunar landing was conducted by apollo 11. commanded by neil armstrong with astronauts buzz aldrin and michael collins, apollo 11 was one of the most significant missions in nasa ' s history, marking the end of the space race when the soviet union gave up its lunar ambitions. as the first human to step on the surface of the moon, neil armstrong uttered the now famous words : that ' s one small step for man, one giant leap for mankind. nasa would conduct six total lunar landings as part of the apollo program, with apollo 17 concluding the program in 1972. = = = = end of apollo = = = = wernher von braun had advocated for nasa to develop a space station since the agency was created. in 1973, following the end of the apollo lunar missions, nasa launched its first space station, skylab, on the final launch of the saturn v. skylab reused a significant amount of apollo and saturn hardware, with a repurposed saturn v third stage serving as the primary module for the space station. damage to skylab during its launch required spacewalks to be performed by the first crew to make it habitable and operational. skylab hosted nine missions and was decommissioned in 1974 and deorbited in 1979, two years prior to the first launch of the space shuttle and any possibility of boosting its orbit. in 1975, the apollo β soyuz mission was the first ever international spaceflight and a major diplomatic accomplishment between the cold war rivals, which also marked the last flight of the apollo capsule. flown in 1975, a us apollo spacecraft docked with a soviet soyuz capsule. = = = interplanetary exploration and space science = = = during the 1960s, nasa started its space science and interplanetary probe program. the mariner program was its flagship
the union of space telescopes and interstellar spaceships guarantees that if extraterrestrial civilizations were common, someone would have come here long ago.
. doppler radar can measure a moving object ' s velocity, by measuring the change in frequency of the return radio waves due to the doppler effect. radar sets mainly use high frequencies in the microwave bands, because these frequencies create strong reflections from objects the size of vehicles and can be focused into narrow beams with compact antennas. parabolic ( dish ) antennas are widely used. in most radars the transmitting antenna also serves as the receiving antenna ; this is called a monostatic radar. a radar which uses separate transmitting and receiving antennas is called a bistatic radar. airport surveillance radar β in aviation, radar is the main tool of air traffic control. a rotating dish antenna sweeps a vertical fan - shaped beam of microwaves around the airspace and the radar set shows the location of aircraft as " blips " of light on a display called a radar screen. airport radar operates at 2. 7 β 2. 9 ghz in the microwave s band. in large airports the radar image is displayed on multiple screens in an operations room called the tracon ( terminal radar approach control ), where air traffic controllers direct the aircraft by radio to maintain safe aircraft separation. secondary surveillance radar β aircraft carry radar transponders, transceivers which when triggered by the incoming radar signal transmit a return microwave signal. this causes the aircraft to show up more strongly on the radar screen. the radar which triggers the transponder and receives the return beam, usually mounted on top of the primary radar dish, is called the secondary surveillance radar. since radar cannot measure an aircraft ' s altitude with any accuracy, the transponder also transmits back the aircraft ' s altitude measured by its altimeter, and an id number identifying the aircraft, which is displayed on the radar screen. electronic countermeasures ( ecm ) β military defensive electronic systems designed to degrade enemy radar effectiveness, or deceive it with false information, to prevent enemies from locating local forces. it often consists of powerful microwave transmitters that can mimic enemy radar signals to create false target indications on the enemy radar screens. marine radar β an s or x band radar on ships used to detect nearby ships and obstructions like bridges. a rotating antenna sweeps a vertical fan - shaped beam of microwaves around the water surface surrounding the craft out to the horizon. weather radar β a doppler radar which maps weather precipitation intensities and wind speeds with the echoes returned from raindrops and their radial velocity by their doppler shift. phased - array radar β a radar set
one phenomenological explanation of superluminal propagation of neutrinos, which may have been observed by opera and minos, is that neutrinos travel faster inside of matter than in vacuum. if so neutrinos exhibit refraction inside matter and should exhibit other manifestations of refraction, such as deflection and reflection. such refraction would be easily detectable through the momentum imparted to appropriately shaped refractive material inserted into the neutrino beam. for numi this could be as large as ~ 10g cm / s. if these effect were found, they would provide new ways of manipulating and detecting neutrinos. reasons why this scenario seems implausible are given, however it is still worthwhile to conduct simple searches for differential refraction of neutrinos.
Question: In order to record the passing of a distant comet, which instruments would be the most useful?
A) camera, tripod, and telescope
B) telescope, binoculars, and camera
C) camera with flash attachment, and tripod
D) telescope with camera attachment and camera
|
D) telescope with camera attachment and camera
|
Context:
10 kgy most food, which is ( with regard to warming ) physically equivalent to water, would warm by only about 2. 5 Β°c ( 4. 5 Β°f ). the specialty of processing food by ionizing radiation is the fact, that the energy density per atomic transition is very high, it can cleave molecules and induce ionization ( hence the name ) which cannot be achieved by mere heating. this is the reason for new beneficial effects, however at the same time, for new concerns. the treatment of solid food by ionizing radiation can provide an effect similar to heat pasteurization of liquids, such as milk. however, the use of the term, cold pasteurization, to describe irradiated foods is controversial, because pasteurization and irradiation are fundamentally different processes, although the intended end results can in some cases be similar. detractors of food irradiation have concerns about the health hazards of induced radioactivity. a report for the industry advocacy group american council on science and health entitled " irradiated foods " states : " the types of radiation sources approved for the treatment of foods have specific energy levels well below that which would cause any element in food to become radioactive. food undergoing irradiation does not become any more radioactive than luggage passing through an airport x - ray scanner or teeth that have been x - rayed. " food irradiation is currently permitted by over 40 countries and volumes are estimated to exceed 500, 000 metric tons ( 490, 000 long tons ; 550, 000 short tons ) annually worldwide. food irradiation is essentially a non - nuclear technology ; it relies on the use of ionizing radiation which may be generated by accelerators for electrons and conversion into bremsstrahlung, but which may use also gamma - rays from nuclear decay. there is a worldwide industry for processing by ionizing radiation, the majority by number and by processing power using accelerators. food irradiation is only a niche application compared to medical supplies, plastic materials, raw materials, gemstones, cables and wires, etc. = = accidents = = nuclear accidents, because of the powerful forces involved, are often very dangerous. historically, the first incidents involved fatal radiation exposure. marie curie died from aplastic anemia which resulted from her high levels of exposure. two scientists, an american and canadian respectively, harry daghlian and louis slotin, died after mishandling the same plutonium mass. unlike conventional weapons, the intense light, heat, and explosive force is
or solid as is the case with water ( h2o ) ; a liquid at room temperature because its molecules are bound by hydrogen bonds. whereas hydrogen sulfide ( h2s ) is a gas at room temperature and standard pressure, as its molecules are bound by weaker dipole β dipole interactions. the transfer of energy from one chemical substance to another depends on the size of energy quanta emitted from one substance. however, heat energy is often transferred more easily from almost any substance to another because the phonons responsible for vibrational and rotational energy levels in a substance have much less energy than photons invoked for the electronic energy transfer. thus, because vibrational and rotational energy levels are more closely spaced than electronic energy levels, heat is more easily transferred between substances relative to light or other forms of electronic energy. for example, ultraviolet electromagnetic radiation is not transferred with as much efficacy from one substance to another as thermal or electrical energy. the existence of characteristic energy levels for different chemical substances is useful for their identification by the analysis of spectral lines. different kinds of spectra are often used in chemical spectroscopy, e. g. ir, microwave, nmr, esr, etc. spectroscopy is also used to identify the composition of remote objects β like stars and distant galaxies β by analyzing their radiation spectra. the term chemical energy is often used to indicate the potential of a chemical substance to undergo a transformation through a chemical reaction or to transform other chemical substances. = = = reaction = = = when a chemical substance is transformed as a result of its interaction with another substance or with energy, a chemical reaction is said to have occurred. a chemical reaction is therefore a concept related to the " reaction " of a substance when it comes in close contact with another, whether as a mixture or a solution ; exposure to some form of energy, or both. it results in some energy exchange between the constituents of the reaction as well as with the system environment, which may be designed vessels β often laboratory glassware. chemical reactions can result in the formation or dissociation of molecules, that is, molecules breaking apart to form two or more molecules or rearrangement of atoms within or across molecules. chemical reactions usually involve the making or breaking of chemical bonds. oxidation, reduction, dissociation, acid β base neutralization and molecular rearrangement are some examples of common chemical reactions. a chemical reaction can be symbolically depicted through a chemical equation. while in a non - nuclear chemical reaction the number and kind of atoms on both sides of the equation are equal, for
a comparison of the sensitivities of methods which allow us to determine the coordinates of a moving hot body is made.
factor e β e / k t { \ displaystyle e ^ { - e / kt } } β that is the probability of a molecule to have energy greater than or equal to e at the given temperature t. this exponential dependence of a reaction rate on temperature is known as the arrhenius equation. the activation energy necessary for a chemical reaction to occur can be in the form of heat, light, electricity or mechanical force in the form of ultrasound. a related concept free energy, which also incorporates entropy considerations, is a very useful means for predicting the feasibility of a reaction and determining the state of equilibrium of a chemical reaction, in chemical thermodynamics. a reaction is feasible only if the total change in the gibbs free energy is negative, Ξ΄ g β€ 0 { \ displaystyle \ delta g \ leq 0 \, } ; if it is equal to zero the chemical reaction is said to be at equilibrium. there exist only limited possible states of energy for electrons, atoms and molecules. these are determined by the rules of quantum mechanics, which require quantization of energy of a bound system. the atoms / molecules in a higher energy state are said to be excited. the molecules / atoms of substance in an excited energy state are often much more reactive ; that is, more amenable to chemical reactions. the phase of a substance is invariably determined by its energy and the energy of its surroundings. when the intermolecular forces of a substance are such that the energy of the surroundings is not sufficient to overcome them, it occurs in a more ordered phase like liquid or solid as is the case with water ( h2o ) ; a liquid at room temperature because its molecules are bound by hydrogen bonds. whereas hydrogen sulfide ( h2s ) is a gas at room temperature and standard pressure, as its molecules are bound by weaker dipole β dipole interactions. the transfer of energy from one chemical substance to another depends on the size of energy quanta emitted from one substance. however, heat energy is often transferred more easily from almost any substance to another because the phonons responsible for vibrational and rotational energy levels in a substance have much less energy than photons invoked for the electronic energy transfer. thus, because vibrational and rotational energy levels are more closely spaced than electronic energy levels, heat is more easily transferred between substances relative to light or other forms of electronic energy. for example, ultraviolet electromagnetic radiation is not transferred with as much efficacy from one substance to another as thermal or electrical energy. the existence of characteristic
; that is, more amenable to chemical reactions. the phase of a substance is invariably determined by its energy and the energy of its surroundings. when the intermolecular forces of a substance are such that the energy of the surroundings is not sufficient to overcome them, it occurs in a more ordered phase like liquid or solid as is the case with water ( h2o ) ; a liquid at room temperature because its molecules are bound by hydrogen bonds. whereas hydrogen sulfide ( h2s ) is a gas at room temperature and standard pressure, as its molecules are bound by weaker dipole β dipole interactions. the transfer of energy from one chemical substance to another depends on the size of energy quanta emitted from one substance. however, heat energy is often transferred more easily from almost any substance to another because the phonons responsible for vibrational and rotational energy levels in a substance have much less energy than photons invoked for the electronic energy transfer. thus, because vibrational and rotational energy levels are more closely spaced than electronic energy levels, heat is more easily transferred between substances relative to light or other forms of electronic energy. for example, ultraviolet electromagnetic radiation is not transferred with as much efficacy from one substance to another as thermal or electrical energy. the existence of characteristic energy levels for different chemical substances is useful for their identification by the analysis of spectral lines. different kinds of spectra are often used in chemical spectroscopy, e. g. ir, microwave, nmr, esr, etc. spectroscopy is also used to identify the composition of remote objects β like stars and distant galaxies β by analyzing their radiation spectra. the term chemical energy is often used to indicate the potential of a chemical substance to undergo a transformation through a chemical reaction or to transform other chemical substances. = = = reaction = = = when a chemical substance is transformed as a result of its interaction with another substance or with energy, a chemical reaction is said to have occurred. a chemical reaction is therefore a concept related to the " reaction " of a substance when it comes in close contact with another, whether as a mixture or a solution ; exposure to some form of energy, or both. it results in some energy exchange between the constituents of the reaction as well as with the system environment, which may be designed vessels β often laboratory glassware. chemical reactions can result in the formation or dissociation of molecules, that is, molecules breaking apart to form two or more molecules or rearrangement of atoms within or across molecules. chemical reactions usually involve the making or breaking of chemical bonds
is also higher at high temperature, as shown by carnot ' s theorem. in a conventional metallic engine, much of the energy released from the fuel must be dissipated as waste heat in order to prevent a meltdown of the metallic parts. despite all of these desirable properties, such engines are not in production because the manufacturing of ceramic parts in the requisite precision and durability is difficult. imperfection in the ceramic leads to cracks, which can lead to potentially dangerous equipment failure. such engines are possible in laboratory settings, but mass - production is not feasible with current technology. work is being done in developing ceramic parts for gas turbine engines. currently, even blades made of advanced metal alloys used in the engines ' hot section require cooling and careful limiting of operating temperatures. turbine engines made with ceramics could operate more efficiently, giving aircraft greater range and payload for a set amount of fuel. recently, there have been advances in ceramics which include bio - ceramics, such as dental implants and synthetic bones. hydroxyapatite, the natural mineral component of bone, has been made synthetically from a number of biological and chemical sources and can be formed into ceramic materials. orthopedic implants made from these materials bond readily to bone and other tissues in the body without rejection or inflammatory reactions. because of this, they are of great interest for gene delivery and tissue engineering scaffolds. most hydroxyapatite ceramics are very porous and lack mechanical strength and are used to coat metal orthopedic devices to aid in forming a bond to bone or as bone fillers. they are also used as fillers for orthopedic plastic screws to aid in reducing the inflammation and increase absorption of these plastic materials. work is being done to make strong, fully dense nano crystalline hydroxyapatite ceramic materials for orthopedic weight bearing devices, replacing foreign metal and plastic orthopedic materials with a synthetic, but naturally occurring, bone mineral. ultimately these ceramic materials may be used as bone replacements or with the incorporation of protein collagens, synthetic bones. durable actinide - containing ceramic materials have many applications such as in nuclear fuels for burning excess pu and in chemically - inert sources of alpha irradiation for power supply of unmanned space vehicles or to produce electricity for microelectronic devices. both use and disposal of radioactive actinides require their immobilization in a durable host material. nuclear waste long - lived radionuclides such as actinides are immobilized using chemical
possible states of energy for electrons, atoms and molecules. these are determined by the rules of quantum mechanics, which require quantization of energy of a bound system. the atoms / molecules in a higher energy state are said to be excited. the molecules / atoms of substance in an excited energy state are often much more reactive ; that is, more amenable to chemical reactions. the phase of a substance is invariably determined by its energy and the energy of its surroundings. when the intermolecular forces of a substance are such that the energy of the surroundings is not sufficient to overcome them, it occurs in a more ordered phase like liquid or solid as is the case with water ( h2o ) ; a liquid at room temperature because its molecules are bound by hydrogen bonds. whereas hydrogen sulfide ( h2s ) is a gas at room temperature and standard pressure, as its molecules are bound by weaker dipole β dipole interactions. the transfer of energy from one chemical substance to another depends on the size of energy quanta emitted from one substance. however, heat energy is often transferred more easily from almost any substance to another because the phonons responsible for vibrational and rotational energy levels in a substance have much less energy than photons invoked for the electronic energy transfer. thus, because vibrational and rotational energy levels are more closely spaced than electronic energy levels, heat is more easily transferred between substances relative to light or other forms of electronic energy. for example, ultraviolet electromagnetic radiation is not transferred with as much efficacy from one substance to another as thermal or electrical energy. the existence of characteristic energy levels for different chemical substances is useful for their identification by the analysis of spectral lines. different kinds of spectra are often used in chemical spectroscopy, e. g. ir, microwave, nmr, esr, etc. spectroscopy is also used to identify the composition of remote objects β like stars and distant galaxies β by analyzing their radiation spectra. the term chemical energy is often used to indicate the potential of a chemical substance to undergo a transformation through a chemical reaction or to transform other chemical substances. = = = reaction = = = when a chemical substance is transformed as a result of its interaction with another substance or with energy, a chemical reaction is said to have occurred. a chemical reaction is therefore a concept related to the " reaction " of a substance when it comes in close contact with another, whether as a mixture or a solution ; exposure to some form of energy, or both. it results in some energy exchange between the constituents of the reaction as well as
endothermic reactions, the reaction absorbs heat from the surroundings. chemical reactions are invariably not possible unless the reactants surmount an energy barrier known as the activation energy. the speed of a chemical reaction ( at given temperature t ) is related to the activation energy e, by the boltzmann ' s population factor e β e / k t { \ displaystyle e ^ { - e / kt } } β that is the probability of a molecule to have energy greater than or equal to e at the given temperature t. this exponential dependence of a reaction rate on temperature is known as the arrhenius equation. the activation energy necessary for a chemical reaction to occur can be in the form of heat, light, electricity or mechanical force in the form of ultrasound. a related concept free energy, which also incorporates entropy considerations, is a very useful means for predicting the feasibility of a reaction and determining the state of equilibrium of a chemical reaction, in chemical thermodynamics. a reaction is feasible only if the total change in the gibbs free energy is negative, Ξ΄ g β€ 0 { \ displaystyle \ delta g \ leq 0 \, } ; if it is equal to zero the chemical reaction is said to be at equilibrium. there exist only limited possible states of energy for electrons, atoms and molecules. these are determined by the rules of quantum mechanics, which require quantization of energy of a bound system. the atoms / molecules in a higher energy state are said to be excited. the molecules / atoms of substance in an excited energy state are often much more reactive ; that is, more amenable to chemical reactions. the phase of a substance is invariably determined by its energy and the energy of its surroundings. when the intermolecular forces of a substance are such that the energy of the surroundings is not sufficient to overcome them, it occurs in a more ordered phase like liquid or solid as is the case with water ( h2o ) ; a liquid at room temperature because its molecules are bound by hydrogen bonds. whereas hydrogen sulfide ( h2s ) is a gas at room temperature and standard pressure, as its molecules are bound by weaker dipole β dipole interactions. the transfer of energy from one chemical substance to another depends on the size of energy quanta emitted from one substance. however, heat energy is often transferred more easily from almost any substance to another because the phonons responsible for vibrational and rotational energy levels in a substance have much less energy than photons invoked for the electronic energy transfer
. a reaction is said to be exergonic if the final state is lower on the energy scale than the initial state ; in the case of endergonic reactions the situation is the reverse. a reaction is said to be exothermic if the reaction releases heat to the surroundings ; in the case of endothermic reactions, the reaction absorbs heat from the surroundings. chemical reactions are invariably not possible unless the reactants surmount an energy barrier known as the activation energy. the speed of a chemical reaction ( at given temperature t ) is related to the activation energy e, by the boltzmann ' s population factor e β e / k t { \ displaystyle e ^ { - e / kt } } β that is the probability of a molecule to have energy greater than or equal to e at the given temperature t. this exponential dependence of a reaction rate on temperature is known as the arrhenius equation. the activation energy necessary for a chemical reaction to occur can be in the form of heat, light, electricity or mechanical force in the form of ultrasound. a related concept free energy, which also incorporates entropy considerations, is a very useful means for predicting the feasibility of a reaction and determining the state of equilibrium of a chemical reaction, in chemical thermodynamics. a reaction is feasible only if the total change in the gibbs free energy is negative, Ξ΄ g β€ 0 { \ displaystyle \ delta g \ leq 0 \, } ; if it is equal to zero the chemical reaction is said to be at equilibrium. there exist only limited possible states of energy for electrons, atoms and molecules. these are determined by the rules of quantum mechanics, which require quantization of energy of a bound system. the atoms / molecules in a higher energy state are said to be excited. the molecules / atoms of substance in an excited energy state are often much more reactive ; that is, more amenable to chemical reactions. the phase of a substance is invariably determined by its energy and the energy of its surroundings. when the intermolecular forces of a substance are such that the energy of the surroundings is not sufficient to overcome them, it occurs in a more ordered phase like liquid or solid as is the case with water ( h2o ) ; a liquid at room temperature because its molecules are bound by hydrogen bonds. whereas hydrogen sulfide ( h2s ) is a gas at room temperature and standard pressure, as its molecules are bound by weaker dipole β dipole interactions. the transfer of
activation energy necessary for a chemical reaction to occur can be in the form of heat, light, electricity or mechanical force in the form of ultrasound. a related concept free energy, which also incorporates entropy considerations, is a very useful means for predicting the feasibility of a reaction and determining the state of equilibrium of a chemical reaction, in chemical thermodynamics. a reaction is feasible only if the total change in the gibbs free energy is negative, Ξ΄ g β€ 0 { \ displaystyle \ delta g \ leq 0 \, } ; if it is equal to zero the chemical reaction is said to be at equilibrium. there exist only limited possible states of energy for electrons, atoms and molecules. these are determined by the rules of quantum mechanics, which require quantization of energy of a bound system. the atoms / molecules in a higher energy state are said to be excited. the molecules / atoms of substance in an excited energy state are often much more reactive ; that is, more amenable to chemical reactions. the phase of a substance is invariably determined by its energy and the energy of its surroundings. when the intermolecular forces of a substance are such that the energy of the surroundings is not sufficient to overcome them, it occurs in a more ordered phase like liquid or solid as is the case with water ( h2o ) ; a liquid at room temperature because its molecules are bound by hydrogen bonds. whereas hydrogen sulfide ( h2s ) is a gas at room temperature and standard pressure, as its molecules are bound by weaker dipole β dipole interactions. the transfer of energy from one chemical substance to another depends on the size of energy quanta emitted from one substance. however, heat energy is often transferred more easily from almost any substance to another because the phonons responsible for vibrational and rotational energy levels in a substance have much less energy than photons invoked for the electronic energy transfer. thus, because vibrational and rotational energy levels are more closely spaced than electronic energy levels, heat is more easily transferred between substances relative to light or other forms of electronic energy. for example, ultraviolet electromagnetic radiation is not transferred with as much efficacy from one substance to another as thermal or electrical energy. the existence of characteristic energy levels for different chemical substances is useful for their identification by the analysis of spectral lines. different kinds of spectra are often used in chemical spectroscopy, e. g. ir, microwave, nmr, esr, etc. spectroscopy is also used to identify the composition of remote objects β like stars and distant galaxies β by
Question: Temperature is the measurement of which type of energy?
A) chemical
B) mechanical
C) potential
D) kinetic
|
D) kinetic
|
Context:
best - known and controversial applications of genetic engineering is the creation and use of genetically modified crops or genetically modified livestock to produce genetically modified food. crops have been developed to increase production, increase tolerance to abiotic stresses, alter the composition of the food, or to produce novel products. the first crops to be released commercially on a large scale provided protection from insect pests or tolerance to herbicides. fungal and virus resistant crops have also been developed or are in development. this makes the insect and weed management of crops easier and can indirectly increase crop yield. gm crops that directly improve yield by accelerating growth or making the plant more hardy ( by improving salt, cold or drought tolerance ) are also under development. in 2016 salmon have been genetically modified with growth hormones to reach normal adult size much faster. gmos have been developed that modify the quality of produce by increasing the nutritional value or providing more industrially useful qualities or quantities. the amflora potato produces a more industrially useful blend of starches. soybeans and canola have been genetically modified to produce more healthy oils. the first commercialised gm food was a tomato that had delayed ripening, increasing its shelf life. plants and animals have been engineered to produce materials they do not normally make. pharming uses crops and animals as bioreactors to produce vaccines, drug intermediates, or the drugs themselves ; the useful product is purified from the harvest and then used in the standard pharmaceutical production process. cows and goats have been engineered to express drugs and other proteins in their milk, and in 2009 the fda approved a drug produced in goat milk. = = = other applications = = = genetic engineering has potential applications in conservation and natural area management. gene transfer through viral vectors has been proposed as a means of controlling invasive species as well as vaccinating threatened fauna from disease. transgenic trees have been suggested as a way to confer resistance to pathogens in wild populations. with the increasing risks of maladaptation in organisms as a result of climate change and other perturbations, facilitated adaptation through gene tweaking could be one solution to reducing extinction risks. applications of genetic engineering in conservation are thus far mostly theoretical and have yet to be put into practice. genetic engineering is also being used to create microbial art. some bacteria have been genetically engineered to create black and white photographs. novelty items such as lavender - colored carnations, blue roses, and glowing fish, have also been produced through genetic engineering. = = regulation = = the regulation of genetic engineering
cellular and molecular biology of cereals, grasses and monocots generally. model plants such as arabidopsis thaliana are used for studying the molecular biology of plant cells and the chloroplast. ideally, these organisms have small genomes that are well known or completely sequenced, small stature and short generation times. corn has been used to study mechanisms of photosynthesis and phloem loading of sugar in c4 plants. the single celled green alga chlamydomonas reinhardtii, while not an embryophyte itself, contains a green - pigmented chloroplast related to that of land plants, making it useful for study. a red alga cyanidioschyzon merolae has also been used to study some basic chloroplast functions. spinach, peas, soybeans and a moss physcomitrella patens are commonly used to study plant cell biology. agrobacterium tumefaciens, a soil rhizosphere bacterium, can attach to plant cells and infect them with a callus - inducing ti plasmid by horizontal gene transfer, causing a callus infection called crown gall disease. schell and van montagu ( 1977 ) hypothesised that the ti plasmid could be a natural vector for introducing the nif gene responsible for nitrogen fixation in the root nodules of legumes and other plant species. today, genetic modification of the ti plasmid is one of the main techniques for introduction of transgenes to plants and the creation of genetically modified crops. = = = epigenetics = = = epigenetics is the study of heritable changes in gene function that cannot be explained by changes in the underlying dna sequence but cause the organism ' s genes to behave ( or " express themselves " ) differently. one example of epigenetic change is the marking of the genes by dna methylation which determines whether they will be expressed or not. gene expression can also be controlled by repressor proteins that attach to silencer regions of the dna and prevent that region of the dna code from being expressed. epigenetic marks may be added or removed from the dna during programmed stages of development of the plant, and are responsible, for example, for the differences between anthers, petals and normal leaves, despite the fact that they all have the same underlying genetic code. epigenetic changes may be temporary or may remain through successive cell divisions for the remainder of
the best - suited crops ( e. g., those with the highest yields ) to produce enough food to support a growing population. as crops and fields became increasingly large and difficult to maintain, it was discovered that specific organisms and their by - products could effectively fertilize, restore nitrogen, and control pests. throughout the history of agriculture, farmers have inadvertently altered the genetics of their crops through introducing them to new environments and breeding them with other plants β one of the first forms of biotechnology. these processes also were included in early fermentation of beer. these processes were introduced in early mesopotamia, egypt, china and india, and still use the same basic biological methods. in brewing, malted grains ( containing enzymes ) convert starch from grains into sugar and then adding specific yeasts to produce beer. in this process, carbohydrates in the grains broke down into alcohols, such as ethanol. later, other cultures produced the process of lactic acid fermentation, which produced other preserved foods, such as soy sauce. fermentation was also used in this time period to produce leavened bread. although the process of fermentation was not fully understood until louis pasteur ' s work in 1857, it is still the first use of biotechnology to convert a food source into another form. before the time of charles darwin ' s work and life, animal and plant scientists had already used selective breeding. darwin added to that body of work with his scientific observations about the ability of science to change species. these accounts contributed to darwin ' s theory of natural selection. for thousands of years, humans have used selective breeding to improve the production of crops and livestock to use them for food. in selective breeding, organisms with desirable characteristics are mated to produce offspring with the same characteristics. for example, this technique was used with corn to produce the largest and sweetest crops. in the early twentieth century scientists gained a greater understanding of microbiology and explored ways of manufacturing specific products. in 1917, chaim weizmann first used a pure microbiological culture in an industrial process, that of manufacturing corn starch using clostridium acetobutylicum, to produce acetone, which the united kingdom desperately needed to manufacture explosives during world war i. biotechnology has also led to the development of antibiotics. in 1928, alexander fleming discovered the mold penicillium. his work led to the purification of the antibiotic formed by the mold by howard florey, ernst boris chain and norman heatley β to form
industrial applications. this branch of biotechnology is the most used for the industries of refining and combustion principally on the production of bio - oils with photosynthetic micro - algae. green biotechnology is biotechnology applied to agricultural processes. an example would be the selection and domestication of plants via micropropagation. another example is the designing of transgenic plants to grow under specific environments in the presence ( or absence ) of chemicals. one hope is that green biotechnology might produce more environmentally friendly solutions than traditional industrial agriculture. an example of this is the engineering of a plant to express a pesticide, thereby ending the need of external application of pesticides. an example of this would be bt corn. whether or not green biotechnology products such as this are ultimately more environmentally friendly is a topic of considerable debate. it is commonly considered as the next phase of green revolution, which can be seen as a platform to eradicate world hunger by using technologies which enable the production of more fertile and resistant, towards biotic and abiotic stress, plants and ensures application of environmentally friendly fertilizers and the use of biopesticides, it is mainly focused on the development of agriculture. on the other hand, some of the uses of green biotechnology involve microorganisms to clean and reduce waste. red biotechnology is the use of biotechnology in the medical and pharmaceutical industries, and health preservation. this branch involves the production of vaccines and antibiotics, regenerative therapies, creation of artificial organs and new diagnostics of diseases. as well as the development of hormones, stem cells, antibodies, sirna and diagnostic tests. white biotechnology, also known as industrial biotechnology, is biotechnology applied to industrial processes. an example is the designing of an organism to produce a useful chemical. another example is the using of enzymes as industrial catalysts to either produce valuable chemicals or destroy hazardous / polluting chemicals. white biotechnology tends to consume less in resources than traditional processes used to produce industrial goods. yellow biotechnology refers to the use of biotechnology in food production ( food industry ), for example in making wine ( winemaking ), cheese ( cheesemaking ), and beer ( brewing ) by fermentation. it has also been used to refer to biotechnology applied to insects. this includes biotechnology - based approaches for the control of harmful insects, the characterisation and utilisation of active ingredients or genes of insects for research, or application in agriculture and medicine and various other approaches. gray biotechnology is dedicated to environmental applications, and focused on the maintenance of biodiversity and the remotion of poll
the broad definition of " utilizing a biotechnological system to make products ". indeed, the cultivation of plants may be viewed as the earliest biotechnological enterprise. agriculture has been theorized to have become the dominant way of producing food since the neolithic revolution. through early biotechnology, the earliest farmers selected and bred the best - suited crops ( e. g., those with the highest yields ) to produce enough food to support a growing population. as crops and fields became increasingly large and difficult to maintain, it was discovered that specific organisms and their by - products could effectively fertilize, restore nitrogen, and control pests. throughout the history of agriculture, farmers have inadvertently altered the genetics of their crops through introducing them to new environments and breeding them with other plants β one of the first forms of biotechnology. these processes also were included in early fermentation of beer. these processes were introduced in early mesopotamia, egypt, china and india, and still use the same basic biological methods. in brewing, malted grains ( containing enzymes ) convert starch from grains into sugar and then adding specific yeasts to produce beer. in this process, carbohydrates in the grains broke down into alcohols, such as ethanol. later, other cultures produced the process of lactic acid fermentation, which produced other preserved foods, such as soy sauce. fermentation was also used in this time period to produce leavened bread. although the process of fermentation was not fully understood until louis pasteur ' s work in 1857, it is still the first use of biotechnology to convert a food source into another form. before the time of charles darwin ' s work and life, animal and plant scientists had already used selective breeding. darwin added to that body of work with his scientific observations about the ability of science to change species. these accounts contributed to darwin ' s theory of natural selection. for thousands of years, humans have used selective breeding to improve the production of crops and livestock to use them for food. in selective breeding, organisms with desirable characteristics are mated to produce offspring with the same characteristics. for example, this technique was used with corn to produce the largest and sweetest crops. in the early twentieth century scientists gained a greater understanding of microbiology and explored ways of manufacturing specific products. in 1917, chaim weizmann first used a pure microbiological culture in an industrial process, that of manufacturing corn starch using clostridium acetobutylicum, to produce acetone, which the united
process of lactic acid fermentation, which produced other preserved foods, such as soy sauce. fermentation was also used in this time period to produce leavened bread. although the process of fermentation was not fully understood until louis pasteur ' s work in 1857, it is still the first use of biotechnology to convert a food source into another form. before the time of charles darwin ' s work and life, animal and plant scientists had already used selective breeding. darwin added to that body of work with his scientific observations about the ability of science to change species. these accounts contributed to darwin ' s theory of natural selection. for thousands of years, humans have used selective breeding to improve the production of crops and livestock to use them for food. in selective breeding, organisms with desirable characteristics are mated to produce offspring with the same characteristics. for example, this technique was used with corn to produce the largest and sweetest crops. in the early twentieth century scientists gained a greater understanding of microbiology and explored ways of manufacturing specific products. in 1917, chaim weizmann first used a pure microbiological culture in an industrial process, that of manufacturing corn starch using clostridium acetobutylicum, to produce acetone, which the united kingdom desperately needed to manufacture explosives during world war i. biotechnology has also led to the development of antibiotics. in 1928, alexander fleming discovered the mold penicillium. his work led to the purification of the antibiotic formed by the mold by howard florey, ernst boris chain and norman heatley β to form what we today know as penicillin. in 1940, penicillin became available for medicinal use to treat bacterial infections in humans. the field of modern biotechnology is generally thought of as having been born in 1971 when paul berg ' s ( stanford ) experiments in gene splicing had early success. herbert w. boyer ( univ. calif. at san francisco ) and stanley n. cohen ( stanford ) significantly advanced the new technology in 1972 by transferring genetic material into a bacterium, such that the imported material would be reproduced. the commercial viability of a biotechnology industry was significantly expanded on june 16, 1980, when the united states supreme court ruled that a genetically modified microorganism could be patented in the case of diamond v. chakrabarty. indian - born ananda chakrabarty, working for general electric, had modified a bacterium ( of the genus pseudomonas ) capable of breaking down crude oil, which he proposed to
eat them. plants and other photosynthetic organisms are at the base of most food chains because they use the energy from the sun and nutrients from the soil and atmosphere, converting them into a form that can be used by animals. this is what ecologists call the first trophic level. the modern forms of the major staple foods, such as hemp, teff, maize, rice, wheat and other cereal grasses, pulses, bananas and plantains, as well as hemp, flax and cotton grown for their fibres, are the outcome of prehistoric selection over thousands of years from among wild ancestral plants with the most desirable characteristics. botanists study how plants produce food and how to increase yields, for example through plant breeding, making their work important to humanity ' s ability to feed the world and provide food security for future generations. botanists also study weeds, which are a considerable problem in agriculture, and the biology and control of plant pathogens in agriculture and natural ecosystems. ethnobotany is the study of the relationships between plants and people. when applied to the investigation of historical plant β people relationships ethnobotany may be referred to as archaeobotany or palaeoethnobotany. some of the earliest plant - people relationships arose between the indigenous people of canada in identifying edible plants from inedible plants. this relationship the indigenous people had with plants was recorded by ethnobotanists. = = plant biochemistry = = plant biochemistry is the study of the chemical processes used by plants. some of these processes are used in their primary metabolism like the photosynthetic calvin cycle and crassulacean acid metabolism. others make specialised materials like the cellulose and lignin used to build their bodies, and secondary products like resins and aroma compounds. plants and various other groups of photosynthetic eukaryotes collectively known as " algae " have unique organelles known as chloroplasts. chloroplasts are thought to be descended from cyanobacteria that formed endosymbiotic relationships with ancient plant and algal ancestors. chloroplasts and cyanobacteria contain the blue - green pigment chlorophyll a. chlorophyll a ( as well as its plant and green algal - specific cousin chlorophyll b ) absorbs light in the blue - violet and orange / red parts of the spectrum while reflecting and transmitting the green light that we see as the characteristic colour
on a large scale provided protection from insect pests or tolerance to herbicides. fungal and virus resistant crops have also been developed or are in development. this makes the insect and weed management of crops easier and can indirectly increase crop yield. gm crops that directly improve yield by accelerating growth or making the plant more hardy ( by improving salt, cold or drought tolerance ) are also under development. in 2016 salmon have been genetically modified with growth hormones to reach normal adult size much faster. gmos have been developed that modify the quality of produce by increasing the nutritional value or providing more industrially useful qualities or quantities. the amflora potato produces a more industrially useful blend of starches. soybeans and canola have been genetically modified to produce more healthy oils. the first commercialised gm food was a tomato that had delayed ripening, increasing its shelf life. plants and animals have been engineered to produce materials they do not normally make. pharming uses crops and animals as bioreactors to produce vaccines, drug intermediates, or the drugs themselves ; the useful product is purified from the harvest and then used in the standard pharmaceutical production process. cows and goats have been engineered to express drugs and other proteins in their milk, and in 2009 the fda approved a drug produced in goat milk. = = = other applications = = = genetic engineering has potential applications in conservation and natural area management. gene transfer through viral vectors has been proposed as a means of controlling invasive species as well as vaccinating threatened fauna from disease. transgenic trees have been suggested as a way to confer resistance to pathogens in wild populations. with the increasing risks of maladaptation in organisms as a result of climate change and other perturbations, facilitated adaptation through gene tweaking could be one solution to reducing extinction risks. applications of genetic engineering in conservation are thus far mostly theoretical and have yet to be put into practice. genetic engineering is also being used to create microbial art. some bacteria have been genetically engineered to create black and white photographs. novelty items such as lavender - colored carnations, blue roses, and glowing fish, have also been produced through genetic engineering. = = regulation = = the regulation of genetic engineering concerns the approaches taken by governments to assess and manage the risks associated with the development and release of gmos. the development of a regulatory framework began in 1975, at asilomar, california. the asilomar meeting recommended a set of voluntary guidelines regarding the use of recombinant technology. as the technology improved
the designing of transgenic plants to grow under specific environments in the presence ( or absence ) of chemicals. one hope is that green biotechnology might produce more environmentally friendly solutions than traditional industrial agriculture. an example of this is the engineering of a plant to express a pesticide, thereby ending the need of external application of pesticides. an example of this would be bt corn. whether or not green biotechnology products such as this are ultimately more environmentally friendly is a topic of considerable debate. it is commonly considered as the next phase of green revolution, which can be seen as a platform to eradicate world hunger by using technologies which enable the production of more fertile and resistant, towards biotic and abiotic stress, plants and ensures application of environmentally friendly fertilizers and the use of biopesticides, it is mainly focused on the development of agriculture. on the other hand, some of the uses of green biotechnology involve microorganisms to clean and reduce waste. red biotechnology is the use of biotechnology in the medical and pharmaceutical industries, and health preservation. this branch involves the production of vaccines and antibiotics, regenerative therapies, creation of artificial organs and new diagnostics of diseases. as well as the development of hormones, stem cells, antibodies, sirna and diagnostic tests. white biotechnology, also known as industrial biotechnology, is biotechnology applied to industrial processes. an example is the designing of an organism to produce a useful chemical. another example is the using of enzymes as industrial catalysts to either produce valuable chemicals or destroy hazardous / polluting chemicals. white biotechnology tends to consume less in resources than traditional processes used to produce industrial goods. yellow biotechnology refers to the use of biotechnology in food production ( food industry ), for example in making wine ( winemaking ), cheese ( cheesemaking ), and beer ( brewing ) by fermentation. it has also been used to refer to biotechnology applied to insects. this includes biotechnology - based approaches for the control of harmful insects, the characterisation and utilisation of active ingredients or genes of insects for research, or application in agriculture and medicine and various other approaches. gray biotechnology is dedicated to environmental applications, and focused on the maintenance of biodiversity and the remotion of pollutants. brown biotechnology is related to the management of arid lands and deserts. one application is the creation of enhanced seeds that resist extreme environmental conditions of arid regions, which is related to the innovation, creation of agriculture techniques and management of resources. violet biotechnology is related to law, ethical and philosophical issues around biotechnology. micro
. throughout the history of agriculture, farmers have inadvertently altered the genetics of their crops through introducing them to new environments and breeding them with other plants β one of the first forms of biotechnology. these processes also were included in early fermentation of beer. these processes were introduced in early mesopotamia, egypt, china and india, and still use the same basic biological methods. in brewing, malted grains ( containing enzymes ) convert starch from grains into sugar and then adding specific yeasts to produce beer. in this process, carbohydrates in the grains broke down into alcohols, such as ethanol. later, other cultures produced the process of lactic acid fermentation, which produced other preserved foods, such as soy sauce. fermentation was also used in this time period to produce leavened bread. although the process of fermentation was not fully understood until louis pasteur ' s work in 1857, it is still the first use of biotechnology to convert a food source into another form. before the time of charles darwin ' s work and life, animal and plant scientists had already used selective breeding. darwin added to that body of work with his scientific observations about the ability of science to change species. these accounts contributed to darwin ' s theory of natural selection. for thousands of years, humans have used selective breeding to improve the production of crops and livestock to use them for food. in selective breeding, organisms with desirable characteristics are mated to produce offspring with the same characteristics. for example, this technique was used with corn to produce the largest and sweetest crops. in the early twentieth century scientists gained a greater understanding of microbiology and explored ways of manufacturing specific products. in 1917, chaim weizmann first used a pure microbiological culture in an industrial process, that of manufacturing corn starch using clostridium acetobutylicum, to produce acetone, which the united kingdom desperately needed to manufacture explosives during world war i. biotechnology has also led to the development of antibiotics. in 1928, alexander fleming discovered the mold penicillium. his work led to the purification of the antibiotic formed by the mold by howard florey, ernst boris chain and norman heatley β to form what we today know as penicillin. in 1940, penicillin became available for medicinal use to treat bacterial infections in humans. the field of modern biotechnology is generally thought of as having been born in 1971 when paul berg ' s ( stanford ) experiments in gene splicing had early success. herbert w. boyer
Question: Scientists are discovering ways to make some crops use less water. A person working in which job would most benefit from this research?
A) a farmer
B) a doctor
C) a waiter
D) a cook
|
A) a farmer
|
Context:
various forms that are characteristic of its life cycle. there are four key processes that underlie development : determination, differentiation, morphogenesis, and growth. determination sets the developmental fate of a cell, which becomes more restrictive during development. differentiation is the process by which specialized cells arise from less specialized cells such as stem cells. stem cells are undifferentiated or partially differentiated cells that can differentiate into various types of cells and proliferate indefinitely to produce more of the same stem cell. cellular differentiation dramatically changes a cell ' s size, shape, membrane potential, metabolic activity, and responsiveness to signals, which are largely due to highly controlled modifications in gene expression and epigenetics. with a few exceptions, cellular differentiation almost never involves a change in the dna sequence itself. thus, different cells can have very different physical characteristics despite having the same genome. morphogenesis, or the development of body form, is the result of spatial differences in gene expression. a small fraction of the genes in an organism ' s genome called the developmental - genetic toolkit control the development of that organism. these toolkit genes are highly conserved among phyla, meaning that they are ancient and very similar in widely separated groups of animals. differences in deployment of toolkit genes affect the body plan and the number, identity, and pattern of body parts. among the most important toolkit genes are the hox genes. hox genes determine where repeating parts, such as the many vertebrae of snakes, will grow in a developing embryo or larva. = = evolution = = = = = evolutionary processes = = = evolution is a central organizing concept in biology. it is the change in heritable characteristics of populations over successive generations. in artificial selection, animals were selectively bred for specific traits. given that traits are inherited, populations contain a varied mix of traits, and reproduction is able to increase any population, darwin argued that in the natural world, it was nature that played the role of humans in selecting for specific traits. darwin inferred that individuals who possessed heritable traits better adapted to their environments are more likely to survive and produce more offspring than other individuals. he further inferred that this would lead to the accumulation of favorable traits over successive generations, thereby increasing the match between the organisms and their environment. = = = speciation = = = a species is a group of organisms that mate with one another and speciation is the process by which one lineage splits into two lineages as a result of having evolved independently from each other
to chromatin, which is a complex of dna and protein found in eukaryotic cells. = = = genes, development, and evolution = = = development is the process by which a multicellular organism ( plant or animal ) goes through a series of changes, starting from a single cell, and taking on various forms that are characteristic of its life cycle. there are four key processes that underlie development : determination, differentiation, morphogenesis, and growth. determination sets the developmental fate of a cell, which becomes more restrictive during development. differentiation is the process by which specialized cells arise from less specialized cells such as stem cells. stem cells are undifferentiated or partially differentiated cells that can differentiate into various types of cells and proliferate indefinitely to produce more of the same stem cell. cellular differentiation dramatically changes a cell ' s size, shape, membrane potential, metabolic activity, and responsiveness to signals, which are largely due to highly controlled modifications in gene expression and epigenetics. with a few exceptions, cellular differentiation almost never involves a change in the dna sequence itself. thus, different cells can have very different physical characteristics despite having the same genome. morphogenesis, or the development of body form, is the result of spatial differences in gene expression. a small fraction of the genes in an organism ' s genome called the developmental - genetic toolkit control the development of that organism. these toolkit genes are highly conserved among phyla, meaning that they are ancient and very similar in widely separated groups of animals. differences in deployment of toolkit genes affect the body plan and the number, identity, and pattern of body parts. among the most important toolkit genes are the hox genes. hox genes determine where repeating parts, such as the many vertebrae of snakes, will grow in a developing embryo or larva. = = evolution = = = = = evolutionary processes = = = evolution is a central organizing concept in biology. it is the change in heritable characteristics of populations over successive generations. in artificial selection, animals were selectively bred for specific traits. given that traits are inherited, populations contain a varied mix of traits, and reproduction is able to increase any population, darwin argued that in the natural world, it was nature that played the role of humans in selecting for specific traits. darwin inferred that individuals who possessed heritable traits better adapted to their environments are more likely to survive and produce more offspring than other individuals. he further inferred that this would lead to the
combination of chemical and mechanical stimulation. cell proliferation and differentiation are largely influenced by mechanical and biochemical cues in the surrounding extracellular matrix environment. bioreactors are typically developed to replicate the specific physiological environment of the tissue being grown ( e. g., flex and fluid shearing for heart tissue growth ). this can allow specialized cell lines to thrive in cultures replicating their native environments, but it also makes bioreactors attractive tools for culturing stem cells. a successful stem - cell - based bioreactor is effective at expanding stem cells with uniform properties and / or promoting controlled, reproducible differentiation into selected mature cell types. there are a variety of bioreactors designed for 3d cell cultures. there are small plastic cylindrical chambers, as well as glass chambers, with regulated internal humidity and moisture specifically engineered for the purpose of growing cells in three dimensions. the bioreactor uses bioactive synthetic materials such as polyethylene terephthalate membranes to surround the spheroid cells in an environment that maintains high levels of nutrients. they are easy to open and close, so that cell spheroids can be removed for testing, yet the chamber is able to maintain 100 % humidity throughout. this humidity is important to achieve maximum cell growth and function. the bioreactor chamber is part of a larger device that rotates to ensure equal cell growth in each direction across three dimensions. quinxell technologies now under quintech life sciences from singapore has developed a bioreactor known as the tisxell biaxial bioreactor which is specially designed for the purpose of tissue engineering. it is the first bioreactor in the world to have a spherical glass chamber with biaxial rotation ; specifically to mimic the rotation of the fetus in the womb ; which provides a conducive environment for the growth of tissues. multiple forms of mechanical stimulation have also been combined into a single bioreactor. using gene expression analysis, one academic study found that applying a combination of cyclic strain and ultrasound stimulation to pre - osteoblast cells in a bioreactor accelerated matrix maturation and differentiation. the technology of this combined stimulation bioreactor could be used to grow bone cells more quickly and effectively in future clinical stem cell therapies. mc2 biotek has also developed a bioreactor known as prototissue that uses gas exchange to maintain high oxygen levels within the cell chamber ; improving upon previous bioreactors, since the higher oxygen levels help the cell grow and undergo normal cell respiration. active
the cell ' s life. some epigenetic changes have been shown to be heritable, while others are reset in the germ cells. epigenetic changes in eukaryotic biology serve to regulate the process of cellular differentiation. during morphogenesis, totipotent stem cells become the various pluripotent cell lines of the embryo, which in turn become fully differentiated cells. a single fertilised egg cell, the zygote, gives rise to the many different plant cell types including parenchyma, xylem vessel elements, phloem sieve tubes, guard cells of the epidermis, etc. as it continues to divide. the process results from the epigenetic activation of some genes and inhibition of others. unlike animals, many plant cells, particularly those of the parenchyma, do not terminally differentiate, remaining totipotent with the ability to give rise to a new individual plant. exceptions include highly lignified cells, the sclerenchyma and xylem which are dead at maturity, and the phloem sieve tubes which lack nuclei. while plants use many of the same epigenetic mechanisms as animals, such as chromatin remodelling, an alternative hypothesis is that plants set their gene expression patterns using positional information from the environment and surrounding cells to determine their developmental fate. epigenetic changes can lead to paramutations, which do not follow the mendelian heritage rules. these epigenetic marks are carried from one generation to the next, with one allele inducing a change on the other. = = plant evolution = = the chloroplasts of plants have a number of biochemical, structural and genetic similarities to cyanobacteria, ( commonly but incorrectly known as " blue - green algae " ) and are thought to be derived from an ancient endosymbiotic relationship between an ancestral eukaryotic cell and a cyanobacterial resident. the algae are a polyphyletic group and are placed in various divisions, some more closely related to plants than others. there are many differences between them in features such as cell wall composition, biochemistry, pigmentation, chloroplast structure and nutrient reserves. the algal division charophyta, sister to the green algal division chlorophyta, is considered to contain the ancestor of true plants. the charophyte class charophyceae and the land plant sub - kingdom embryophy
can be activated by inducers are called inducible genes, in contrast to constitutive genes that are almost constantly active. in contrast to both, structural genes encode proteins that are not involved in gene regulation. in addition to regulatory events involving the promoter, gene expression can also be regulated by epigenetic changes to chromatin, which is a complex of dna and protein found in eukaryotic cells. = = = genes, development, and evolution = = = development is the process by which a multicellular organism ( plant or animal ) goes through a series of changes, starting from a single cell, and taking on various forms that are characteristic of its life cycle. there are four key processes that underlie development : determination, differentiation, morphogenesis, and growth. determination sets the developmental fate of a cell, which becomes more restrictive during development. differentiation is the process by which specialized cells arise from less specialized cells such as stem cells. stem cells are undifferentiated or partially differentiated cells that can differentiate into various types of cells and proliferate indefinitely to produce more of the same stem cell. cellular differentiation dramatically changes a cell ' s size, shape, membrane potential, metabolic activity, and responsiveness to signals, which are largely due to highly controlled modifications in gene expression and epigenetics. with a few exceptions, cellular differentiation almost never involves a change in the dna sequence itself. thus, different cells can have very different physical characteristics despite having the same genome. morphogenesis, or the development of body form, is the result of spatial differences in gene expression. a small fraction of the genes in an organism ' s genome called the developmental - genetic toolkit control the development of that organism. these toolkit genes are highly conserved among phyla, meaning that they are ancient and very similar in widely separated groups of animals. differences in deployment of toolkit genes affect the body plan and the number, identity, and pattern of body parts. among the most important toolkit genes are the hox genes. hox genes determine where repeating parts, such as the many vertebrae of snakes, will grow in a developing embryo or larva. = = evolution = = = = = evolutionary processes = = = evolution is a central organizing concept in biology. it is the change in heritable characteristics of populations over successive generations. in artificial selection, animals were selectively bred for specific traits. given that traits are inherited, populations contain a varied mix of traits, and reproduction is able to increase any population,
or removed from the dna during programmed stages of development of the plant, and are responsible, for example, for the differences between anthers, petals and normal leaves, despite the fact that they all have the same underlying genetic code. epigenetic changes may be temporary or may remain through successive cell divisions for the remainder of the cell ' s life. some epigenetic changes have been shown to be heritable, while others are reset in the germ cells. epigenetic changes in eukaryotic biology serve to regulate the process of cellular differentiation. during morphogenesis, totipotent stem cells become the various pluripotent cell lines of the embryo, which in turn become fully differentiated cells. a single fertilised egg cell, the zygote, gives rise to the many different plant cell types including parenchyma, xylem vessel elements, phloem sieve tubes, guard cells of the epidermis, etc. as it continues to divide. the process results from the epigenetic activation of some genes and inhibition of others. unlike animals, many plant cells, particularly those of the parenchyma, do not terminally differentiate, remaining totipotent with the ability to give rise to a new individual plant. exceptions include highly lignified cells, the sclerenchyma and xylem which are dead at maturity, and the phloem sieve tubes which lack nuclei. while plants use many of the same epigenetic mechanisms as animals, such as chromatin remodelling, an alternative hypothesis is that plants set their gene expression patterns using positional information from the environment and surrounding cells to determine their developmental fate. epigenetic changes can lead to paramutations, which do not follow the mendelian heritage rules. these epigenetic marks are carried from one generation to the next, with one allele inducing a change on the other. = = plant evolution = = the chloroplasts of plants have a number of biochemical, structural and genetic similarities to cyanobacteria, ( commonly but incorrectly known as " blue - green algae " ) and are thought to be derived from an ancient endosymbiotic relationship between an ancestral eukaryotic cell and a cyanobacterial resident. the algae are a polyphyletic group and are placed in various divisions, some more closely related to plants than others. there are many differences between them in features such as cell wall composition, biochemistry,
induced stem cells ( isc ) as related to the donor. = = = stem cells = = = stem cells are undifferentiated cells with the ability to divide in culture and give rise to different forms of specialized cells. stem cells are divided into " adult " and " embryonic " stem cells according to their source. while there is still a large ethical debate related to the use of embryonic stem cells, it is thought that another alternative source β induced pluripotent stem cells β may be useful for the repair of diseased or damaged tissues, or may be used to grow new organs. totipotent cells are stem cells which can divide into further stem cells or differentiate into any cell type in the body, including extra - embryonic tissue. pluripotent cells are stem cells which can differentiate into any cell type in the body except extra - embryonic tissue. induced pluripotent stem cells ( ipscs ) are subclass of pluripotent stem cells resembling embryonic stem cells ( escs ) that have been derived from adult differentiated cells. ipscs are created by altering the expression of transcriptional factors in adult cells until they become like embryonic stem cells. multipotent stem cells can be differentiated into any cell within the same class, such as blood or bone. a common example of multipotent cells is mesenchymal stem cells ( mscs ). = = scaffolds = = scaffolds are materials that have been engineered to cause desirable cellular interactions to contribute to the formation of new functional tissues for medical purposes. cells are often ' seeded ' into these structures capable of supporting three - dimensional tissue formation. scaffolds mimic the extracellular matrix of the native tissue, recapitulating the in vivo milieu and allowing cells to influence their own microenvironments. they usually serve at least one of the following purposes : allowing cell attachment and migration, delivering and retaining cells and biochemical factors, enabling diffusion of vital cell nutrients and expressed products, and exerting certain mechanical and biological influences to modify the behaviour of the cell phase. in 2009, an interdisciplinary team led by the thoracic surgeon thorsten walles implanted the first bioartificial transplant that provides an innate vascular network for post - transplant graft supply successfully into a patient awaiting tracheal reconstruction. to achieve the goal of tissue reconstruction, scaffolds must meet some specific requirements. high porosity and adequate pore size are necessary to facilitate cell seed
pluripotent cell lines of the embryo, which in turn become fully differentiated cells. a single fertilised egg cell, the zygote, gives rise to the many different plant cell types including parenchyma, xylem vessel elements, phloem sieve tubes, guard cells of the epidermis, etc. as it continues to divide. the process results from the epigenetic activation of some genes and inhibition of others. unlike animals, many plant cells, particularly those of the parenchyma, do not terminally differentiate, remaining totipotent with the ability to give rise to a new individual plant. exceptions include highly lignified cells, the sclerenchyma and xylem which are dead at maturity, and the phloem sieve tubes which lack nuclei. while plants use many of the same epigenetic mechanisms as animals, such as chromatin remodelling, an alternative hypothesis is that plants set their gene expression patterns using positional information from the environment and surrounding cells to determine their developmental fate. epigenetic changes can lead to paramutations, which do not follow the mendelian heritage rules. these epigenetic marks are carried from one generation to the next, with one allele inducing a change on the other. = = plant evolution = = the chloroplasts of plants have a number of biochemical, structural and genetic similarities to cyanobacteria, ( commonly but incorrectly known as " blue - green algae " ) and are thought to be derived from an ancient endosymbiotic relationship between an ancestral eukaryotic cell and a cyanobacterial resident. the algae are a polyphyletic group and are placed in various divisions, some more closely related to plants than others. there are many differences between them in features such as cell wall composition, biochemistry, pigmentation, chloroplast structure and nutrient reserves. the algal division charophyta, sister to the green algal division chlorophyta, is considered to contain the ancestor of true plants. the charophyte class charophyceae and the land plant sub - kingdom embryophyta together form the monophyletic group or clade streptophytina. nonvascular land plants are embryophytes that lack the vascular tissues xylem and phloem. they include mosses, liverworts and hornworts. pteridophytic vascular plants with true xyle
such as electromagnetic forces, mechanical pressures, or fluid pressures to the tissue. these systems can be two - or three - dimensional setups. bioreactors can be used in both academic and industry applications. general - use and application - specific bioreactors are also commercially available, which may provide static chemical stimulation or a combination of chemical and mechanical stimulation. cell proliferation and differentiation are largely influenced by mechanical and biochemical cues in the surrounding extracellular matrix environment. bioreactors are typically developed to replicate the specific physiological environment of the tissue being grown ( e. g., flex and fluid shearing for heart tissue growth ). this can allow specialized cell lines to thrive in cultures replicating their native environments, but it also makes bioreactors attractive tools for culturing stem cells. a successful stem - cell - based bioreactor is effective at expanding stem cells with uniform properties and / or promoting controlled, reproducible differentiation into selected mature cell types. there are a variety of bioreactors designed for 3d cell cultures. there are small plastic cylindrical chambers, as well as glass chambers, with regulated internal humidity and moisture specifically engineered for the purpose of growing cells in three dimensions. the bioreactor uses bioactive synthetic materials such as polyethylene terephthalate membranes to surround the spheroid cells in an environment that maintains high levels of nutrients. they are easy to open and close, so that cell spheroids can be removed for testing, yet the chamber is able to maintain 100 % humidity throughout. this humidity is important to achieve maximum cell growth and function. the bioreactor chamber is part of a larger device that rotates to ensure equal cell growth in each direction across three dimensions. quinxell technologies now under quintech life sciences from singapore has developed a bioreactor known as the tisxell biaxial bioreactor which is specially designed for the purpose of tissue engineering. it is the first bioreactor in the world to have a spherical glass chamber with biaxial rotation ; specifically to mimic the rotation of the fetus in the womb ; which provides a conducive environment for the growth of tissues. multiple forms of mechanical stimulation have also been combined into a single bioreactor. using gene expression analysis, one academic study found that applying a combination of cyclic strain and ultrasound stimulation to pre - osteoblast cells in a bioreactor accelerated matrix maturation and differentiation. the technology of this combined stimulation bioreactor could be used to grow bone cells more quickly and effectively
##ilage generated without the use of exogenous scaffold material. in this methodology, all material in the construct is cellular produced directly by the cells. bioartificial heart : doris taylor ' s lab constructed a biocompatible rat heart by re - cellularising a de - cellularised rat heart. this scaffold and cells were placed in a bioreactor, where it matured to become a partially or fully transplantable organ. the work was called a " landmark ". the lab first stripped the cells away from a rat heart ( a process called " decellularization " ) and then injected rat stem cells into the decellularized rat heart. tissue - engineered blood vessels : blood vessels that have been grown in a lab and can be used to repair damaged blood vessels without eliciting an immune response. tissue engineered blood vessels have been developed by many different approaches. they could be implanted as pre - seeded cellularized blood vessels, as acellular vascular grafts made with decellularized vessels or synthetic vascular grafts. artificial skin constructed from human skin cells embedded in a hydrogel, such as in the case of bio - printed constructs for battlefield burn repairs. artificial bone marrow : bone marrow cultured in vitro to be transplanted serves as a " just cells " approach to tissue engineering. tissue engineered bone : a structural matrix can be composed of metals such as titanium, polymers of varying degradation rates, or certain types of ceramics. materials are often chosen to recruit osteoblasts to aid in reforming the bone and returning biological function. various types of cells can be added directly into the matrix to expedite the process. laboratory - grown penis : decellularized scaffolds of rabbit penises were recellularised with smooth muscle and endothelial cells. the organ was then transplanted to live rabbits and functioned comparably to the native organ, suggesting potential as treatment for genital trauma. oral mucosa tissue engineering uses a cells and scaffold approach to replicate the 3 dimensional structure and function of oral mucosa. = = cells as building blocks = = cells are one of the main components for the success of tissue engineering approaches. tissue engineering uses cells as strategies for creation / replacement of new tissue. examples include fibroblasts used for skin repair or renewal, chondrocytes used for cartilage repair ( maci β fda approved product ), and hepatocytes used in liver support systems cells can be used alone or with
Question: Stem cells are cells that can differentiate into many different kinds of cells. Scientists have discovered that the environment supporting the cells influences stem-cell differentiation. For example, stem cells grown on strong, flexible surfaces became muscle cells. What most likely happened to stem cells grown on a hard, rigid surface?
A) They became nerve cells.
B) They became red blood cells.
C) They became bone cells.
D) They became skin cells.
|
C) They became bone cells.
|
Context:
there are a variety of bioreactors designed for 3d cell cultures. there are small plastic cylindrical chambers, as well as glass chambers, with regulated internal humidity and moisture specifically engineered for the purpose of growing cells in three dimensions. the bioreactor uses bioactive synthetic materials such as polyethylene terephthalate membranes to surround the spheroid cells in an environment that maintains high levels of nutrients. they are easy to open and close, so that cell spheroids can be removed for testing, yet the chamber is able to maintain 100 % humidity throughout. this humidity is important to achieve maximum cell growth and function. the bioreactor chamber is part of a larger device that rotates to ensure equal cell growth in each direction across three dimensions. quinxell technologies now under quintech life sciences from singapore has developed a bioreactor known as the tisxell biaxial bioreactor which is specially designed for the purpose of tissue engineering. it is the first bioreactor in the world to have a spherical glass chamber with biaxial rotation ; specifically to mimic the rotation of the fetus in the womb ; which provides a conducive environment for the growth of tissues. multiple forms of mechanical stimulation have also been combined into a single bioreactor. using gene expression analysis, one academic study found that applying a combination of cyclic strain and ultrasound stimulation to pre - osteoblast cells in a bioreactor accelerated matrix maturation and differentiation. the technology of this combined stimulation bioreactor could be used to grow bone cells more quickly and effectively in future clinical stem cell therapies. mc2 biotek has also developed a bioreactor known as prototissue that uses gas exchange to maintain high oxygen levels within the cell chamber ; improving upon previous bioreactors, since the higher oxygen levels help the cell grow and undergo normal cell respiration. active areas of research on bioreactors includes increasing production scale and refining the physiological environment, both of which could improve the efficiency and efficacy of bioreactors in research or clinical use. bioreactors are currently used to study, among other things, cell and tissue level therapies, cell and tissue response to specific physiological environment changes, and development of disease and injury. = = = long fiber generation = = = in 2013, a group from the university of tokyo developed cell laden fibers up to a meter in length and on the order of 100 ΞΌm in size. these fibers were created using a microfluidic device that forms a
river valley during ancient times. the papyrus was harvested by field workers and brought to processing centers where it was cut into thin strips. the strips were then laid - out side by side and covered in plant resin. the second layer of strips was laid on perpendicularly, then both pressed together until the sheet was dry. the sheets were then joined to form a roll and later used for writing. egyptian society made several significant advances during dynastic periods in many areas of technology. according to hossam elanzeery, they were the first civilization to use timekeeping devices such as sundials, shadow clocks, and obelisks and successfully leveraged their knowledge of astronomy to create a calendar model that society still uses today. they developed shipbuilding technology that saw them progress from papyrus reed vessels to cedar wood ships while also pioneering the use of rope trusses and stem - mounted rudders. the egyptians also used their knowledge of anatomy to lay the foundation for many modern medical techniques and practiced the earliest known version of neuroscience. elanzeery also states that they used and furthered mathematical science, as evidenced in the building of the pyramids. ancient egyptians also invented and pioneered many food technologies that have become the basis of modern food technology processes. based on paintings and reliefs found in tombs, as well as archaeological artifacts, scholars like paul t nicholson believe that the ancient egyptians established systematic farming practices, engaged in cereal processing, brewed beer and baked bread, processed meat, practiced viticulture and created the basis for modern wine production, and created condiments to complement, preserve and mask the flavors of their food. = = = = indus valley = = = = the indus valley civilization, situated in a resource - rich area ( in modern pakistan and northwestern india ), is notable for its early application of city planning, sanitation technologies, and plumbing. indus valley construction and architecture, called ' vaastu shastra ', suggests a thorough understanding of materials engineering, hydrology, and sanitation. = = = = china = = = = the chinese made many first - known discoveries and developments. major technological contributions from china include the earliest known form of the binary code and epigenetic sequencing, early seismological detectors, matches, paper, helicopter rotor, raised - relief map, the double - action piston pump, cast iron, water powered blast furnace bellows, the iron plough, the multi - tube seed drill, the wheelbarrow, the parachute, the compass, the rudder, the crossbow, the south pointing chariot and gunpowder
great pyramid of giza, which is 481 feet ( 147 meters ) high. they also made writing medium similar to paper from papyrus, which joshua mark states is the foundation for modern paper. papyrus is a plant ( cyperus papyrus ) which grew in plentiful amounts in the egyptian delta and throughout the nile river valley during ancient times. the papyrus was harvested by field workers and brought to processing centers where it was cut into thin strips. the strips were then laid - out side by side and covered in plant resin. the second layer of strips was laid on perpendicularly, then both pressed together until the sheet was dry. the sheets were then joined to form a roll and later used for writing. egyptian society made several significant advances during dynastic periods in many areas of technology. according to hossam elanzeery, they were the first civilization to use timekeeping devices such as sundials, shadow clocks, and obelisks and successfully leveraged their knowledge of astronomy to create a calendar model that society still uses today. they developed shipbuilding technology that saw them progress from papyrus reed vessels to cedar wood ships while also pioneering the use of rope trusses and stem - mounted rudders. the egyptians also used their knowledge of anatomy to lay the foundation for many modern medical techniques and practiced the earliest known version of neuroscience. elanzeery also states that they used and furthered mathematical science, as evidenced in the building of the pyramids. ancient egyptians also invented and pioneered many food technologies that have become the basis of modern food technology processes. based on paintings and reliefs found in tombs, as well as archaeological artifacts, scholars like paul t nicholson believe that the ancient egyptians established systematic farming practices, engaged in cereal processing, brewed beer and baked bread, processed meat, practiced viticulture and created the basis for modern wine production, and created condiments to complement, preserve and mask the flavors of their food. = = = = indus valley = = = = the indus valley civilization, situated in a resource - rich area ( in modern pakistan and northwestern india ), is notable for its early application of city planning, sanitation technologies, and plumbing. indus valley construction and architecture, called ' vaastu shastra ', suggests a thorough understanding of materials engineering, hydrology, and sanitation. = = = = china = = = = the chinese made many first - known discoveries and developments. major technological contributions from china include the earliest known form of the binary code and epigenetic sequencing, early seismological detectors,
such as electromagnetic forces, mechanical pressures, or fluid pressures to the tissue. these systems can be two - or three - dimensional setups. bioreactors can be used in both academic and industry applications. general - use and application - specific bioreactors are also commercially available, which may provide static chemical stimulation or a combination of chemical and mechanical stimulation. cell proliferation and differentiation are largely influenced by mechanical and biochemical cues in the surrounding extracellular matrix environment. bioreactors are typically developed to replicate the specific physiological environment of the tissue being grown ( e. g., flex and fluid shearing for heart tissue growth ). this can allow specialized cell lines to thrive in cultures replicating their native environments, but it also makes bioreactors attractive tools for culturing stem cells. a successful stem - cell - based bioreactor is effective at expanding stem cells with uniform properties and / or promoting controlled, reproducible differentiation into selected mature cell types. there are a variety of bioreactors designed for 3d cell cultures. there are small plastic cylindrical chambers, as well as glass chambers, with regulated internal humidity and moisture specifically engineered for the purpose of growing cells in three dimensions. the bioreactor uses bioactive synthetic materials such as polyethylene terephthalate membranes to surround the spheroid cells in an environment that maintains high levels of nutrients. they are easy to open and close, so that cell spheroids can be removed for testing, yet the chamber is able to maintain 100 % humidity throughout. this humidity is important to achieve maximum cell growth and function. the bioreactor chamber is part of a larger device that rotates to ensure equal cell growth in each direction across three dimensions. quinxell technologies now under quintech life sciences from singapore has developed a bioreactor known as the tisxell biaxial bioreactor which is specially designed for the purpose of tissue engineering. it is the first bioreactor in the world to have a spherical glass chamber with biaxial rotation ; specifically to mimic the rotation of the fetus in the womb ; which provides a conducive environment for the growth of tissues. multiple forms of mechanical stimulation have also been combined into a single bioreactor. using gene expression analysis, one academic study found that applying a combination of cyclic strain and ultrasound stimulation to pre - osteoblast cells in a bioreactor accelerated matrix maturation and differentiation. the technology of this combined stimulation bioreactor could be used to grow bone cells more quickly and effectively
are commonly referred to as " cross - hatching ". phantom β ( not shown ) are alternately long - and double short - dashed thin lines used to represent a feature or component that is not part of the specified part or assembly. e. g. billet ends that may be used for testing, or the machined product that is the focus of a tooling drawing. lines can also be classified by a letter classification in which each line is given a letter. type a lines show the outline of the feature of an object. they are the thickest lines on a drawing and done with a pencil softer than hb. type b lines are dimension lines and are used for dimensioning, projecting, extending, or leaders. a harder pencil should be used, such as a 2h pencil. type c lines are used for breaks when the whole object is not shown. these are freehand drawn and only for short breaks. 2h pencil type d lines are similar to type c, except these are zigzagged and only for longer breaks. 2h pencil type e lines indicate hidden outlines of internal features of an object. these are dotted lines. 2h pencil type f lines are type e lines, except these are used for drawings in electrotechnology. 2h pencil type g lines are used for centre lines. these are dotted lines, but a long line of 10 β 20 mm, then a 1 mm gap, then a small line of 2 mm. 2h pencil type h lines are the same as type g, except that every second long line is thicker. these indicate the cutting plane of an object. 2h pencil type k lines indicate the alternate positions of an object and the line taken by that object. these are drawn with a long line of 10 β 20 mm, then a small gap, then a small line of 2 mm, then a gap, then another small line. 2h pencil. = = = multiple views and projections = = = in most cases, a single view is not sufficient to show all necessary features, and several views are used. types of views include the following : = = = = multiview projection = = = = a multiview projection is a type of orthographic projection that shows the object as it looks from the front, right, left, top, bottom, or back ( e. g. the primary views ), and is typically positioned relative to each other according to the rules of either first - angle or third - angle projection. the origin and vector direction of the projectors (
casting, also called the lost wax process, die casting, centrifugal casting, both vertical and horizontal, and continuous castings. each of these forms has advantages for certain metals and applications considering factors like magnetism and corrosion. forging β a red - hot billet is hammered into shape. rolling β a billet is passed through successively narrower rollers to create a sheet. extrusion β a hot and malleable metal is forced under pressure through a die, which shapes it before it cools. machining β lathes, milling machines and drills cut the cold metal to shape. sintering β a powdered metal is heated in a non - oxidizing environment after being compressed into a die. fabrication β sheets of metal are cut with guillotines or gas cutters and bent and welded into structural shape. laser cladding β metallic powder is blown through a movable laser beam ( e. g. mounted on a nc 5 - axis machine ). the resulting melted metal reaches a substrate to form a melt pool. by moving the laser head, it is possible to stack the tracks and build up a three - dimensional piece. 3d printing β sintering or melting amorphous powder metal in a 3d space to make any object to shape. cold - working processes, in which the product ' s shape is altered by rolling, fabrication or other processes, while the product is cold, can increase the strength of the product by a process called work hardening. work hardening creates microscopic defects in the metal, which resist further changes of shape. = = = heat treatment = = = metals can be heat - treated to alter the properties of strength, ductility, toughness, hardness and resistance to corrosion. common heat treatment processes include annealing, precipitation strengthening, quenching, and tempering : annealing process softens the metal by heating it and then allowing it to cool very slowly, which gets rid of stresses in the metal and makes the grain structure large and soft - edged so that, when the metal is hit or stressed it dents or perhaps bends, rather than breaking ; it is also easier to sand, grind, or cut annealed metal. quenching is the process of cooling metal very quickly after heating, thus " freezing " the metal ' s molecules in the very hard martensite form, which makes the metal harder. tempering relieves stresses in the metal that were caused by the hardening process ; tempering makes the metal less hard while making it better able to sustain
unversed in geometry enter here, " and also turned out many notable philosophers. plato ' s student aristotle introduced empiricism and the notion that universal truths can be arrived at via observation and induction, thereby laying the foundations of the scientific method. aristotle also produced many biological writings that were empirical in nature, focusing on biological causation and the diversity of life. he made countless observations of nature, especially the habits and attributes of plants and animals on lesbos, classified more than 540 animal species, and dissected at least 50. aristotle ' s writings profoundly influenced subsequent islamic and european scholarship, though they were eventually superseded in the scientific revolution. aristotle also contributed to theories of the elements and the cosmos. he believed that the celestial bodies ( such as the planets and the sun ) had something called an unmoved mover that put the celestial bodies in motion. aristotle tried to explain everything through mathematics and physics, but sometimes explained things such as the motion of celestial bodies through a higher power such as god. aristotle did not have the technological advancements that would have explained the motion of celestial bodies. in addition, aristotle had many views on the elements. he believed that everything was derived of the elements earth, water, air, fire, and lastly the aether. the aether was a celestial element, and therefore made up the matter of the celestial bodies. the elements of earth, water, air and fire were derived of a combination of two of the characteristics of hot, wet, cold, and dry, and all had their inevitable place and motion. the motion of these elements begins with earth being the closest to " the earth, " then water, air, fire, and finally aether. in addition to the makeup of all things, aristotle came up with theories as to why things did not return to their natural motion. he understood that water sits above earth, air above water, and fire above air in their natural state. he explained that although all elements must return to their natural state, the human body and other living things have a constraint on the elements β thus not allowing the elements making one who they are to return to their natural state. the important legacy of this period included substantial advances in factual knowledge, especially in anatomy, zoology, botany, mineralogy, geography, mathematics and astronomy ; an awareness of the importance of certain scientific problems, especially those related to the problem of change and its causes ; and a recognition of the methodological importance of applying mathematics to natural phenomena and of undertaking empirical research. in the hellenistic age scholars
are continuous lines used to depict edges directly visible from a particular angle. hidden β are short - dashed lines that may be used to represent edges that are not directly visible. center β are alternately long - and short - dashed lines that may be used to represent the axes of circular features. cutting plane β are thin, medium - dashed lines, or thick alternately long - and double short - dashed that may be used to define sections for section views. section β are thin lines in a pattern ( pattern determined by the material being " cut " or " sectioned " ) used to indicate surfaces in section views resulting from " cutting ". section lines are commonly referred to as " cross - hatching ". phantom β ( not shown ) are alternately long - and double short - dashed thin lines used to represent a feature or component that is not part of the specified part or assembly. e. g. billet ends that may be used for testing, or the machined product that is the focus of a tooling drawing. lines can also be classified by a letter classification in which each line is given a letter. type a lines show the outline of the feature of an object. they are the thickest lines on a drawing and done with a pencil softer than hb. type b lines are dimension lines and are used for dimensioning, projecting, extending, or leaders. a harder pencil should be used, such as a 2h pencil. type c lines are used for breaks when the whole object is not shown. these are freehand drawn and only for short breaks. 2h pencil type d lines are similar to type c, except these are zigzagged and only for longer breaks. 2h pencil type e lines indicate hidden outlines of internal features of an object. these are dotted lines. 2h pencil type f lines are type e lines, except these are used for drawings in electrotechnology. 2h pencil type g lines are used for centre lines. these are dotted lines, but a long line of 10 β 20 mm, then a 1 mm gap, then a small line of 2 mm. 2h pencil type h lines are the same as type g, except that every second long line is thicker. these indicate the cutting plane of an object. 2h pencil type k lines indicate the alternate positions of an object and the line taken by that object. these are drawn with a long line of 10 β 20 mm, then a small gap, then a small line of 2 mm, then a gap, then another small line. 2h
allow specialized cell lines to thrive in cultures replicating their native environments, but it also makes bioreactors attractive tools for culturing stem cells. a successful stem - cell - based bioreactor is effective at expanding stem cells with uniform properties and / or promoting controlled, reproducible differentiation into selected mature cell types. there are a variety of bioreactors designed for 3d cell cultures. there are small plastic cylindrical chambers, as well as glass chambers, with regulated internal humidity and moisture specifically engineered for the purpose of growing cells in three dimensions. the bioreactor uses bioactive synthetic materials such as polyethylene terephthalate membranes to surround the spheroid cells in an environment that maintains high levels of nutrients. they are easy to open and close, so that cell spheroids can be removed for testing, yet the chamber is able to maintain 100 % humidity throughout. this humidity is important to achieve maximum cell growth and function. the bioreactor chamber is part of a larger device that rotates to ensure equal cell growth in each direction across three dimensions. quinxell technologies now under quintech life sciences from singapore has developed a bioreactor known as the tisxell biaxial bioreactor which is specially designed for the purpose of tissue engineering. it is the first bioreactor in the world to have a spherical glass chamber with biaxial rotation ; specifically to mimic the rotation of the fetus in the womb ; which provides a conducive environment for the growth of tissues. multiple forms of mechanical stimulation have also been combined into a single bioreactor. using gene expression analysis, one academic study found that applying a combination of cyclic strain and ultrasound stimulation to pre - osteoblast cells in a bioreactor accelerated matrix maturation and differentiation. the technology of this combined stimulation bioreactor could be used to grow bone cells more quickly and effectively in future clinical stem cell therapies. mc2 biotek has also developed a bioreactor known as prototissue that uses gas exchange to maintain high oxygen levels within the cell chamber ; improving upon previous bioreactors, since the higher oxygen levels help the cell grow and undergo normal cell respiration. active areas of research on bioreactors includes increasing production scale and refining the physiological environment, both of which could improve the efficiency and efficacy of bioreactors in research or clinical use. bioreactors are currently used to study, among other things, cell and tissue level therapies, cell and tissue response to specific physiological
three of what is called the six simple machines, from which all machines are based. these machines are the inclined plane, the wedge, and the lever, which allowed the ancient egyptians to move millions of limestone blocks which weighed approximately 3. 5 tons ( 7, 000 lbs. ) each into place to create structures like the great pyramid of giza, which is 481 feet ( 147 meters ) high. they also made writing medium similar to paper from papyrus, which joshua mark states is the foundation for modern paper. papyrus is a plant ( cyperus papyrus ) which grew in plentiful amounts in the egyptian delta and throughout the nile river valley during ancient times. the papyrus was harvested by field workers and brought to processing centers where it was cut into thin strips. the strips were then laid - out side by side and covered in plant resin. the second layer of strips was laid on perpendicularly, then both pressed together until the sheet was dry. the sheets were then joined to form a roll and later used for writing. egyptian society made several significant advances during dynastic periods in many areas of technology. according to hossam elanzeery, they were the first civilization to use timekeeping devices such as sundials, shadow clocks, and obelisks and successfully leveraged their knowledge of astronomy to create a calendar model that society still uses today. they developed shipbuilding technology that saw them progress from papyrus reed vessels to cedar wood ships while also pioneering the use of rope trusses and stem - mounted rudders. the egyptians also used their knowledge of anatomy to lay the foundation for many modern medical techniques and practiced the earliest known version of neuroscience. elanzeery also states that they used and furthered mathematical science, as evidenced in the building of the pyramids. ancient egyptians also invented and pioneered many food technologies that have become the basis of modern food technology processes. based on paintings and reliefs found in tombs, as well as archaeological artifacts, scholars like paul t nicholson believe that the ancient egyptians established systematic farming practices, engaged in cereal processing, brewed beer and baked bread, processed meat, practiced viticulture and created the basis for modern wine production, and created condiments to complement, preserve and mask the flavors of their food. = = = = indus valley = = = = the indus valley civilization, situated in a resource - rich area ( in modern pakistan and northwestern india ), is notable for its early application of city planning, sanitation technologies, and plumbing. indus valley construction and architecture, called ' vaastu
Question: Sherry is going to use her computer to draw a three-dimensional box. She wants to be able to specify the exact dimensions and angles and to rotate the box as needed on the screen. Which of the following types of programs would Sherry most likely use for this purpose?
A) word processing software
B) math spreadsheet program
C) Internet-based search engine
D) computer-aided design program
|
D) computer-aided design program
|
Context:
consisting of several distinct layers, often referred to as spheres : the lithosphere, the hydrosphere, the atmosphere, and the biosphere, this concept of spheres is a useful tool for understanding the earth ' s surface and its various processes these correspond to rocks, water, air and life. also included by some are the cryosphere ( corresponding to ice ) as a distinct portion of the hydrosphere and the pedosphere ( corresponding to soil ) as an active and intermixed sphere. the following fields of science are generally categorized within the earth sciences : geology describes the rocky parts of the earth ' s crust ( or lithosphere ) and its historic development. major subdisciplines are mineralogy and petrology, geomorphology, paleontology, stratigraphy, structural geology, engineering geology, and sedimentology. physical geography focuses on geography as an earth science. physical geography is the study of earth ' s seasons, climate, atmosphere, soil, streams, landforms, and oceans. physical geography can be divided into several branches or related fields, as follows : geomorphology, biogeography, environmental geography, palaeogeography, climatology, meteorology, coastal geography, hydrology, ecology, glaciology. geophysics and geodesy investigate the shape of the earth, its reaction to forces and its magnetic and gravity fields. geophysicists explore the earth ' s core and mantle as well as the tectonic and seismic activity of the lithosphere. geophysics is commonly used to supplement the work of geologists in developing a comprehensive understanding of crustal geology, particularly in mineral and petroleum exploration. seismologists use geophysics to understand plate tectonic movement, as well as predict seismic activity. geochemistry studies the processes that control the abundance, composition, and distribution of chemical compounds and isotopes in geologic environments. geochemists use the tools and principles of chemistry to study the earth ' s composition, structure, processes, and other physical aspects. major subdisciplines are aqueous geochemistry, cosmochemistry, isotope geochemistry and biogeochemistry. soil science covers the outermost layer of the earth ' s crust that is subject to soil formation processes ( or pedosphere ). major subdivisions in this field of study include edaphology and pedology. ecology covers the interactions between organisms and their environment. this field of study differentiates the study of earth
earth. it emphasizes the study of how humans use and interact with freshwater supplies. study of water ' s movement is closely related to geomorphology and other branches of earth science. applied hydrology involves engineering to maintain aquatic environments and distribute water supplies. subdisciplines of hydrology include oceanography, hydrogeology, ecohydrology, and glaciology. oceanography is the study of oceans. hydrogeology is the study of groundwater. it includes the mapping of groundwater supplies and the analysis of groundwater contaminants. applied hydrogeology seeks to prevent contamination of groundwater and mineral springs and make it available as drinking water. the earliest exploitation of groundwater resources dates back to 3000 bc, and hydrogeology as a science was developed by hydrologists beginning in the 17th century. ecohydrology is the study of ecological systems in the hydrosphere. it can be divided into the physical study of aquatic ecosystems and the biological study of aquatic organisms. ecohydrology includes the effects that organisms and aquatic ecosystems have on one another as well as how these ecoystems are affected by humans. glaciology is the study of the cryosphere, including glaciers and coverage of the earth by ice and snow. concerns of glaciology include access to glacial freshwater, mitigation of glacial hazards, obtaining resources that exist beneath frozen land, and addressing the effects of climate change on the cryosphere. = = ecology = = ecology is the study of the biosphere. this includes the study of nature and of how living things interact with the earth and one another and the consequences of that. it considers how living things use resources such as oxygen, water, and nutrients from the earth to sustain themselves. it also considers how humans and other living creatures cause changes to nature. = = physical geography = = physical geography is the study of earth ' s systems and how they interact with one another as part of a single self - contained system. it incorporates astronomy, mathematical geography, meteorology, climatology, geology, geomorphology, biology, biogeography, pedology, and soils geography. physical geography is distinct from human geography, which studies the human populations on earth, though it does include human effects on the environment. = = methodology = = methodologies vary depending on the nature of the subjects being studied. studies typically fall into one of three categories : observational, experimental, or theoretical. earth scientists often conduct sophisticated computer analysis or visit an interesting location to study earth phenomena (
##sphere ( or lithosphere ). earth science can be considered to be a branch of planetary science but with a much older history. = = geology = = geology is broadly the study of earth ' s structure, substance, and processes. geology is largely the study of the lithosphere, or earth ' s surface, including the crust and rocks. it includes the physical characteristics and processes that occur in the lithosphere as well as how they are affected by geothermal energy. it incorporates aspects of chemistry, physics, and biology as elements of geology interact. historical geology is the application of geology to interpret earth history and how it has changed over time. geochemistry studies the chemical components and processes of the earth. geophysics studies the physical properties of the earth. paleontology studies fossilized biological material in the lithosphere. planetary geology studies geoscience as it pertains to extraterrestrial bodies. geomorphology studies the origin of landscapes. structural geology studies the deformation of rocks to produce mountains and lowlands. resource geology studies how energy resources can be obtained from minerals. environmental geology studies how pollution and contaminants affect soil and rock. mineralogy is the study of minerals and includes the study of mineral formation, crystal structure, hazards associated with minerals, and the physical and chemical properties of minerals. petrology is the study of rocks, including the formation and composition of rocks. petrography is a branch of petrology that studies the typology and classification of rocks. = = earth ' s interior = = plate tectonics, mountain ranges, volcanoes, and earthquakes are geological phenomena that can be explained in terms of physical and chemical processes in the earth ' s crust. beneath the earth ' s crust lies the mantle which is heated by the radioactive decay of heavy elements. the mantle is not quite solid and consists of magma which is in a state of semi - perpetual convection. this convection process causes the lithospheric plates to move, albeit slowly. the resulting process is known as plate tectonics. areas of the crust where new crust is created are called divergent boundaries, those where it is brought back into the earth are convergent boundaries and those where plates slide past each other, but no new lithospheric material is created or destroyed, are referred to as transform ( or conservative ) boundaries. earthquakes result from the movement of the lithospheric plates, and they often occur near convergent boundaries where parts of the crust are forced into the earth as
are the cryosphere ( corresponding to ice ) as a distinct portion of the hydrosphere and the pedosphere ( corresponding to soil ) as an active and intermixed sphere. the following fields of science are generally categorized within the earth sciences : geology describes the rocky parts of the earth ' s crust ( or lithosphere ) and its historic development. major subdisciplines are mineralogy and petrology, geomorphology, paleontology, stratigraphy, structural geology, engineering geology, and sedimentology. physical geography focuses on geography as an earth science. physical geography is the study of earth ' s seasons, climate, atmosphere, soil, streams, landforms, and oceans. physical geography can be divided into several branches or related fields, as follows : geomorphology, biogeography, environmental geography, palaeogeography, climatology, meteorology, coastal geography, hydrology, ecology, glaciology. geophysics and geodesy investigate the shape of the earth, its reaction to forces and its magnetic and gravity fields. geophysicists explore the earth ' s core and mantle as well as the tectonic and seismic activity of the lithosphere. geophysics is commonly used to supplement the work of geologists in developing a comprehensive understanding of crustal geology, particularly in mineral and petroleum exploration. seismologists use geophysics to understand plate tectonic movement, as well as predict seismic activity. geochemistry studies the processes that control the abundance, composition, and distribution of chemical compounds and isotopes in geologic environments. geochemists use the tools and principles of chemistry to study the earth ' s composition, structure, processes, and other physical aspects. major subdisciplines are aqueous geochemistry, cosmochemistry, isotope geochemistry and biogeochemistry. soil science covers the outermost layer of the earth ' s crust that is subject to soil formation processes ( or pedosphere ). major subdivisions in this field of study include edaphology and pedology. ecology covers the interactions between organisms and their environment. this field of study differentiates the study of earth from other planets in the solar system, earth being the only planet teeming with life. hydrology, oceanography and limnology are studies which focus on the movement, distribution, and quality of the water and involve all the components of the hydrologic cycle on the earth and its atmosphere ( or hydrosphere ). "
earth science or geoscience includes all fields of natural science related to the planet earth. this is a branch of science dealing with the physical, chemical, and biological complex constitutions and synergistic linkages of earth ' s four spheres : the biosphere, hydrosphere / cryosphere, atmosphere, and geosphere ( or lithosphere ). earth science can be considered to be a branch of planetary science but with a much older history. = = geology = = geology is broadly the study of earth ' s structure, substance, and processes. geology is largely the study of the lithosphere, or earth ' s surface, including the crust and rocks. it includes the physical characteristics and processes that occur in the lithosphere as well as how they are affected by geothermal energy. it incorporates aspects of chemistry, physics, and biology as elements of geology interact. historical geology is the application of geology to interpret earth history and how it has changed over time. geochemistry studies the chemical components and processes of the earth. geophysics studies the physical properties of the earth. paleontology studies fossilized biological material in the lithosphere. planetary geology studies geoscience as it pertains to extraterrestrial bodies. geomorphology studies the origin of landscapes. structural geology studies the deformation of rocks to produce mountains and lowlands. resource geology studies how energy resources can be obtained from minerals. environmental geology studies how pollution and contaminants affect soil and rock. mineralogy is the study of minerals and includes the study of mineral formation, crystal structure, hazards associated with minerals, and the physical and chemical properties of minerals. petrology is the study of rocks, including the formation and composition of rocks. petrography is a branch of petrology that studies the typology and classification of rocks. = = earth ' s interior = = plate tectonics, mountain ranges, volcanoes, and earthquakes are geological phenomena that can be explained in terms of physical and chemical processes in the earth ' s crust. beneath the earth ' s crust lies the mantle which is heated by the radioactive decay of heavy elements. the mantle is not quite solid and consists of magma which is in a state of semi - perpetual convection. this convection process causes the lithospheric plates to move, albeit slowly. the resulting process is known as plate tectonics. areas of the crust where new crust is created are called divergent boundaries, those where it is brought back into the earth are convergent boundaries and
three major planets, venus, earth, and mercury formed out of the solar nebula. a fourth planetesimal, theia, also formed near earth where it collided in a giant impact, rebounding as the planet mars. during this impact earth lost $ { \ approx } 4 $ \ % of its crust and mantle that is now is found on mars and the moon. at the antipode of the giant impact, $ \ approx $ 60 \ % of earth ' s crust, atmosphere, and a large amount of mantle were ejected into space forming the moon. the lost crust never reformed and became the earth ' s ocean basins. the theia impact site corresponds to indian ocean gravitational anomaly on earth and the hellas basin on mars. the dynamics of the giant impact are consistent with the rotational rates and axial tilts of both earth and mars. the giant impact removed sufficient co $ _ 2 $ from earth ' s atmosphere to avoid a runaway greenhouse effect, initiated plate tectonics, and gave life time to form near geothermal vents at the continental margins. mercury formed near venus where on a close approach it was slingshot into the sun ' s convective zone losing 94 \ % of its mass, much of which remains there today. black carbon, from co $ _ 2 $ decomposed by the intense heat, is still found on the surface of mercury. arriving at 616 km / s, mercury dramatically altered the sun ' s rotational energy, explaining both its anomalously slow rotation rate and axial tilt. these results are quantitatively supported by mass balances, the current locations of the terrestrial planets, and the orientations of their major orbital axes.
antipodes, that is to say, men on the opposite side of the earth, where the sun rises when it sets to us, men who walk with their feet opposite ours that is on no ground credible. and, indeed, it is not affirmed that this has been learned by historical knowledge, but by scientific conjecture, on the ground that the earth is suspended within the concavity of the sky, and that it has as much room on the one side of it as on the other : hence they say that the part that is beneath must also be inhabited. but they do not remark that, although it be supposed or scientifically demonstrated that the world is of a round and spherical form, yet it does not follow that the other side of the earth is bare of water ; nor even, though it be bare, does it immediately follow that it is peopled. for scripture, which proves the truth of its historical statements by the accomplishment of its prophecies, gives no false information ; and it is too absurd to say, that some men might have taken ship and traversed the whole wide ocean, and crossed from this side of the world to the other, and that thus even the inhabitants of that distant region are descended from that one first man. some historians do not view augustine ' s scriptural commentaries as endorsing any particular cosmological model, endorsing instead the view that augustine shared the common view of his contemporaries that the earth is spherical, in line with his endorsement of science in de genesi ad litteram. c. p. e. nothaft, responding to writers like leo ferrari who described augustine as endorsing a flat earth, says that "... other recent writers on the subject treat augustine ' s acceptance of the earth ' s spherical shape as a well - established fact ". while it always remained a minority view, from the mid - fourth to the seventh centuries ad, the flat - earth view experienced a revival, around the time when diodorus of tarsus founded the exegetical school known as the school of antioch, which sought to counter what he saw as the pagan cosmology of the greeks with a return to the traditional cosmology. the writings of diodorus did not survive, but are reconstructed from later criticism. this revival primarily took place in the east syriac world ( with little influence on the latin west ) where it gained proponents such as ephrem the syrian and in the popular hexaemeral homilies of jacob of serugh. chrys
##hosphere ) and its historic development. major subdisciplines are mineralogy and petrology, geomorphology, paleontology, stratigraphy, structural geology, engineering geology, and sedimentology. physical geography focuses on geography as an earth science. physical geography is the study of earth ' s seasons, climate, atmosphere, soil, streams, landforms, and oceans. physical geography can be divided into several branches or related fields, as follows : geomorphology, biogeography, environmental geography, palaeogeography, climatology, meteorology, coastal geography, hydrology, ecology, glaciology. geophysics and geodesy investigate the shape of the earth, its reaction to forces and its magnetic and gravity fields. geophysicists explore the earth ' s core and mantle as well as the tectonic and seismic activity of the lithosphere. geophysics is commonly used to supplement the work of geologists in developing a comprehensive understanding of crustal geology, particularly in mineral and petroleum exploration. seismologists use geophysics to understand plate tectonic movement, as well as predict seismic activity. geochemistry studies the processes that control the abundance, composition, and distribution of chemical compounds and isotopes in geologic environments. geochemists use the tools and principles of chemistry to study the earth ' s composition, structure, processes, and other physical aspects. major subdisciplines are aqueous geochemistry, cosmochemistry, isotope geochemistry and biogeochemistry. soil science covers the outermost layer of the earth ' s crust that is subject to soil formation processes ( or pedosphere ). major subdivisions in this field of study include edaphology and pedology. ecology covers the interactions between organisms and their environment. this field of study differentiates the study of earth from other planets in the solar system, earth being the only planet teeming with life. hydrology, oceanography and limnology are studies which focus on the movement, distribution, and quality of the water and involve all the components of the hydrologic cycle on the earth and its atmosphere ( or hydrosphere ). " sub - disciplines of hydrology include hydrometeorology, surface water hydrology, hydrogeology, watershed science, forest hydrology, and water chemistry. " glaciology covers the icy parts of the earth ( or cryosphere ). atmospheric sciences cover the gaseous parts of the earth ( or atmosphere
water content and the internal evolution of terrestrial planets and icy bodies are closely linked. the distribution of water in planetary systems is controlled by the temperature structure in the protoplanetary disk and dynamics and migration of planetesimals and planetary embryos. this results in the formation of planetesimals and planetary embryos with a great variety of compositions, water contents and degrees of oxidation. the internal evolution and especially the formation time of planetesimals relative to the timescale of radiogenic heating by short - lived 26al decay may govern the amount of hydrous silicates and leftover rock - ice mixtures available in the late stages of their evolution. in turn, water content may affect the early internal evolution of the planetesimals and in particular metal - silicate separation processes. moreover, water content may contribute to an increase of oxygen fugacity and thus affect the concentrations of siderophile elements within the silicate reservoirs of solar system objects. finally, the water content strongly influences the differentiation rate of the icy moons, controls their internal evolution and governs the alteration processes occurring in their deep interiors.
the scientific revolution. aristotle also contributed to theories of the elements and the cosmos. he believed that the celestial bodies ( such as the planets and the sun ) had something called an unmoved mover that put the celestial bodies in motion. aristotle tried to explain everything through mathematics and physics, but sometimes explained things such as the motion of celestial bodies through a higher power such as god. aristotle did not have the technological advancements that would have explained the motion of celestial bodies. in addition, aristotle had many views on the elements. he believed that everything was derived of the elements earth, water, air, fire, and lastly the aether. the aether was a celestial element, and therefore made up the matter of the celestial bodies. the elements of earth, water, air and fire were derived of a combination of two of the characteristics of hot, wet, cold, and dry, and all had their inevitable place and motion. the motion of these elements begins with earth being the closest to " the earth, " then water, air, fire, and finally aether. in addition to the makeup of all things, aristotle came up with theories as to why things did not return to their natural motion. he understood that water sits above earth, air above water, and fire above air in their natural state. he explained that although all elements must return to their natural state, the human body and other living things have a constraint on the elements β thus not allowing the elements making one who they are to return to their natural state. the important legacy of this period included substantial advances in factual knowledge, especially in anatomy, zoology, botany, mineralogy, geography, mathematics and astronomy ; an awareness of the importance of certain scientific problems, especially those related to the problem of change and its causes ; and a recognition of the methodological importance of applying mathematics to natural phenomena and of undertaking empirical research. in the hellenistic age scholars frequently employed the principles developed in earlier greek thought : the application of mathematics and deliberate empirical research, in their scientific investigations. thus, clear unbroken lines of influence lead from ancient greek and hellenistic philosophers, to medieval muslim philosophers and scientists, to the european renaissance and enlightenment, to the secular sciences of the modern day. neither reason nor inquiry began with the ancient greeks, but the socratic method did, along with the idea of forms, give great advances in geometry, logic, and the natural sciences. according to benjamin farrington, former professor of classics at swansea university : " men were weighing for thousands of years before archimedes worked out the
Question: Scientists who have studied the hydrosphere concluded that about 71% of Earth is covered with water. Which part of the hydrosphere is most often associated with the formation of organic rocks?
A) lakes
B) oceans
C) ponds
D) rivers
|
B) oceans
|
Context:
and nucleotides. carbohydrates include monomers and polymers of sugars. lipids are the only class of macromolecules that are not made up of polymers. they include steroids, phospholipids, and fats, largely nonpolar and hydrophobic ( water - repelling ) substances. proteins are the most diverse of the macromolecules. they include enzymes, transport proteins, large signaling molecules, antibodies, and structural proteins. the basic unit ( or monomer ) of a protein is an amino acid. twenty amino acids are used in proteins. nucleic acids are polymers of nucleotides. their function is to store, transmit, and express hereditary information. = = cells = = cell theory states that cells are the fundamental units of life, that all living things are composed of one or more cells, and that all cells arise from preexisting cells through cell division. most cells are very small, with diameters ranging from 1 to 100 micrometers and are therefore only visible under a light or electron microscope. there are generally two types of cells : eukaryotic cells, which contain a nucleus, and prokaryotic cells, which do not. prokaryotes are single - celled organisms such as bacteria, whereas eukaryotes can be single - celled or multicellular. in multicellular organisms, every cell in the organism ' s body is derived ultimately from a single cell in a fertilized egg. = = = cell structure = = = every cell is enclosed within a cell membrane that separates its cytoplasm from the extracellular space. a cell membrane consists of a lipid bilayer, including cholesterols that sit between phospholipids to maintain their fluidity at various temperatures. cell membranes are semipermeable, allowing small molecules such as oxygen, carbon dioxide, and water to pass through while restricting the movement of larger molecules and charged particles such as ions. cell membranes also contain membrane proteins, including integral membrane proteins that go across the membrane serving as membrane transporters, and peripheral proteins that loosely attach to the outer side of the cell membrane, acting as enzymes shaping the cell. cell membranes are involved in various cellular processes such as cell adhesion, storing electrical energy, and cell signalling and serve as the attachment surface for several extracellular structures such as a cell wall, glycocalyx, and cytoskeleton. within the cytoplasm of a cell
not always mean it is required, especially when dealing with genetic or functional redundancy. tracking experiments, which seek to gain information about the localisation and interaction of the desired protein. one way to do this is to replace the wild - type gene with a ' fusion ' gene, which is a juxtaposition of the wild - type gene with a reporting element such as green fluorescent protein ( gfp ) that will allow easy visualisation of the products of the genetic modification. while this is a useful technique, the manipulation can destroy the function of the gene, creating secondary effects and possibly calling into question the results of the experiment. more sophisticated techniques are now in development that can track protein products without mitigating their function, such as the addition of small sequences that will serve as binding motifs to monoclonal antibodies. expression studies aim to discover where and when specific proteins are produced. in these experiments, the dna sequence before the dna that codes for a protein, known as a gene ' s promoter, is reintroduced into an organism with the protein coding region replaced by a reporter gene such as gfp or an enzyme that catalyses the production of a dye. thus the time and place where a particular protein is produced can be observed. expression studies can be taken a step further by altering the promoter to find which pieces are crucial for the proper expression of the gene and are actually bound by transcription factor proteins ; this process is known as promoter bashing. = = = industrial = = = organisms can have their cells transformed with a gene coding for a useful protein, such as an enzyme, so that they will overexpress the desired protein. mass quantities of the protein can then be manufactured by growing the transformed organism in bioreactor equipment using industrial fermentation, and then purifying the protein. some genes do not work well in bacteria, so yeast, insect cells or mammalian cells can also be used. these techniques are used to produce medicines such as insulin, human growth hormone, and vaccines, supplements such as tryptophan, aid in the production of food ( chymosin in cheese making ) and fuels. other applications with genetically engineered bacteria could involve making them perform tasks outside their natural cycle, such as making biofuels, cleaning up oil spills, carbon and other toxic waste and detecting arsenic in drinking water. certain genetically modified microbes can also be used in biomining and bioremediation, due to their ability to extract heavy metals from their environment and incorporate them into compounds that are more easily recover
prominent functional groups that can be found in organisms : amino group, carboxyl group, carbonyl group, hydroxyl group, phosphate group, and sulfhydryl group. in 1953, the miller β urey experiment showed that organic compounds could be synthesized abiotically within a closed system mimicking the conditions of early earth, thus suggesting that complex organic molecules could have arisen spontaneously in early earth ( see abiogenesis ). = = = macromolecules = = = macromolecules are large molecules made up of smaller subunits or monomers. monomers include sugars, amino acids, and nucleotides. carbohydrates include monomers and polymers of sugars. lipids are the only class of macromolecules that are not made up of polymers. they include steroids, phospholipids, and fats, largely nonpolar and hydrophobic ( water - repelling ) substances. proteins are the most diverse of the macromolecules. they include enzymes, transport proteins, large signaling molecules, antibodies, and structural proteins. the basic unit ( or monomer ) of a protein is an amino acid. twenty amino acids are used in proteins. nucleic acids are polymers of nucleotides. their function is to store, transmit, and express hereditary information. = = cells = = cell theory states that cells are the fundamental units of life, that all living things are composed of one or more cells, and that all cells arise from preexisting cells through cell division. most cells are very small, with diameters ranging from 1 to 100 micrometers and are therefore only visible under a light or electron microscope. there are generally two types of cells : eukaryotic cells, which contain a nucleus, and prokaryotic cells, which do not. prokaryotes are single - celled organisms such as bacteria, whereas eukaryotes can be single - celled or multicellular. in multicellular organisms, every cell in the organism ' s body is derived ultimately from a single cell in a fertilized egg. = = = cell structure = = = every cell is enclosed within a cell membrane that separates its cytoplasm from the extracellular space. a cell membrane consists of a lipid bilayer, including cholesterols that sit between phospholipids to maintain their fluidity at various temperatures. cell membranes are semipermeable, allowing small molecules such as
water - repelling ) substances. proteins are the most diverse of the macromolecules. they include enzymes, transport proteins, large signaling molecules, antibodies, and structural proteins. the basic unit ( or monomer ) of a protein is an amino acid. twenty amino acids are used in proteins. nucleic acids are polymers of nucleotides. their function is to store, transmit, and express hereditary information. = = cells = = cell theory states that cells are the fundamental units of life, that all living things are composed of one or more cells, and that all cells arise from preexisting cells through cell division. most cells are very small, with diameters ranging from 1 to 100 micrometers and are therefore only visible under a light or electron microscope. there are generally two types of cells : eukaryotic cells, which contain a nucleus, and prokaryotic cells, which do not. prokaryotes are single - celled organisms such as bacteria, whereas eukaryotes can be single - celled or multicellular. in multicellular organisms, every cell in the organism ' s body is derived ultimately from a single cell in a fertilized egg. = = = cell structure = = = every cell is enclosed within a cell membrane that separates its cytoplasm from the extracellular space. a cell membrane consists of a lipid bilayer, including cholesterols that sit between phospholipids to maintain their fluidity at various temperatures. cell membranes are semipermeable, allowing small molecules such as oxygen, carbon dioxide, and water to pass through while restricting the movement of larger molecules and charged particles such as ions. cell membranes also contain membrane proteins, including integral membrane proteins that go across the membrane serving as membrane transporters, and peripheral proteins that loosely attach to the outer side of the cell membrane, acting as enzymes shaping the cell. cell membranes are involved in various cellular processes such as cell adhesion, storing electrical energy, and cell signalling and serve as the attachment surface for several extracellular structures such as a cell wall, glycocalyx, and cytoskeleton. within the cytoplasm of a cell, there are many biomolecules such as proteins and nucleic acids. in addition to biomolecules, eukaryotic cells have specialized structures called organelles that have their own lipid bilayers or are spatially units. these organelles include the cell nucleus, which contains most of
founded in 1976 and started the production of human proteins. genetically engineered human insulin was produced in 1978 and insulin - producing bacteria were commercialised in 1982. genetically modified food has been sold since 1994, with the release of the flavr savr tomato. the flavr savr was engineered to have a longer shelf life, but most current gm crops are modified to increase resistance to insects and herbicides. glofish, the first gmo designed as a pet, was sold in the united states in december 2003. in 2016 salmon modified with a growth hormone were sold. genetic engineering has been applied in numerous fields including research, medicine, industrial biotechnology and agriculture. in research, gmos are used to study gene function and expression through loss of function, gain of function, tracking and expression experiments. by knocking out genes responsible for certain conditions it is possible to create animal model organisms of human diseases. as well as producing hormones, vaccines and other drugs, genetic engineering has the potential to cure genetic diseases through gene therapy. chinese hamster ovary ( cho ) cells are used in industrial genetic engineering. additionally mrna vaccines are made through genetic engineering to prevent infections by viruses such as covid - 19. the same techniques that are used to produce drugs can also have industrial applications such as producing enzymes for laundry detergent, cheeses and other products. the rise of commercialised genetically modified crops has provided economic benefit to farmers in many different countries, but has also been the source of most of the controversy surrounding the technology. this has been present since its early use ; the first field trials were destroyed by anti - gm activists. although there is a scientific consensus that currently available food derived from gm crops poses no greater risk to human health than conventional food, critics consider gm food safety a leading concern. gene flow, impact on non - target organisms, control of the food supply and intellectual property rights have also been raised as potential issues. these concerns have led to the development of a regulatory framework, which started in 1975. it has led to an international treaty, the cartagena protocol on biosafety, that was adopted in 2000. individual countries have developed their own regulatory systems regarding gmos, with the most marked differences occurring between the united states and europe. = = overview = = genetic engineering is a process that alters the genetic structure of an organism by either removing or introducing dna, or modifying existing genetic material in situ. unlike traditional animal and plant breeding, which involves doing multiple crosses and then selecting for the organism with the desired phenotype,
mutation and thereby determine the role of particular genes. it is used especially frequently in developmental biology. when this is done by creating a library of genes with point mutations at every position in the area of interest, or even every position in the whole gene, this is called " scanning mutagenesis ". the simplest method, and the first to be used, is " alanine scanning ", where every position in turn is mutated to the unreactive amino acid alanine. gain of function experiments, the logical counterpart of knockouts. these are sometimes performed in conjunction with knockout experiments to more finely establish the function of the desired gene. the process is much the same as that in knockout engineering, except that the construct is designed to increase the function of the gene, usually by providing extra copies of the gene or inducing synthesis of the protein more frequently. gain of function is used to tell whether or not a protein is sufficient for a function, but does not always mean it is required, especially when dealing with genetic or functional redundancy. tracking experiments, which seek to gain information about the localisation and interaction of the desired protein. one way to do this is to replace the wild - type gene with a ' fusion ' gene, which is a juxtaposition of the wild - type gene with a reporting element such as green fluorescent protein ( gfp ) that will allow easy visualisation of the products of the genetic modification. while this is a useful technique, the manipulation can destroy the function of the gene, creating secondary effects and possibly calling into question the results of the experiment. more sophisticated techniques are now in development that can track protein products without mitigating their function, such as the addition of small sequences that will serve as binding motifs to monoclonal antibodies. expression studies aim to discover where and when specific proteins are produced. in these experiments, the dna sequence before the dna that codes for a protein, known as a gene ' s promoter, is reintroduced into an organism with the protein coding region replaced by a reporter gene such as gfp or an enzyme that catalyses the production of a dye. thus the time and place where a particular protein is produced can be observed. expression studies can be taken a step further by altering the promoter to find which pieces are crucial for the proper expression of the gene and are actually bound by transcription factor proteins ; this process is known as promoter bashing. = = = industrial = = = organisms can have their cells transformed with a gene coding for a useful protein, such as an enzyme,
defective gene with a functioning one. it is an important tool in research that allows the function of specific genes to be studied. drugs, vaccines and other products have been harvested from organisms engineered to produce them. crops have been developed that aid food security by increasing yield, nutritional value and tolerance to environmental stresses. the dna can be introduced directly into the host organism or into a cell that is then fused or hybridised with the host. this relies on recombinant nucleic acid techniques to form new combinations of heritable genetic material followed by the incorporation of that material either indirectly through a vector system or directly through micro - injection, macro - injection or micro - encapsulation. genetic engineering does not normally include traditional breeding, in vitro fertilisation, induction of polyploidy, mutagenesis and cell fusion techniques that do not use recombinant nucleic acids or a genetically modified organism in the process. however, some broad definitions of genetic engineering include selective breeding. cloning and stem cell research, although not considered genetic engineering, are closely related and genetic engineering can be used within them. synthetic biology is an emerging discipline that takes genetic engineering a step further by introducing artificially synthesised material into an organism. plants, animals or microorganisms that have been changed through genetic engineering are termed genetically modified organisms or gmos. if genetic material from another species is added to the host, the resulting organism is called transgenic. if genetic material from the same species or a species that can naturally breed with the host is used the resulting organism is called cisgenic. if genetic engineering is used to remove genetic material from the target organism the resulting organism is termed a knockout organism. in europe genetic modification is synonymous with genetic engineering while within the united states of america and canada genetic modification can also be used to refer to more conventional breeding methods. = = history = = humans have altered the genomes of species for thousands of years through selective breeding, or artificial selection : 1 : 1 as contrasted with natural selection. more recently, mutation breeding has used exposure to chemicals or radiation to produce a high frequency of random mutations, for selective breeding purposes. genetic engineering as the direct manipulation of dna by humans outside breeding and mutations has only existed since the 1970s. the term " genetic engineering " was coined by the russian - born geneticist nikolay timofeev - ressovsky in his 1934 paper " the experimental production of mutations ", published in the british journal biological reviews. jack williamson used the term in his science fiction novel dragon '
on a large scale provided protection from insect pests or tolerance to herbicides. fungal and virus resistant crops have also been developed or are in development. this makes the insect and weed management of crops easier and can indirectly increase crop yield. gm crops that directly improve yield by accelerating growth or making the plant more hardy ( by improving salt, cold or drought tolerance ) are also under development. in 2016 salmon have been genetically modified with growth hormones to reach normal adult size much faster. gmos have been developed that modify the quality of produce by increasing the nutritional value or providing more industrially useful qualities or quantities. the amflora potato produces a more industrially useful blend of starches. soybeans and canola have been genetically modified to produce more healthy oils. the first commercialised gm food was a tomato that had delayed ripening, increasing its shelf life. plants and animals have been engineered to produce materials they do not normally make. pharming uses crops and animals as bioreactors to produce vaccines, drug intermediates, or the drugs themselves ; the useful product is purified from the harvest and then used in the standard pharmaceutical production process. cows and goats have been engineered to express drugs and other proteins in their milk, and in 2009 the fda approved a drug produced in goat milk. = = = other applications = = = genetic engineering has potential applications in conservation and natural area management. gene transfer through viral vectors has been proposed as a means of controlling invasive species as well as vaccinating threatened fauna from disease. transgenic trees have been suggested as a way to confer resistance to pathogens in wild populations. with the increasing risks of maladaptation in organisms as a result of climate change and other perturbations, facilitated adaptation through gene tweaking could be one solution to reducing extinction risks. applications of genetic engineering in conservation are thus far mostly theoretical and have yet to be put into practice. genetic engineering is also being used to create microbial art. some bacteria have been genetically engineered to create black and white photographs. novelty items such as lavender - colored carnations, blue roses, and glowing fish, have also been produced through genetic engineering. = = regulation = = the regulation of genetic engineering concerns the approaches taken by governments to assess and manage the risks associated with the development and release of gmos. the development of a regulatory framework began in 1975, at asilomar, california. the asilomar meeting recommended a set of voluntary guidelines regarding the use of recombinant technology. as the technology improved
small ubiquitin - related modifier ( sumo ) proteins are widely expressed in eukaryotic cells, which are reversibly coupled to their substrates by motif recognition, called sumoylation. two interesting questions are 1 ) how many potential sumo substrates may be included in mammalian proteomes, such as human and mouse, 2 ) and given a sumo substrate, can we recognize its sumoylation sites? to answer these two questions, previous prediction systems of sumo substrates mainly adopted the pattern recognition methods, which could get high sensitivity with relatively too many potential false positives. so we use phylogenetic conservation between mouse and human to reduce the number of potential false positives.
phenotypic analysis. the new genetic material can be inserted randomly within the host genome or targeted to a specific location. the technique of gene targeting uses homologous recombination to make desired changes to a specific endogenous gene. this tends to occur at a relatively low frequency in plants and animals and generally requires the use of selectable markers. the frequency of gene targeting can be greatly enhanced through genome editing. genome editing uses artificially engineered nucleases that create specific double - stranded breaks at desired locations in the genome, and use the cell ' s endogenous mechanisms to repair the induced break by the natural processes of homologous recombination and nonhomologous end - joining. there are four families of engineered nucleases : meganucleases, zinc finger nucleases, transcription activator - like effector nucleases ( talens ), and the cas9 - guiderna system ( adapted from crispr ). talen and crispr are the two most commonly used and each has its own advantages. talens have greater target specificity, while crispr is easier to design and more efficient. in addition to enhancing gene targeting, engineered nucleases can be used to introduce mutations at endogenous genes that generate a gene knockout. = = applications = = genetic engineering has applications in medicine, research, industry and agriculture and can be used on a wide range of plants, animals and microorganisms. bacteria, the first organisms to be genetically modified, can have plasmid dna inserted containing new genes that code for medicines or enzymes that process food and other substrates. plants have been modified for insect protection, herbicide resistance, virus resistance, enhanced nutrition, tolerance to environmental pressures and the production of edible vaccines. most commercialised gmos are insect resistant or herbicide tolerant crop plants. genetically modified animals have been used for research, model animals and the production of agricultural or pharmaceutical products. the genetically modified animals include animals with genes knocked out, increased susceptibility to disease, hormones for extra growth and the ability to express proteins in their milk. = = = medicine = = = genetic engineering has many applications to medicine that include the manufacturing of drugs, creation of model animals that mimic human conditions and gene therapy. one of the earliest uses of genetic engineering was to mass - produce human insulin in bacteria. this application has now been applied to human growth hormones, follicle stimulating hormones ( for treating infertility ), human albumin,
Question: Which property of proteins best allows them to perform an enormously diverse range of functions?
A) their large, chemical-energy reserve
B) their large, stored-electric potential
C) the complexity of their folded structure
D) the wide variety of elements they include
|
C) the complexity of their folded structure
|
Context:
the project consists to determine, mathematically, the trajectory that will take an artificial satellite to fight against the air resistance. during our work, we had to consider that our satellite will crash to the surface of our planet. we started our study by understanding the system of forces that are acting between our satellite and the earth. in this work, we had to study the second law of newton by taking knowledge of the air friction, the speed of the satellite which helped us to find the equation that relates the trajectory of the satellite itself, its speed and the density of the air depending on the altitude. finally, we had to find a mathematic relation that links the density with the altitude and then we had to put it into our movement equation. in order to verify our model, we ' ll see what happens if we give a zero velocity to the satellite.
the group velocity of light has been measured at eight different wavelengths between 385 nm and 532 nm in the mediterranean sea at a depth of about 2. 2 km with the antares optical beacon systems. a parametrisation of the dependence of the refractive index on wavelength based on the salinity, pressure and temperature of the sea water at the antares site is in good agreement with these measurements.
are the stone - paved streets of the city - state of ur, dating to c. 4, 000 bce, and timber roads leading through the swamps of glastonbury, england, dating to around the same period. the first long - distance road, which came into use around 3, 500 bce, spanned 2, 400 km from the persian gulf to the mediterranean sea, but was not paved and was only partially maintained. in around 2, 000 bce, the minoans on the greek island of crete built a 50 km road leading from the palace of gortyn on the south side of the island, through the mountains, to the palace of knossos on the north side of the island. unlike the earlier road, the minoan road was completely paved. ancient minoan private homes had running water. a bathtub virtually identical to modern ones was unearthed at the palace of knossos. several minoan private homes also had toilets, which could be flushed by pouring water down the drain. the ancient romans had many public flush toilets, which emptied into an extensive sewage system. the primary sewer in rome was the cloaca maxima ; construction began on it in the sixth century bce and it is still in use today. the ancient romans also had a complex system of aqueducts, which were used to transport water across long distances. the first roman aqueduct was built in 312 bce. the eleventh and final ancient roman aqueduct was built in 226 ce. put together, the roman aqueducts extended over 450 km, but less than 70 km of this was above ground and supported by arches. = = = pre - modern = = = innovations continued through the middle ages with the introduction of silk production ( in asia and later europe ), the horse collar, and horseshoes. simple machines ( such as the lever, the screw, and the pulley ) were combined into more complicated tools, such as the wheelbarrow, windmills, and clocks. a system of universities developed and spread scientific ideas and practices, including oxford and cambridge. the renaissance era produced many innovations, including the introduction of the movable type printing press to europe, which facilitated the communication of knowledge. technology became increasingly influenced by science, beginning a cycle of mutual advancement. = = = modern = = = starting in the united kingdom in the 18th century, the discovery of steam power set off the industrial revolution, which saw wide - ranging technological discoveries, particularly in the areas of agriculture, manufacturing, mining, metallurgy, and transport, and the
, 400 km from the persian gulf to the mediterranean sea, but was not paved and was only partially maintained. in around 2, 000 bce, the minoans on the greek island of crete built a 50 km road leading from the palace of gortyn on the south side of the island, through the mountains, to the palace of knossos on the north side of the island. unlike the earlier road, the minoan road was completely paved. ancient minoan private homes had running water. a bathtub virtually identical to modern ones was unearthed at the palace of knossos. several minoan private homes also had toilets, which could be flushed by pouring water down the drain. the ancient romans had many public flush toilets, which emptied into an extensive sewage system. the primary sewer in rome was the cloaca maxima ; construction began on it in the sixth century bce and it is still in use today. the ancient romans also had a complex system of aqueducts, which were used to transport water across long distances. the first roman aqueduct was built in 312 bce. the eleventh and final ancient roman aqueduct was built in 226 ce. put together, the roman aqueducts extended over 450 km, but less than 70 km of this was above ground and supported by arches. = = = pre - modern = = = innovations continued through the middle ages with the introduction of silk production ( in asia and later europe ), the horse collar, and horseshoes. simple machines ( such as the lever, the screw, and the pulley ) were combined into more complicated tools, such as the wheelbarrow, windmills, and clocks. a system of universities developed and spread scientific ideas and practices, including oxford and cambridge. the renaissance era produced many innovations, including the introduction of the movable type printing press to europe, which facilitated the communication of knowledge. technology became increasingly influenced by science, beginning a cycle of mutual advancement. = = = modern = = = starting in the united kingdom in the 18th century, the discovery of steam power set off the industrial revolution, which saw wide - ranging technological discoveries, particularly in the areas of agriculture, manufacturing, mining, metallurgy, and transport, and the widespread application of the factory system. this was followed a century later by the second industrial revolution which led to rapid scientific discovery, standardization, and mass production. new technologies were developed, including sewage systems, electricity, light bulbs, electric motors, railroads, automobiles, and airplanes. these technological advances led to significant developments in medicine
earth. each satellite has an onboard atomic clock and transmits a continuous radio signal containing a precise time signal as well as its current position. two frequencies are used, 1. 2276 and 1. 57542 ghz. since the velocity of radio waves is virtually constant, the delay of the radio signal from a satellite is proportional to the distance of the receiver from the satellite. by receiving the signals from at least four satellites a gps receiver can calculate its position on earth by comparing the arrival time of the radio signals. since each satellite ' s position is known precisely at any given time, from the delay the position of the receiver can be calculated by a microprocessor in the receiver. the position can be displayed as latitude and longitude, or as a marker on an electronic map. gps receivers are incorporated in almost all cellphones and in vehicles such as automobiles, aircraft, and ships, and are used to guide drones, missiles, cruise missiles, and even artillery shells to their target, and handheld gps receivers are produced for hikers and the military. radio beacon β a fixed location terrestrial radio transmitter which transmits a continuous radio signal used by aircraft and ships for navigation. the locations of beacons are plotted on navigational maps used by aircraft and ships. vhf omnidirectional range ( vor ) β a worldwide aircraft radio navigation system consisting of fixed ground radio beacons transmitting between 108. 00 and 117. 95 mhz in the very high frequency ( vhf ) band. an automated navigational instrument on the aircraft displays a bearing to a nearby vor transmitter. a vor beacon transmits two signals simultaneously on different frequencies. a directional antenna transmits a beam of radio waves that rotates like a lighthouse at a fixed rate, 30 times per second. when the directional beam is facing north, an omnidirectional antenna transmits a pulse. by measuring the difference in phase of these two signals, an aircraft can determine its bearing ( or " radial " ) from the station accurately. by taking a bearing on two vor beacons an aircraft can determine its position ( called a " fix " ) to an accuracy of about 90 metres ( 300 ft ). most vor beacons also have a distance measuring capability, called distance measuring equipment ( dme ) ; these are called vor / dme ' s. the aircraft transmits a radio signal to the vor / dme beacon and a transponder transmits a return signal. from the propagation delay between the transmitted and received signal the aircraft can calculate
large scale manned space flight within the solar system is still confronted with the solution of two problems : 1. a propulsion system to transport large payloads with short transit times between different planetary orbits. 2. a cost effective lifting of large payloads into earth orbit. for the solution of the first problem a deuterium fusion bomb propulsion system is proposed where a thermonuclear detonation wave is ignited in a small cylindrical assembly of deuterium with a gigavolt - multimegampere proton beam, drawn from the magnetically insulated spacecraft acting in the ultrahigh vacuum of space as a gigavolt capacitor. for the solution of the second problem, the ignition is done by argon ion lasers driven by high explosives, with the lasers destroyed in the fusion explosion and becoming part of the exhaust.
if $ g $ is an omega - stable group with a normal definable subgroup $ h $, then the sylow - $ 2 $ - subgroups of $ g / h $ are the images of the sylow - $ 2 $ - subgroups of $ g $.
time interval between the incident and scattered photon in raman effect and absorption of photon and emission of electron in photoelectric effect has not been determined till now. this is because there is no such high level instrument discovered till now to detect time interval to such a small level. but this can be calculated theoretically by applying a basic principle of physics like impulse is equal to the change in momentum. considering the collision between electron and photon as perfect inelastic collision in photoelectric effect, elastic and inelastic collision in raman effect and elastic collision in plane mirror reflection and the interaction between electron and photon as strong gravitational interaction we calculate the required time interval. during these phenomena there is lattice vibration which can be quantized as phonon particles.
there are no limits for the speeds of light and particles in general relativity ( gr ). four examples illustrate this basic result, which is too often neglected.
metres ) by small portable navigation instruments, by timing the arrival of radio signals from the satellites. these are the most widely used navigation systems today. the main satellite navigation systems are the us global positioning system ( gps ), russia ' s glonass, china ' s beidou navigation satellite system ( bds ) and the european union ' s galileo. global positioning system ( gps ) β the most widely used satellite navigation system, maintained by the us air force, which uses a constellation of 31 satellites in low earth orbit. the orbits of the satellites are distributed so at any time at least four satellites are above the horizon over each point on earth. each satellite has an onboard atomic clock and transmits a continuous radio signal containing a precise time signal as well as its current position. two frequencies are used, 1. 2276 and 1. 57542 ghz. since the velocity of radio waves is virtually constant, the delay of the radio signal from a satellite is proportional to the distance of the receiver from the satellite. by receiving the signals from at least four satellites a gps receiver can calculate its position on earth by comparing the arrival time of the radio signals. since each satellite ' s position is known precisely at any given time, from the delay the position of the receiver can be calculated by a microprocessor in the receiver. the position can be displayed as latitude and longitude, or as a marker on an electronic map. gps receivers are incorporated in almost all cellphones and in vehicles such as automobiles, aircraft, and ships, and are used to guide drones, missiles, cruise missiles, and even artillery shells to their target, and handheld gps receivers are produced for hikers and the military. radio beacon β a fixed location terrestrial radio transmitter which transmits a continuous radio signal used by aircraft and ships for navigation. the locations of beacons are plotted on navigational maps used by aircraft and ships. vhf omnidirectional range ( vor ) β a worldwide aircraft radio navigation system consisting of fixed ground radio beacons transmitting between 108. 00 and 117. 95 mhz in the very high frequency ( vhf ) band. an automated navigational instrument on the aircraft displays a bearing to a nearby vor transmitter. a vor beacon transmits two signals simultaneously on different frequencies. a directional antenna transmits a beam of radio waves that rotates like a lighthouse at a fixed rate, 30 times per second. when the directional beam is facing north, an omnidirectional antenna transmits a pulse. by measuring the difference in phase of
Question: The mathematical model for calculating speed is shown below. Speed = distance/time. An airplane traveled 700 kilometers in two hours during a trip. What was the average speed of the plane during the trip?
A) 5.8 kilometers per hour
B) 350 kilometers per hour
C) 1400 kilometers per hour
D) 84,000 kilometers per hour
|
B) 350 kilometers per hour
|
Context:
aquatic and most of the aquatic photosynthetic eukaryotic organisms are collectively described as algae, which is a term of convenience as not all algae are closely related. algae comprise several distinct clades such as glaucophytes, which are microscopic freshwater algae that may have resembled in form to the early unicellular ancestor of plantae. unlike glaucophytes, the other algal clades such as red and green algae are multicellular. green algae comprise three major clades : chlorophytes, coleochaetophytes, and stoneworts. fungi are eukaryotes that digest foods outside their bodies, secreting digestive enzymes that break down large food molecules before absorbing them through their cell membranes. many fungi are also saprobes, feeding on dead organic matter, making them important decomposers in ecological systems. animals are multicellular eukaryotes. with few exceptions, animals consume organic material, breathe oxygen, are able to move, can reproduce sexually, and grow from a hollow sphere of cells, the blastula, during embryonic development. over 1. 5 million living animal species have been described β of which around 1 million are insects β but it has been estimated there are over 7 million animal species in total. they have complex interactions with each other and their environments, forming intricate food webs. = = = viruses = = = viruses are submicroscopic infectious agents that replicate inside the cells of organisms. viruses infect all types of life forms, from animals and plants to microorganisms, including bacteria and archaea. more than 6, 000 virus species have been described in detail. viruses are found in almost every ecosystem on earth and are the most numerous type of biological entity. the origins of viruses in the evolutionary history of life are unclear : some may have evolved from plasmids β pieces of dna that can move between cells β while others may have evolved from bacteria. in evolution, viruses are an important means of horizontal gene transfer, which increases genetic diversity in a way analogous to sexual reproduction. because viruses possess some but not all characteristics of life, they have been described as " organisms at the edge of life ", and as self - replicators. = = ecology = = ecology is the study of the distribution and abundance of life, the interaction between organisms and their environment. = = = ecosystems = = = the community of living ( biotic ) organisms in conjunction with the nonliving ( abiotic ) components ( e.
cortisol, corticosterone and aldosterone activate full - length glucocorticoid receptor ( gr ) from elephant shark, a cartilaginous fish belonging to the oldest group of jawed vertebrates. activation by aldosterone a mineralocorticoid, indicates partial divergence of elephant shark gr from the mr. progesterone activates elephant shark mr, but not elephant shark gr. progesterone inhibits steroid binding to elephant shark gr, but not to human gr. deletion of the n - terminal domain ( ntd ) from elephant shark gr ( truncated gr ) reduced the response to corticosteroids, while truncated and full - length elephant shark mr had similar responses to corticosteroids. chimeras of elephant shark gr ntd fused to mr dbd + lbd had increased activation by corticosteroids and progesterone compared to full - length elephant shark mr. elephant shark mr ntd fused to gr dbd + lbd had similar activation as full - length elephant shark mr, indicating that activation of human gr by the ntd evolved early in gr divergence from the mr.
armed with an astrolabe and kepler ' s laws one can arrive at accurate estimates of the orbits of planets.
a 4mj planet with a 15. 8day orbital period has been detected from very precise radial velocity measurements with the coralie echelle spectrograph. a second remote and more massive companion has also been detected. all the planetary companions so far detected in orbit closer than 0. 08 au have a parent star with a statistically higher metal content compared to the metallicity distribution of other stars with planets. different processes occuring during their formation may provide a possible explanation for this observation.
, social and economic status, habits ( including diet, medications, tobacco, alcohol ). the physical examination is the examination of the patient for medical signs of disease that are objective and observable, in contrast to symptoms that are volunteered by the patient and are not necessarily objectively observable. the healthcare provider uses sight, hearing, touch, and sometimes smell ( e. g., in infection, uremia, diabetic ketoacidosis ). four actions are the basis of physical examination : inspection, palpation ( feel ), percussion ( tap to determine resonance characteristics ), and auscultation ( listen ), generally in that order, although auscultation occurs prior to percussion and palpation for abdominal assessments. the clinical examination involves the study of : abdomen and rectum cardiovascular ( heart and blood vessels ) general appearance of the patient and specific indicators of disease ( nutritional status, presence of jaundice, pallor or clubbing ) genitalia ( and pregnancy if the patient is or could be pregnant ) head, eye, ear, nose, and throat ( heent ) musculoskeletal ( including spine and extremities ) neurological ( consciousness, awareness, brain, vision, cranial nerves, spinal cord and peripheral nerves ) psychiatric ( orientation, mental state, mood, evidence of abnormal perception or thought ). respiratory ( large airways and lungs ) skin vital signs including height, weight, body temperature, blood pressure, pulse, respiration rate, and hemoglobin oxygen saturation it is to likely focus on areas of interest highlighted in the medical history and may not include everything listed above. the treatment plan may include ordering additional medical laboratory tests and medical imaging studies, starting therapy, referral to a specialist, or watchful observation. a follow - up may be advised. depending upon the health insurance plan and the managed care system, various forms of " utilization review ", such as prior authorization of tests, may place barriers on accessing expensive services. the medical decision - making ( mdm ) process includes the analysis and synthesis of all the above data to come up with a list of possible diagnoses ( the differential diagnoses ), along with an idea of what needs to be done to obtain a definitive diagnosis that would explain the patient ' s problem. on subsequent visits, the process may be repeated in an abbreviated manner to obtain any new history, symptoms, physical findings, lab or imaging results, or specialist consultations. = = institutions = = contemporary
eat them. plants and other photosynthetic organisms are at the base of most food chains because they use the energy from the sun and nutrients from the soil and atmosphere, converting them into a form that can be used by animals. this is what ecologists call the first trophic level. the modern forms of the major staple foods, such as hemp, teff, maize, rice, wheat and other cereal grasses, pulses, bananas and plantains, as well as hemp, flax and cotton grown for their fibres, are the outcome of prehistoric selection over thousands of years from among wild ancestral plants with the most desirable characteristics. botanists study how plants produce food and how to increase yields, for example through plant breeding, making their work important to humanity ' s ability to feed the world and provide food security for future generations. botanists also study weeds, which are a considerable problem in agriculture, and the biology and control of plant pathogens in agriculture and natural ecosystems. ethnobotany is the study of the relationships between plants and people. when applied to the investigation of historical plant β people relationships ethnobotany may be referred to as archaeobotany or palaeoethnobotany. some of the earliest plant - people relationships arose between the indigenous people of canada in identifying edible plants from inedible plants. this relationship the indigenous people had with plants was recorded by ethnobotanists. = = plant biochemistry = = plant biochemistry is the study of the chemical processes used by plants. some of these processes are used in their primary metabolism like the photosynthetic calvin cycle and crassulacean acid metabolism. others make specialised materials like the cellulose and lignin used to build their bodies, and secondary products like resins and aroma compounds. plants and various other groups of photosynthetic eukaryotes collectively known as " algae " have unique organelles known as chloroplasts. chloroplasts are thought to be descended from cyanobacteria that formed endosymbiotic relationships with ancient plant and algal ancestors. chloroplasts and cyanobacteria contain the blue - green pigment chlorophyll a. chlorophyll a ( as well as its plant and green algal - specific cousin chlorophyll b ) absorbs light in the blue - violet and orange / red parts of the spectrum while reflecting and transmitting the green light that we see as the characteristic colour
a detailed analysis of primordial nucleosynthesis predictions for light element abundances is performed. contents : 1. the standard cosmology : an overview. 2. primordial nucleosynthesis. 3. the born rates for n < - > p reactions. 4. finite nucleon mass corrections. 5. qed thermal radiative corrections. 6. calculations of big bang nucleosynthesis. results.
which could be used as tools, primarily in the form of choppers or scrapers. these tools greatly aided the early humans in their hunter - gatherer lifestyle to perform a variety of tasks including butchering carcasses ( and breaking bones to get at the marrow ) ; chopping wood ; cracking open nuts ; skinning an animal for its hide, and even forming other tools out of softer materials such as bone and wood. the earliest stone tools were irrelevant, being little more than a fractured rock. in the acheulian era, beginning approximately 1. 65 million years ago, methods of working these stones into specific shapes, such as hand axes emerged. this early stone age is described as the lower paleolithic. the middle paleolithic, approximately 300, 000 years ago, saw the introduction of the prepared - core technique, where multiple blades could be rapidly formed from a single core stone. the upper paleolithic, beginning approximately 40, 000 years ago, saw the introduction of pressure flaking, where a wood, bone, or antler punch could be used to shape a stone very finely. the end of the last ice age about 10, 000 years ago is taken as the end point of the upper paleolithic and the beginning of the epipaleolithic / mesolithic. the mesolithic technology included the use of microliths as composite stone tools, along with wood, bone, and antler tools. the later stone age, during which the rudiments of agricultural technology were developed, is called the neolithic period. during this period, polished stone tools were made from a variety of hard rocks such as flint, jade, jadeite, and greenstone, largely by working exposures as quarries, but later the valuable rocks were pursued by tunneling underground, the first steps in mining technology. the polished axes were used for forest clearance and the establishment of crop farming and were so effective as to remain in use when bronze and iron appeared. these stone axes were used alongside a continued use of stone tools such as a range of projectiles, knives, and scrapers, as well as tools, made from organic materials such as wood, bone, and antler. stone age cultures developed music and engaged in organized warfare. stone age humans developed ocean - worthy outrigger canoe technology, leading to migration across the malay archipelago, across the indian ocean to madagascar and also across the pacific ocean, which required knowledge of the ocean currents, weather patterns, sailing, and celestial navigation. although paleolithic cultures
earth. it emphasizes the study of how humans use and interact with freshwater supplies. study of water ' s movement is closely related to geomorphology and other branches of earth science. applied hydrology involves engineering to maintain aquatic environments and distribute water supplies. subdisciplines of hydrology include oceanography, hydrogeology, ecohydrology, and glaciology. oceanography is the study of oceans. hydrogeology is the study of groundwater. it includes the mapping of groundwater supplies and the analysis of groundwater contaminants. applied hydrogeology seeks to prevent contamination of groundwater and mineral springs and make it available as drinking water. the earliest exploitation of groundwater resources dates back to 3000 bc, and hydrogeology as a science was developed by hydrologists beginning in the 17th century. ecohydrology is the study of ecological systems in the hydrosphere. it can be divided into the physical study of aquatic ecosystems and the biological study of aquatic organisms. ecohydrology includes the effects that organisms and aquatic ecosystems have on one another as well as how these ecoystems are affected by humans. glaciology is the study of the cryosphere, including glaciers and coverage of the earth by ice and snow. concerns of glaciology include access to glacial freshwater, mitigation of glacial hazards, obtaining resources that exist beneath frozen land, and addressing the effects of climate change on the cryosphere. = = ecology = = ecology is the study of the biosphere. this includes the study of nature and of how living things interact with the earth and one another and the consequences of that. it considers how living things use resources such as oxygen, water, and nutrients from the earth to sustain themselves. it also considers how humans and other living creatures cause changes to nature. = = physical geography = = physical geography is the study of earth ' s systems and how they interact with one another as part of a single self - contained system. it incorporates astronomy, mathematical geography, meteorology, climatology, geology, geomorphology, biology, biogeography, pedology, and soils geography. physical geography is distinct from human geography, which studies the human populations on earth, though it does include human effects on the environment. = = methodology = = methodologies vary depending on the nature of the subjects being studied. studies typically fall into one of three categories : observational, experimental, or theoretical. earth scientists often conduct sophisticated computer analysis or visit an interesting location to study earth phenomena (
the group velocity of light has been measured at eight different wavelengths between 385 nm and 532 nm in the mediterranean sea at a depth of about 2. 2 km with the antares optical beacon systems. a parametrisation of the dependence of the refractive index on wavelength based on the salinity, pressure and temperature of the sea water at the antares site is in good agreement with these measurements.
Question: What is the best way for a marine scientist to determine the eating habits of sharks?
A) Tag one shark and monitor its eating habits for one day.
B) Tag one shark and monitor its eating habits for one year.
C) Tag several sharks and monitor their eating habits for one day.
D) Tag several sharks and monitor their eating habits for one year.
|
D) Tag several sharks and monitor their eating habits for one year.
|
Context:
three of what is called the six simple machines, from which all machines are based. these machines are the inclined plane, the wedge, and the lever, which allowed the ancient egyptians to move millions of limestone blocks which weighed approximately 3. 5 tons ( 7, 000 lbs. ) each into place to create structures like the great pyramid of giza, which is 481 feet ( 147 meters ) high. they also made writing medium similar to paper from papyrus, which joshua mark states is the foundation for modern paper. papyrus is a plant ( cyperus papyrus ) which grew in plentiful amounts in the egyptian delta and throughout the nile river valley during ancient times. the papyrus was harvested by field workers and brought to processing centers where it was cut into thin strips. the strips were then laid - out side by side and covered in plant resin. the second layer of strips was laid on perpendicularly, then both pressed together until the sheet was dry. the sheets were then joined to form a roll and later used for writing. egyptian society made several significant advances during dynastic periods in many areas of technology. according to hossam elanzeery, they were the first civilization to use timekeeping devices such as sundials, shadow clocks, and obelisks and successfully leveraged their knowledge of astronomy to create a calendar model that society still uses today. they developed shipbuilding technology that saw them progress from papyrus reed vessels to cedar wood ships while also pioneering the use of rope trusses and stem - mounted rudders. the egyptians also used their knowledge of anatomy to lay the foundation for many modern medical techniques and practiced the earliest known version of neuroscience. elanzeery also states that they used and furthered mathematical science, as evidenced in the building of the pyramids. ancient egyptians also invented and pioneered many food technologies that have become the basis of modern food technology processes. based on paintings and reliefs found in tombs, as well as archaeological artifacts, scholars like paul t nicholson believe that the ancient egyptians established systematic farming practices, engaged in cereal processing, brewed beer and baked bread, processed meat, practiced viticulture and created the basis for modern wine production, and created condiments to complement, preserve and mask the flavors of their food. = = = = indus valley = = = = the indus valley civilization, situated in a resource - rich area ( in modern pakistan and northwestern india ), is notable for its early application of city planning, sanitation technologies, and plumbing. indus valley construction and architecture, called ' vaastu
great pyramid of giza, which is 481 feet ( 147 meters ) high. they also made writing medium similar to paper from papyrus, which joshua mark states is the foundation for modern paper. papyrus is a plant ( cyperus papyrus ) which grew in plentiful amounts in the egyptian delta and throughout the nile river valley during ancient times. the papyrus was harvested by field workers and brought to processing centers where it was cut into thin strips. the strips were then laid - out side by side and covered in plant resin. the second layer of strips was laid on perpendicularly, then both pressed together until the sheet was dry. the sheets were then joined to form a roll and later used for writing. egyptian society made several significant advances during dynastic periods in many areas of technology. according to hossam elanzeery, they were the first civilization to use timekeeping devices such as sundials, shadow clocks, and obelisks and successfully leveraged their knowledge of astronomy to create a calendar model that society still uses today. they developed shipbuilding technology that saw them progress from papyrus reed vessels to cedar wood ships while also pioneering the use of rope trusses and stem - mounted rudders. the egyptians also used their knowledge of anatomy to lay the foundation for many modern medical techniques and practiced the earliest known version of neuroscience. elanzeery also states that they used and furthered mathematical science, as evidenced in the building of the pyramids. ancient egyptians also invented and pioneered many food technologies that have become the basis of modern food technology processes. based on paintings and reliefs found in tombs, as well as archaeological artifacts, scholars like paul t nicholson believe that the ancient egyptians established systematic farming practices, engaged in cereal processing, brewed beer and baked bread, processed meat, practiced viticulture and created the basis for modern wine production, and created condiments to complement, preserve and mask the flavors of their food. = = = = indus valley = = = = the indus valley civilization, situated in a resource - rich area ( in modern pakistan and northwestern india ), is notable for its early application of city planning, sanitation technologies, and plumbing. indus valley construction and architecture, called ' vaastu shastra ', suggests a thorough understanding of materials engineering, hydrology, and sanitation. = = = = china = = = = the chinese made many first - known discoveries and developments. major technological contributions from china include the earliest known form of the binary code and epigenetic sequencing, early seismological detectors,
if wood has been with us since time immemorial, being part of our environment, housing and tools, now wood has gain momentum, as it is clear that wood improves our life style. because of the healthiness, resistance, ecology and comfort, wood is important for all of us, no matter what our life style is. woodtouch project aims to open a completely new market for furniture and interior design companies, enabling touch interaction between the user and wooden furniture surfaces. why not switch on or dim the lights touching a wooden table? why not turn on the heating system? why not use wood as a touch sensitive surface for domotic control? the furniture designed with this novel technology, offers a wooden outer image and has different touch sensitive areas over the ones the user is able to control all sorts of electric appliances touching over a wooden surface.
properties and the desired end effect. the relation between yield stress and grain size is described mathematically by the hall - petch equation which is Ο y = Ο 0 + k y d { \ displaystyle \ sigma _ { y } = \ sigma _ { 0 } + { k _ { y } \ over { \ sqrt { d } } } } where ky is the strengthening coefficient ( a constant unique to each material ), Οo is a materials constant for the starting stress for dislocation movement ( or the resistance of the lattice to dislocation motion ), d is the grain diameter, and Οy is the yield stress. theoretically, a material could be made infinitely strong if the grains are made infinitely small. this is, unfortunately, impossible because the lower limit of grain size is a single unit cell of the material. even then, if the grains of a material are the size of a single unit cell, then the material is in fact amorphous, not crystalline, since there is no long range order, and dislocations can not be defined in an amorphous material. it has been observed experimentally that the microstructure with the highest yield strength is a grain size of about 10 nanometers, because grains smaller than this undergo another yielding mechanism, grain boundary sliding. producing engineering materials with this ideal grain size is difficult because of the limitations of initial particle sizes inherent to nanomaterials and nanotechnology. = = faber - evans model = = the faber - evans model, developed by katherine faber and anthony g. evans, was developed to predict the increase in fracture toughness in ceramics due to crack deflection around second - phase particles that are prone to microcracking in a matrix. the model considers particle morphology, aspect ratio, spacing, and volume fraction of the second phase, as well as the reduction in local stress intensity at the crack tip when the crack is deflected or the crack plane bows. actual crack tortuosity is obtained through imaging techniques, which allows for the direct input of deflection and bowing angles into the model. the model calculates the average strain energy release rate and compares the resulting increase in fracture toughness to that of a flat crack through the plain matrix. the magnitude of the toughening is determined by the mismatch strain caused by thermal contraction incompatibility and the microfracture resistance of the particle / matrix interface. the toughening becomes noticeable with a narrow size distribution of appropriately sized particles, and researchers typically
##ization. the fouling is usually induced faster at higher driving forces. membrane fouling and particle retention in a feed solution also builds up a concentration gradients and particle backflow ( concentration polarization ). the tangential flow devices are more cost and labor - intensive, but they are less susceptible to fouling due to the sweeping effects and high shear rates of the passing flow. the most commonly used synthetic membrane devices ( modules ) are flat sheets / plates, spiral wounds, and hollow fibers. flat membranes used in filtration and separation processes can be enhanced with surface patterning, where microscopic structures are introduced to improve performance. these patterns increase surface area, optimize water flow, and reduce fouling, leading to higher permeability and longer membrane lifespan. research has shown that such modifications can significantly enhance efficiency in water purification, energy applications, and industrial separations. flat plates are usually constructed as circular thin flat membrane surfaces to be used in dead - end geometry modules. spiral wounds are constructed from similar flat membranes but in the form of a " pocket " containing two membrane sheets separated by a highly porous support plate. several such pockets are then wound around a tube to create a tangential flow geometry and to reduce membrane fouling. hollow fiber modules consist of an assembly of self - supporting fibers with dense skin separation layers, and a more open matrix helping to withstand pressure gradients and maintain structural integrity. the hollow fiber modules can contain up to 10, 000 fibers ranging from 200 to 2500 ΞΌm in diameter ; the main advantage of hollow fiber modules is the very large surface area within an enclosed volume, increasing the efficiency of the separation process. the disc tube module uses a cross - flow geometry and consists of a pressure tube and hydraulic discs, which are held by a central tension rod, and membrane cushions that lie between two discs. = = membrane performance and governing equations = = the selection of synthetic membranes for a targeted separation process is usually based on few requirements. membranes have to provide enough mass transfer area to process large amounts of feed stream. the selected membrane has to have high selectivity ( rejection ) properties for certain particles ; it has to resist fouling and to have high mechanical stability. it also needs to be reproducible and to have low manufacturing costs. the main modeling equation for the dead - end filtration at constant pressure drop is represented by darcy ' s law : d v p d t = q = Ξ΄ p ΞΌ a ( 1 r m + r ) { \ displaystyle { \ frac { dv _
during their exposure, nuclear emulsion sheets detect both tracks from experiment - related particles, as well as a considerable amount of background tracks, mainly due to cosmic rays. unless the exposure has been fairly short, it is therefore fairly likely that a fraction of the tracks that have been identified as belonging to the particles the experiment is interested in, are really due to background. a method, which allows to measure this fraction reliably directly from the data, is described.
polished stone tools were made from a variety of hard rocks such as flint, jade, jadeite, and greenstone, largely by working exposures as quarries, but later the valuable rocks were pursued by tunneling underground, the first steps in mining technology. the polished axes were used for forest clearance and the establishment of crop farming and were so effective as to remain in use when bronze and iron appeared. these stone axes were used alongside a continued use of stone tools such as a range of projectiles, knives, and scrapers, as well as tools, made from organic materials such as wood, bone, and antler. stone age cultures developed music and engaged in organized warfare. stone age humans developed ocean - worthy outrigger canoe technology, leading to migration across the malay archipelago, across the indian ocean to madagascar and also across the pacific ocean, which required knowledge of the ocean currents, weather patterns, sailing, and celestial navigation. although paleolithic cultures left no written records, the shift from nomadic life to settlement and agriculture can be inferred from a range of archaeological evidence. such evidence includes ancient tools, cave paintings, and other prehistoric art, such as the venus of willendorf. human remains also provide direct evidence, both through the examination of bones, and the study of mummies. scientists and historians have been able to form significant inferences about the lifestyle and culture of various prehistoric peoples, and especially their technology. = = = ancient = = = = = = = copper and bronze ages = = = = metallic copper occurs on the surface of weathered copper ore deposits and copper was used before copper smelting was known. copper smelting is believed to have originated when the technology of pottery kilns allowed sufficiently high temperatures. the concentration of various elements such as arsenic increase with depth in copper ore deposits and smelting of these ores yields arsenical bronze, which can be sufficiently work hardened to be suitable for making tools. bronze is an alloy of copper with tin ; the latter being found in relatively few deposits globally caused a long time to elapse before true tin bronze became widespread. ( see : tin sources and trade in ancient times ) bronze was a major advancement over stone as a material for making tools, both because of its mechanical properties like strength and ductility and because it could be cast in molds to make intricately shaped objects. bronze significantly advanced shipbuilding technology with better tools and bronze nails. bronze nails replaced the old method of attaching boards of the hull with cord woven through drilled holes. better ships
the propagation of terahertz waves in a dust acoustic wave is investigated numerically. by assuming a sinus profile of the dust number density in the dust acoustic waves, the transmission properties are calculated using finite difference time domain method. it shows that the dust acoustic wave can function similarly as a bragg filter to block the terahertz waves of a certain wavelength. the bandwidth of the filter depends on the density profile of the dust acoustic wave.
river valley during ancient times. the papyrus was harvested by field workers and brought to processing centers where it was cut into thin strips. the strips were then laid - out side by side and covered in plant resin. the second layer of strips was laid on perpendicularly, then both pressed together until the sheet was dry. the sheets were then joined to form a roll and later used for writing. egyptian society made several significant advances during dynastic periods in many areas of technology. according to hossam elanzeery, they were the first civilization to use timekeeping devices such as sundials, shadow clocks, and obelisks and successfully leveraged their knowledge of astronomy to create a calendar model that society still uses today. they developed shipbuilding technology that saw them progress from papyrus reed vessels to cedar wood ships while also pioneering the use of rope trusses and stem - mounted rudders. the egyptians also used their knowledge of anatomy to lay the foundation for many modern medical techniques and practiced the earliest known version of neuroscience. elanzeery also states that they used and furthered mathematical science, as evidenced in the building of the pyramids. ancient egyptians also invented and pioneered many food technologies that have become the basis of modern food technology processes. based on paintings and reliefs found in tombs, as well as archaeological artifacts, scholars like paul t nicholson believe that the ancient egyptians established systematic farming practices, engaged in cereal processing, brewed beer and baked bread, processed meat, practiced viticulture and created the basis for modern wine production, and created condiments to complement, preserve and mask the flavors of their food. = = = = indus valley = = = = the indus valley civilization, situated in a resource - rich area ( in modern pakistan and northwestern india ), is notable for its early application of city planning, sanitation technologies, and plumbing. indus valley construction and architecture, called ' vaastu shastra ', suggests a thorough understanding of materials engineering, hydrology, and sanitation. = = = = china = = = = the chinese made many first - known discoveries and developments. major technological contributions from china include the earliest known form of the binary code and epigenetic sequencing, early seismological detectors, matches, paper, helicopter rotor, raised - relief map, the double - action piston pump, cast iron, water powered blast furnace bellows, the iron plough, the multi - tube seed drill, the wheelbarrow, the parachute, the compass, the rudder, the crossbow, the south pointing chariot and gunpowder
planes at faster rates than other planes, thereby allowing more complicated 3 - d microstructures to be implemented. wet anisotropic etchants are often used in conjunction with boron etch stops wherein the surface of the silicon is heavily doped with boron resulting in a silicon material layer that is resistant to the wet etchants. this has been used in mews pressure sensor manufacturing for example. etching progresses at the same speed in all directions. long and narrow holes in a mask will produce v - shaped grooves in the silicon. the surface of these grooves can be atomically smooth if the etch is carried out correctly, with dimensions and angles being extremely accurate. some single crystal materials, such as silicon, will have different etching rates depending on the crystallographic orientation of the substrate. this is known as anisotropic etching and one of the most common examples is the etching of silicon in koh ( potassium hydroxide ), where si < 111 > planes etch approximately 100 times slower than other planes ( crystallographic orientations ). therefore, etching a rectangular hole in a ( 100 ) - si wafer results in a pyramid shaped etch pit with 54. 7Β° walls, instead of a hole with curved sidewalls as with isotropic etching. hydrofluoric acid is commonly used as an aqueous etchant for silicon dioxide ( sio2, also known as box for soi ), usually in 49 % concentrated form, 5 : 1, 10 : 1 or 20 : 1 boe ( buffered oxide etchant ) or bhf ( buffered hf ). they were first used in medieval times for glass etching. it was used in ic fabrication for patterning the gate oxide until the process step was replaced by rie. hydrofluoric acid is considered one of the more dangerous acids in the cleanroom. electrochemical etching ( ece ) for dopant - selective removal of silicon is a common method to automate and to selectively control etching. an active p β n diode junction is required, and either type of dopant can be the etch - resistant ( " etch - stop " ) material. boron is the most common etch - stop dopant. in combination with wet anisotropic etching as described above, ece has been used successfully for controlling silicon diaphragm thickness in commercial piezoresistive silicon pressure sensors
Question: A student pushes a wooden block across a sheet of sandpaper. Which characteristic of the block increases?
A) hardness
B) mass
C) size
D) temperature
|
D) temperature
|
Context:
behavioral responses to different stimuli, one can understand something about how those stimuli are processed. lewandowski & strohmetz ( 2009 ) reviewed a collection of innovative uses of behavioral measurement in psychology including behavioral traces, behavioral observations, and behavioral choice. behavioral traces are pieces of evidence that indicate behavior occurred, but the actor is not present ( e. g., litter in a parking lot or readings on an electric meter ). behavioral observations involve the direct witnessing of the actor engaging in the behavior ( e. g., watching how close a person sits next to another person ). behavioral choices are when a person selects between two or more options ( e. g., voting behavior, choice of a punishment for another participant ). reaction time. the time between the presentation of a stimulus and an appropriate response can indicate differences between two cognitive processes, and can indicate some things about their nature. for example, if in a search task the reaction times vary proportionally with the number of elements, then it is evident that this cognitive process of searching involves serial instead of parallel processing. psychophysical responses. psychophysical experiments are an old psychological technique, which has been adopted by cognitive psychology. they typically involve making judgments of some physical property, e. g. the loudness of a sound. correlation of subjective scales between individuals can show cognitive or sensory biases as compared to actual physical measurements. some examples include : sameness judgments for colors, tones, textures, etc. threshold differences for colors, tones, textures, etc. eye tracking. this methodology is used to study a variety of cognitive processes, most notably visual perception and language processing. the fixation point of the eyes is linked to an individual ' s focus of attention. thus, by monitoring eye movements, we can study what information is being processed at a given time. eye tracking allows us to study cognitive processes on extremely short time scales. eye movements reflect online decision making during a task, and they provide us with some insight into the ways in which those decisions may be processed. = = = brain imaging = = = brain imaging involves analyzing activity within the brain while performing various tasks. this allows us to link behavior and brain function to help understand how information is processed. different types of imaging techniques vary in their temporal ( time - based ) and spatial ( location - based ) resolution. brain imaging is often used in cognitive neuroscience. single - photon emission computed tomography and positron emission tomography. spect and pet use radioactive isotopes, which are injected into the subject ' s bloodstream
is not present ( e. g., litter in a parking lot or readings on an electric meter ). behavioral observations involve the direct witnessing of the actor engaging in the behavior ( e. g., watching how close a person sits next to another person ). behavioral choices are when a person selects between two or more options ( e. g., voting behavior, choice of a punishment for another participant ). reaction time. the time between the presentation of a stimulus and an appropriate response can indicate differences between two cognitive processes, and can indicate some things about their nature. for example, if in a search task the reaction times vary proportionally with the number of elements, then it is evident that this cognitive process of searching involves serial instead of parallel processing. psychophysical responses. psychophysical experiments are an old psychological technique, which has been adopted by cognitive psychology. they typically involve making judgments of some physical property, e. g. the loudness of a sound. correlation of subjective scales between individuals can show cognitive or sensory biases as compared to actual physical measurements. some examples include : sameness judgments for colors, tones, textures, etc. threshold differences for colors, tones, textures, etc. eye tracking. this methodology is used to study a variety of cognitive processes, most notably visual perception and language processing. the fixation point of the eyes is linked to an individual ' s focus of attention. thus, by monitoring eye movements, we can study what information is being processed at a given time. eye tracking allows us to study cognitive processes on extremely short time scales. eye movements reflect online decision making during a task, and they provide us with some insight into the ways in which those decisions may be processed. = = = brain imaging = = = brain imaging involves analyzing activity within the brain while performing various tasks. this allows us to link behavior and brain function to help understand how information is processed. different types of imaging techniques vary in their temporal ( time - based ) and spatial ( location - based ) resolution. brain imaging is often used in cognitive neuroscience. single - photon emission computed tomography and positron emission tomography. spect and pet use radioactive isotopes, which are injected into the subject ' s bloodstream and taken up by the brain. by observing which areas of the brain take up the radioactive isotope, we can see which areas of the brain are more active than other areas. pet has similar spatial resolution to fmri, but it has extremely poor temporal resolution. electroencephalography. eeg measures the electrical fields
and their competitive or mutualistic interactions with other species. some ecologists even rely on empirical data from indigenous people that is gathered by ethnobotanists. this information can relay a great deal of information on how the land once was thousands of years ago and how it has changed over that time. the goals of plant ecology are to understand the causes of their distribution patterns, productivity, environmental impact, evolution, and responses to environmental change. plants depend on certain edaphic ( soil ) and climatic factors in their environment but can modify these factors too. for example, they can change their environment ' s albedo, increase runoff interception, stabilise mineral soils and develop their organic content, and affect local temperature. plants compete with other organisms in their ecosystem for resources. they interact with their neighbours at a variety of spatial scales in groups, populations and communities that collectively constitute vegetation. regions with characteristic vegetation types and dominant plants as well as similar abiotic and biotic factors, climate, and geography make up biomes like tundra or tropical rainforest. herbivores eat plants, but plants can defend themselves and some species are parasitic or even carnivorous. other organisms form mutually beneficial relationships with plants. for example, mycorrhizal fungi and rhizobia provide plants with nutrients in exchange for food, ants are recruited by ant plants to provide protection, honey bees, bats and other animals pollinate flowers and humans and other animals act as dispersal vectors to spread spores and seeds. = = = plants, climate and environmental change = = = plant responses to climate and other environmental changes can inform our understanding of how these changes affect ecosystem function and productivity. for example, plant phenology can be a useful proxy for temperature in historical climatology, and the biological impact of climate change and global warming. palynology, the analysis of fossil pollen deposits in sediments from thousands or millions of years ago allows the reconstruction of past climates. estimates of atmospheric co2 concentrations since the palaeozoic have been obtained from stomatal densities and the leaf shapes and sizes of ancient land plants. ozone depletion can expose plants to higher levels of ultraviolet radiation - b ( uv - b ), resulting in lower growth rates. moreover, information from studies of community ecology, plant systematics, and taxonomy is essential to understanding vegetation change, habitat destruction and species extinction. = = genetics = = inheritance in plants follows the same fundamental principles of genetics as in other multicellular organisms. gregor mendel discovered the genetic laws of inheritance by studying
on biological causation and the diversity of life. he made countless observations of nature, especially the habits and attributes of plants and animals on lesbos, classified more than 540 animal species, and dissected at least 50. aristotle ' s writings profoundly influenced subsequent islamic and european scholarship, though they were eventually superseded in the scientific revolution. aristotle also contributed to theories of the elements and the cosmos. he believed that the celestial bodies ( such as the planets and the sun ) had something called an unmoved mover that put the celestial bodies in motion. aristotle tried to explain everything through mathematics and physics, but sometimes explained things such as the motion of celestial bodies through a higher power such as god. aristotle did not have the technological advancements that would have explained the motion of celestial bodies. in addition, aristotle had many views on the elements. he believed that everything was derived of the elements earth, water, air, fire, and lastly the aether. the aether was a celestial element, and therefore made up the matter of the celestial bodies. the elements of earth, water, air and fire were derived of a combination of two of the characteristics of hot, wet, cold, and dry, and all had their inevitable place and motion. the motion of these elements begins with earth being the closest to " the earth, " then water, air, fire, and finally aether. in addition to the makeup of all things, aristotle came up with theories as to why things did not return to their natural motion. he understood that water sits above earth, air above water, and fire above air in their natural state. he explained that although all elements must return to their natural state, the human body and other living things have a constraint on the elements β thus not allowing the elements making one who they are to return to their natural state. the important legacy of this period included substantial advances in factual knowledge, especially in anatomy, zoology, botany, mineralogy, geography, mathematics and astronomy ; an awareness of the importance of certain scientific problems, especially those related to the problem of change and its causes ; and a recognition of the methodological importance of applying mathematics to natural phenomena and of undertaking empirical research. in the hellenistic age scholars frequently employed the principles developed in earlier greek thought : the application of mathematics and deliberate empirical research, in their scientific investigations. thus, clear unbroken lines of influence lead from ancient greek and hellenistic philosophers, to medieval muslim philosophers and scientists, to the european renaissance and enlightenment, to the secular sciences of the modern day. neither reason
( or underlined when italics are not available ). the evolutionary relationships and heredity of a group of organisms is called its phylogeny. phylogenetic studies attempt to discover phylogenies. the basic approach is to use similarities based on shared inheritance to determine relationships. as an example, species of pereskia are trees or bushes with prominent leaves. they do not obviously resemble a typical leafless cactus such as an echinocactus. however, both pereskia and echinocactus have spines produced from areoles ( highly specialised pad - like structures ) suggesting that the two genera are indeed related. judging relationships based on shared characters requires care, since plants may resemble one another through convergent evolution in which characters have arisen independently. some euphorbias have leafless, rounded bodies adapted to water conservation similar to those of globular cacti, but characters such as the structure of their flowers make it clear that the two groups are not closely related. the cladistic method takes a systematic approach to characters, distinguishing between those that carry no information about shared evolutionary history β such as those evolved separately in different groups ( homoplasies ) or those left over from ancestors ( plesiomorphies ) β and derived characters, which have been passed down from innovations in a shared ancestor ( apomorphies ). only derived characters, such as the spine - producing areoles of cacti, provide evidence for descent from a common ancestor. the results of cladistic analyses are expressed as cladograms : tree - like diagrams showing the pattern of evolutionary branching and descent. from the 1990s onwards, the predominant approach to constructing phylogenies for living plants has been molecular phylogenetics, which uses molecular characters, particularly dna sequences, rather than morphological characters like the presence or absence of spines and areoles. the difference is that the genetic code itself is used to decide evolutionary relationships, instead of being used indirectly via the characters it gives rise to. clive stace describes this as having " direct access to the genetic basis of evolution. " as a simple example, prior to the use of genetic evidence, fungi were thought either to be plants or to be more closely related to plants than animals. genetic evidence suggests that the true evolutionary relationship of multicelled organisms is as shown in the cladogram below β fungi are more closely related to animals than to plants. in 1998, the angiosperm phylogeny group published a phylogeny for flowering plants based on an analysis of
options ( e. g., voting behavior, choice of a punishment for another participant ). reaction time. the time between the presentation of a stimulus and an appropriate response can indicate differences between two cognitive processes, and can indicate some things about their nature. for example, if in a search task the reaction times vary proportionally with the number of elements, then it is evident that this cognitive process of searching involves serial instead of parallel processing. psychophysical responses. psychophysical experiments are an old psychological technique, which has been adopted by cognitive psychology. they typically involve making judgments of some physical property, e. g. the loudness of a sound. correlation of subjective scales between individuals can show cognitive or sensory biases as compared to actual physical measurements. some examples include : sameness judgments for colors, tones, textures, etc. threshold differences for colors, tones, textures, etc. eye tracking. this methodology is used to study a variety of cognitive processes, most notably visual perception and language processing. the fixation point of the eyes is linked to an individual ' s focus of attention. thus, by monitoring eye movements, we can study what information is being processed at a given time. eye tracking allows us to study cognitive processes on extremely short time scales. eye movements reflect online decision making during a task, and they provide us with some insight into the ways in which those decisions may be processed. = = = brain imaging = = = brain imaging involves analyzing activity within the brain while performing various tasks. this allows us to link behavior and brain function to help understand how information is processed. different types of imaging techniques vary in their temporal ( time - based ) and spatial ( location - based ) resolution. brain imaging is often used in cognitive neuroscience. single - photon emission computed tomography and positron emission tomography. spect and pet use radioactive isotopes, which are injected into the subject ' s bloodstream and taken up by the brain. by observing which areas of the brain take up the radioactive isotope, we can see which areas of the brain are more active than other areas. pet has similar spatial resolution to fmri, but it has extremely poor temporal resolution. electroencephalography. eeg measures the electrical fields generated by large populations of neurons in the cortex by placing a series of electrodes on the scalp of the subject. this technique has an extremely high temporal resolution, but a relatively poor spatial resolution. functional magnetic resonance imaging. fmri measures the relative amount of oxygenated blood flowing to different parts of the brain. more oxygen
octet hyperon charge radii are calculated in a chiral constituent quark model including electromagnetic exchange currents between quarks. in impulse approximation one observes a decrease of the hyperon charge radii with increasing strangeness. this effect is reduced by exchange currents. due to exchange currents, the charge radius of the negatively charged hyperons are close to the proton charge radius.
cell. in juxtacrine signaling, there is direct contact between the signaling and responding cells. finally, hormones are ligands that travel through the circulatory systems of animals or vascular systems of plants to reach their target cells. once a ligand binds with a receptor, it can influence the behavior of another cell, depending on the type of receptor. for instance, neurotransmitters that bind with an inotropic receptor can alter the excitability of a target cell. other types of receptors include protein kinase receptors ( e. g., receptor for the hormone insulin ) and g protein - coupled receptors. activation of g protein - coupled receptors can initiate second messenger cascades. the process by which a chemical or physical signal is transmitted through a cell as a series of molecular events is called signal transduction. = = = cell cycle = = = the cell cycle is a series of events that take place in a cell that cause it to divide into two daughter cells. these events include the duplication of its dna and some of its organelles, and the subsequent partitioning of its cytoplasm into two daughter cells in a process called cell division. in eukaryotes ( i. e., animal, plant, fungal, and protist cells ), there are two distinct types of cell division : mitosis and meiosis. mitosis is part of the cell cycle, in which replicated chromosomes are separated into two new nuclei. cell division gives rise to genetically identical cells in which the total number of chromosomes is maintained. in general, mitosis ( division of the nucleus ) is preceded by the s stage of interphase ( during which the dna is replicated ) and is often followed by telophase and cytokinesis ; which divides the cytoplasm, organelles and cell membrane of one cell into two new cells containing roughly equal shares of these cellular components. the different stages of mitosis all together define the mitotic phase of an animal cell cycle β the division of the mother cell into two genetically identical daughter cells. the cell cycle is a vital process by which a single - celled fertilized egg develops into a mature organism, as well as the process by which hair, skin, blood cells, and some internal organs are renewed. after cell division, each of the daughter cells begin the interphase of a new cycle. in contrast to mitosis, meiosis results in four haploid daughter cells by undergoing one round of dna replication followed by two divisions
, like the woodward β hoffmann rules often come in handy while proposing a mechanism for a chemical reaction. according to the iupac gold book, a chemical reaction is " a process that results in the interconversion of chemical species. " accordingly, a chemical reaction may be an elementary reaction or a stepwise reaction. an additional caveat is made, in that this definition includes cases where the interconversion of conformers is experimentally observable. such detectable chemical reactions normally involve sets of molecular entities as indicated by this definition, but it is often conceptually convenient to use the term also for changes involving single molecular entities ( i. e. ' microscopic chemical events ' ). = = = ions and salts = = = an ion is a charged species, an atom or a molecule, that has lost or gained one or more electrons. when an atom loses an electron and thus has more protons than electrons, the atom is a positively charged ion or cation. when an atom gains an electron and thus has more electrons than protons, the atom is a negatively charged ion or anion. cations and anions can form a crystalline lattice of neutral salts, such as the na + and clβ ions forming sodium chloride, or nacl. examples of polyatomic ions that do not split up during acid β base reactions are hydroxide ( ohβ ) and phosphate ( po43β ). plasma is composed of gaseous matter that has been completely ionized, usually through high temperature. = = = acidity and basicity = = = a substance can often be classified as an acid or a base. there are several different theories which explain acid β base behavior. the simplest is arrhenius theory, which states that an acid is a substance that produces hydronium ions when it is dissolved in water, and a base is one that produces hydroxide ions when dissolved in water. according to brΓΈnsted β lowry acid β base theory, acids are substances that donate a positive hydrogen ion to another substance in a chemical reaction ; by extension, a base is the substance which receives that hydrogen ion. a third common theory is lewis acid β base theory, which is based on the formation of new chemical bonds. lewis theory explains that an acid is a substance which is capable of accepting a pair of electrons from another substance during the process of bond formation, while a base is a substance which can provide a pair of electrons to form a new bond. there are several other ways in which a substance may be classified as an acid
##ry facilitates screening and ranking of clones which bind to the native ( non - denaturated ) forms of antigen proteins. flow cytometry screening has been used for primary screening of a large number ( ~ 1000 ) of hybridoma clones recognizing the native form of the antigen on the cell surface. in the flow cytometry - based screening, a mixture of antigen - negative cells and antigen - positive cells is used as the antigen to be tested for each hybridoma supernatant sample. the b cell that produces the desired antibodies can be cloned to produce many identical daughter clones. supplemental media containing interleukin - 6 ( such as briclone ) are essential for this step. once a hybridoma colony is established, it will continually grow in culture medium like rpmi - 1640 ( with antibiotics and fetal bovine serum ) and produce antibodies. multiwell plates are used initially to grow the hybridomas, and after selection, are changed to larger tissue culture flasks. this maintains the well - being of the hybridomas and provides enough cells for cryopreservation and supernatant for subsequent investigations. the culture supernatant can yield 1 to 60 ΞΌg / ml of monoclonal antibody, which is maintained at - 20 Β°c or lower until required. by using culture supernatant or a purified immunoglobulin preparation, further analysis of a potential monoclonal antibody producing hybridoma can be made in terms of reactivity, specificity, and cross - reactivity. = = applications = = the use of monoclonal antibodies is numerous and includes the prevention, diagnosis, and treatment of disease. for example, monoclonal antibodies can distinguish subsets of b cells and t cells, which is helpful in identifying different types of leukaemias. in addition, specific monoclonal antibodies have been used to define cell surface markers on white blood cells and other cell types. this led to the cluster of differentiation series of markers. these are often referred to as cd markers and define several hundred different cell surface components of cells, each specified by binding of a particular monoclonal antibody. such antibodies are extremely useful for fluorescence - activated cell sorting, the specific isolation of particular types of cells. = = = in diagnostic histopathology = = = with the help of monoclonal antibodies, tissues and organs can be classified based on their expression of certain defined markers, which reflect tissue or cellular genesis. prostate specific antigen, placental alkaline phospha
Question: Amanda is learning about different adaptations of animals. Which is an example of a behavioral adaptation?
A) webbed feet on a duck
B) migration of songbirds
C) sharp eyesight of an eagle
D) camouflaged colors on a frog
|
B) migration of songbirds
|
Context:
defective body parts. inside the body, artificial heart valves are in common use with artificial hearts and lungs seeing less common use but under active technology development. other medical devices and aids that can be considered prosthetics include hearing aids, artificial eyes, palatal obturator, gastric bands, and dentures. prostheses are specifically not orthoses, although given certain circumstances a prosthesis might end up performing some or all of the same functionary benefits as an orthosis. prostheses are technically the complete finished item. for instance, a c - leg knee alone is not a prosthesis, but only a prosthetic component. the complete prosthesis would consist of the attachment system to the residual limb β usually a " socket ", and all the attachment hardware components all the way down to and including the terminal device. despite the technical difference, the terms are often used interchangeably. the terms " prosthetic " and " orthotic " are adjectives used to describe devices such as a prosthetic knee. the terms " prosthetics " and " orthotics " are used to describe the respective allied health fields. an occupational therapist ' s role in prosthetics include therapy, training and evaluations. prosthetic training includes orientation to prosthetics components and terminology, donning and doffing, wearing schedule, and how to care for residual limb and the prosthesis. = = = exoskeletons = = = a powered exoskeleton is a wearable mobile machine that is powered by a system of electric motors, pneumatics, levers, hydraulics, or a combination of technologies that allow for limb movement with increased strength and endurance. its design aims to provide back support, sense the user ' s motion, and send a signal to motors which manage the gears. the exoskeleton supports the shoulder, waist and thigh, and assists movement for lifting and holding heavy items, while lowering back stress. = = = adaptive seating and positioning = = = people with balance and motor function challenges often need specialized equipment to sit or stand safely and securely. this equipment is frequently specialized for specific settings such as in a classroom or nursing home. positioning is often important in seating arrangements to ensure that user ' s body pressure is distributed equally without inhibiting movement in a desired way. positioning devices have been developed to aid in allowing people to stand and bear weight on their legs without risk of a fall.
to dye denim and the artist ' s pigments gamboge and rose madder. sugar, starch, cotton, linen, hemp, some types of rope, wood and particle boards, papyrus and paper, vegetable oils, wax, and natural rubber are examples of commercially important materials made from plant tissues or their secondary products. charcoal, a pure form of carbon made by pyrolysis of wood, has a long history as a metal - smelting fuel, as a filter material and adsorbent and as an artist ' s material and is one of the three ingredients of gunpowder. cellulose, the world ' s most abundant organic polymer, can be converted into energy, fuels, materials and chemical feedstock. products made from cellulose include rayon and cellophane, wallpaper paste, biobutanol and gun cotton. sugarcane, rapeseed and soy are some of the plants with a highly fermentable sugar or oil content that are used as sources of biofuels, important alternatives to fossil fuels, such as biodiesel. sweetgrass was used by native americans to ward off bugs like mosquitoes. these bug repelling properties of sweetgrass were later found by the american chemical society in the molecules phytol and coumarin. = = plant ecology = = plant ecology is the science of the functional relationships between plants and their habitats β the environments where they complete their life cycles. plant ecologists study the composition of local and regional floras, their biodiversity, genetic diversity and fitness, the adaptation of plants to their environment, and their competitive or mutualistic interactions with other species. some ecologists even rely on empirical data from indigenous people that is gathered by ethnobotanists. this information can relay a great deal of information on how the land once was thousands of years ago and how it has changed over that time. the goals of plant ecology are to understand the causes of their distribution patterns, productivity, environmental impact, evolution, and responses to environmental change. plants depend on certain edaphic ( soil ) and climatic factors in their environment but can modify these factors too. for example, they can change their environment ' s albedo, increase runoff interception, stabilise mineral soils and develop their organic content, and affect local temperature. plants compete with other organisms in their ecosystem for resources. they interact with their neighbours at a variety of spatial scales in groups, populations and communities that collectively constitute vegetation. regions with characteristic vegetation types and dominant plants as well as similar abiot
industrial applications. this branch of biotechnology is the most used for the industries of refining and combustion principally on the production of bio - oils with photosynthetic micro - algae. green biotechnology is biotechnology applied to agricultural processes. an example would be the selection and domestication of plants via micropropagation. another example is the designing of transgenic plants to grow under specific environments in the presence ( or absence ) of chemicals. one hope is that green biotechnology might produce more environmentally friendly solutions than traditional industrial agriculture. an example of this is the engineering of a plant to express a pesticide, thereby ending the need of external application of pesticides. an example of this would be bt corn. whether or not green biotechnology products such as this are ultimately more environmentally friendly is a topic of considerable debate. it is commonly considered as the next phase of green revolution, which can be seen as a platform to eradicate world hunger by using technologies which enable the production of more fertile and resistant, towards biotic and abiotic stress, plants and ensures application of environmentally friendly fertilizers and the use of biopesticides, it is mainly focused on the development of agriculture. on the other hand, some of the uses of green biotechnology involve microorganisms to clean and reduce waste. red biotechnology is the use of biotechnology in the medical and pharmaceutical industries, and health preservation. this branch involves the production of vaccines and antibiotics, regenerative therapies, creation of artificial organs and new diagnostics of diseases. as well as the development of hormones, stem cells, antibodies, sirna and diagnostic tests. white biotechnology, also known as industrial biotechnology, is biotechnology applied to industrial processes. an example is the designing of an organism to produce a useful chemical. another example is the using of enzymes as industrial catalysts to either produce valuable chemicals or destroy hazardous / polluting chemicals. white biotechnology tends to consume less in resources than traditional processes used to produce industrial goods. yellow biotechnology refers to the use of biotechnology in food production ( food industry ), for example in making wine ( winemaking ), cheese ( cheesemaking ), and beer ( brewing ) by fermentation. it has also been used to refer to biotechnology applied to insects. this includes biotechnology - based approaches for the control of harmful insects, the characterisation and utilisation of active ingredients or genes of insects for research, or application in agriculture and medicine and various other approaches. gray biotechnology is dedicated to environmental applications, and focused on the maintenance of biodiversity and the remotion of poll
, specialty and engineering plastics. polyvinyl chloride ( pvc ) is widely used, inexpensive, and annual production quantities are large. it lends itself to a vast array of applications, from artificial leather to electrical insulation and cabling, packaging, and containers. its fabrication and processing are simple and well - established. the versatility of pvc is due to the wide range of plasticisers and other additives that it accepts. the term " additives " in polymer science refers to the chemicals and compounds added to the polymer base to modify its material properties. polycarbonate would be normally considered an engineering plastic ( other examples include peek, abs ). such plastics are valued for their superior strengths and other special material properties. they are usually not used for disposable applications, unlike commodity plastics. specialty plastics are materials with unique characteristics, such as ultra - high strength, electrical conductivity, electro - fluorescence, high thermal stability, etc. the dividing lines between the various types of plastics is not based on material but rather on their properties and applications. for example, polyethylene ( pe ) is a cheap, low friction polymer commonly used to make disposable bags for shopping and trash, and is considered a commodity plastic, whereas medium - density polyethylene ( mdpe ) is used for underground gas and water pipes, and another variety called ultra - high - molecular - weight polyethylene ( uhmwpe ) is an engineering plastic which is used extensively as the glide rails for industrial equipment and the low - friction socket in implanted hip joints. = = = metal alloys = = = the alloys of iron ( steel, stainless steel, cast iron, tool steel, alloy steels ) make up the largest proportion of metals today both by quantity and commercial value. iron alloyed with various proportions of carbon gives low, mid and high carbon steels. an iron - carbon alloy is only considered steel if the carbon level is between 0. 01 % and 2. 00 % by weight. for steels, the hardness and tensile strength of the steel is related to the amount of carbon present, with increasing carbon levels also leading to lower ductility and toughness. heat treatment processes such as quenching and tempering can significantly change these properties, however. in contrast, certain metal alloys exhibit unique properties where their size and density remain unchanged across a range of temperatures. cast iron is defined as an iron β carbon alloy with more than 2. 00 %, but less than 6. 67 %
include the manufacturing of drugs, creation of model animals that mimic human conditions and gene therapy. one of the earliest uses of genetic engineering was to mass - produce human insulin in bacteria. this application has now been applied to human growth hormones, follicle stimulating hormones ( for treating infertility ), human albumin, monoclonal antibodies, antihemophilic factors, vaccines and many other drugs. mouse hybridomas, cells fused together to create monoclonal antibodies, have been adapted through genetic engineering to create human monoclonal antibodies. genetically engineered viruses are being developed that can still confer immunity, but lack the infectious sequences. genetic engineering is also used to create animal models of human diseases. genetically modified mice are the most common genetically engineered animal model. they have been used to study and model cancer ( the oncomouse ), obesity, heart disease, diabetes, arthritis, substance abuse, anxiety, aging and parkinson disease. potential cures can be tested against these mouse models. gene therapy is the genetic engineering of humans, generally by replacing defective genes with effective ones. clinical research using somatic gene therapy has been conducted with several diseases, including x - linked scid, chronic lymphocytic leukemia ( cll ), and parkinson ' s disease. in 2012, alipogene tiparvovec became the first gene therapy treatment to be approved for clinical use. in 2015 a virus was used to insert a healthy gene into the skin cells of a boy suffering from a rare skin disease, epidermolysis bullosa, in order to grow, and then graft healthy skin onto 80 percent of the boy ' s body which was affected by the illness. germline gene therapy would result in any change being inheritable, which has raised concerns within the scientific community. in 2015, crispr was used to edit the dna of non - viable human embryos, leading scientists of major world academies to call for a moratorium on inheritable human genome edits. there are also concerns that the technology could be used not just for treatment, but for enhancement, modification or alteration of a human beings ' appearance, adaptability, intelligence, character or behavior. the distinction between cure and enhancement can also be difficult to establish. in november 2018, he jiankui announced that he had edited the genomes of two human embryos, to attempt to disable the ccr5 gene, which codes for a receptor that hiv uses to enter cells. the work was widely condemned as unethical, dangerous,
product of ceramic manufacture, or as an adjective. ceramics is the making of things out of ceramic materials. ceramic engineering, like many sciences, evolved from a different discipline by today ' s standards. materials science engineering is grouped with ceramics engineering to this day. abraham darby first used coke in 1709 in shropshire, england, to improve the yield of a smelting process. coke is now widely used to produce carbide ceramics. potter josiah wedgwood opened the first modern ceramics factory in stoke - on - trent, england, in 1759. austrian chemist carl josef bayer, working for the textile industry in russia, developed a process to separate alumina from bauxite ore in 1888. the bayer process is still used to purify alumina for the ceramic and aluminium industries. brothers pierre and jacques curie discovered piezoelectricity in rochelle salt c. 1880. piezoelectricity is one of the key properties of electroceramics. e. g. acheson heated a mixture of coke and clay in 1893, and invented carborundum, or synthetic silicon carbide. henri moissan also synthesized sic and tungsten carbide in his electric arc furnace in paris about the same time as acheson. karl schroter used liquid - phase sintering to bond or " cement " moissan ' s tungsten carbide particles with cobalt in 1923 in germany. cemented ( metal - bonded ) carbide edges greatly increase the durability of hardened steel cutting tools. w. h. nernst developed cubic - stabilized zirconia in the 1920s in berlin. this material is used as an oxygen sensor in exhaust systems. the main limitation on the use of ceramics in engineering is brittleness. = = = military = = = the military requirements of world war ii encouraged developments, which created a need for high - performance materials and helped speed the development of ceramic science and engineering. throughout the 1960s and 1970s, new types of ceramics were developed in response to advances in atomic energy, electronics, communications, and space travel. the discovery of ceramic superconductors in 1986 has spurred intense research to develop superconducting ceramic parts for electronic devices, electric motors, and transportation equipment. there is an increasing need in the military sector for high - strength, robust materials which have the capability to transmit light around the visible ( 0. 4 β 0. 7 micrometers ) and mid - infrared ( 1 β 5 micrometers ) regions of the spectrum. these materials
##drate - rich plant products such as barley ( beer ), rice ( sake ) and grapes ( wine ). native americans have used various plants as ways of treating illness or disease for thousands of years. this knowledge native americans have on plants has been recorded by enthnobotanists and then in turn has been used by pharmaceutical companies as a way of drug discovery. plants can synthesise coloured dyes and pigments such as the anthocyanins responsible for the red colour of red wine, yellow weld and blue woad used together to produce lincoln green, indoxyl, source of the blue dye indigo traditionally used to dye denim and the artist ' s pigments gamboge and rose madder. sugar, starch, cotton, linen, hemp, some types of rope, wood and particle boards, papyrus and paper, vegetable oils, wax, and natural rubber are examples of commercially important materials made from plant tissues or their secondary products. charcoal, a pure form of carbon made by pyrolysis of wood, has a long history as a metal - smelting fuel, as a filter material and adsorbent and as an artist ' s material and is one of the three ingredients of gunpowder. cellulose, the world ' s most abundant organic polymer, can be converted into energy, fuels, materials and chemical feedstock. products made from cellulose include rayon and cellophane, wallpaper paste, biobutanol and gun cotton. sugarcane, rapeseed and soy are some of the plants with a highly fermentable sugar or oil content that are used as sources of biofuels, important alternatives to fossil fuels, such as biodiesel. sweetgrass was used by native americans to ward off bugs like mosquitoes. these bug repelling properties of sweetgrass were later found by the american chemical society in the molecules phytol and coumarin. = = plant ecology = = plant ecology is the science of the functional relationships between plants and their habitats β the environments where they complete their life cycles. plant ecologists study the composition of local and regional floras, their biodiversity, genetic diversity and fitness, the adaptation of plants to their environment, and their competitive or mutualistic interactions with other species. some ecologists even rely on empirical data from indigenous people that is gathered by ethnobotanists. this information can relay a great deal of information on how the land once was thousands of years ago and how it has changed over that time. the goals of
used by pharmaceutical companies as a way of drug discovery. plants can synthesise coloured dyes and pigments such as the anthocyanins responsible for the red colour of red wine, yellow weld and blue woad used together to produce lincoln green, indoxyl, source of the blue dye indigo traditionally used to dye denim and the artist ' s pigments gamboge and rose madder. sugar, starch, cotton, linen, hemp, some types of rope, wood and particle boards, papyrus and paper, vegetable oils, wax, and natural rubber are examples of commercially important materials made from plant tissues or their secondary products. charcoal, a pure form of carbon made by pyrolysis of wood, has a long history as a metal - smelting fuel, as a filter material and adsorbent and as an artist ' s material and is one of the three ingredients of gunpowder. cellulose, the world ' s most abundant organic polymer, can be converted into energy, fuels, materials and chemical feedstock. products made from cellulose include rayon and cellophane, wallpaper paste, biobutanol and gun cotton. sugarcane, rapeseed and soy are some of the plants with a highly fermentable sugar or oil content that are used as sources of biofuels, important alternatives to fossil fuels, such as biodiesel. sweetgrass was used by native americans to ward off bugs like mosquitoes. these bug repelling properties of sweetgrass were later found by the american chemical society in the molecules phytol and coumarin. = = plant ecology = = plant ecology is the science of the functional relationships between plants and their habitats β the environments where they complete their life cycles. plant ecologists study the composition of local and regional floras, their biodiversity, genetic diversity and fitness, the adaptation of plants to their environment, and their competitive or mutualistic interactions with other species. some ecologists even rely on empirical data from indigenous people that is gathered by ethnobotanists. this information can relay a great deal of information on how the land once was thousands of years ago and how it has changed over that time. the goals of plant ecology are to understand the causes of their distribution patterns, productivity, environmental impact, evolution, and responses to environmental change. plants depend on certain edaphic ( soil ) and climatic factors in their environment but can modify these factors too. for example, they can change their environment ' s albedo, increase runoff interception
the creation of your own reality and your own world. the metaphor i used was humans being like magic markers. for so long, they painted black and white pictures in their life because that ' s all they thought they could do. but they can paint with a different color and make a very vibrant and beautiful picture if they take control. on the single " new skin ", he further elaborated : in " new skin ", i attribute a scab to the present state of society. the way the scab looks in its worst state is gross and chaotic and horrible, that ' s now, but when it breaks away, there ' s a brand new piece of skin that ' s stronger than before. it ' s like creation out of chaos. the song " favorite things ", according to boyd, related to the topic of religion : " my favorite things " is my personal beliefs about religion and how it oppresses the things i enjoy the most. unfortunately, the simplest things, such as thinking for myself, creating my own reality and being whatever the hell i want to be each day of my life, are a sin. to be a good christian basically means to give up the reigns of your life and let some unseen force do it for you. " favorite things " also includes a sample of the 1959 track " flamenco fantasy ", by easy listening group the 101 strings orchestra. the song has a similar title to " my favorite things ", from the mary poppins musical and film, with both songs repeatedly mentioning their titles in the lyrics. however, it does not musically reference " my favorite things ". the single " a certain shade of green " has been described as being a song about procrastination. the line " are you gonna stand around till 2012 a. d.? " is a reference to an interpretation of the mayan calendar which dictated that the world would end on december 21, 2012. boyd did not believe this to be true, but it was on his mind as his mother was researching it for a book called maya memory : the glory that was palenque. while recording " nebula ", boyd said in 1997, " we found out what it ' s like to actually plug a phaser pedal into the wall while it ' s on. it sounds like a laser gun, and that ' s the first sound you hear in ' nebula '. " he added that for the song, " we used these walkie - talkies for children that have this slinky - like coil between them. when
the broad definition of " utilizing a biotechnological system to make products ". indeed, the cultivation of plants may be viewed as the earliest biotechnological enterprise. agriculture has been theorized to have become the dominant way of producing food since the neolithic revolution. through early biotechnology, the earliest farmers selected and bred the best - suited crops ( e. g., those with the highest yields ) to produce enough food to support a growing population. as crops and fields became increasingly large and difficult to maintain, it was discovered that specific organisms and their by - products could effectively fertilize, restore nitrogen, and control pests. throughout the history of agriculture, farmers have inadvertently altered the genetics of their crops through introducing them to new environments and breeding them with other plants β one of the first forms of biotechnology. these processes also were included in early fermentation of beer. these processes were introduced in early mesopotamia, egypt, china and india, and still use the same basic biological methods. in brewing, malted grains ( containing enzymes ) convert starch from grains into sugar and then adding specific yeasts to produce beer. in this process, carbohydrates in the grains broke down into alcohols, such as ethanol. later, other cultures produced the process of lactic acid fermentation, which produced other preserved foods, such as soy sauce. fermentation was also used in this time period to produce leavened bread. although the process of fermentation was not fully understood until louis pasteur ' s work in 1857, it is still the first use of biotechnology to convert a food source into another form. before the time of charles darwin ' s work and life, animal and plant scientists had already used selective breeding. darwin added to that body of work with his scientific observations about the ability of science to change species. these accounts contributed to darwin ' s theory of natural selection. for thousands of years, humans have used selective breeding to improve the production of crops and livestock to use them for food. in selective breeding, organisms with desirable characteristics are mated to produce offspring with the same characteristics. for example, this technique was used with corn to produce the largest and sweetest crops. in the early twentieth century scientists gained a greater understanding of microbiology and explored ways of manufacturing specific products. in 1917, chaim weizmann first used a pure microbiological culture in an industrial process, that of manufacturing corn starch using clostridium acetobutylicum, to produce acetone, which the united
Question: Which of the following is the best example of a custom-made product?
A) graphing calculator
B) light bulb
C) needle nose pliers
D) artificial leg
|
D) artificial leg
|
Context:
made of steel. the shoe is generally wider than the caisson to reduce friction, and the leading edge may be supplied with pressurised bentonite slurry, which swells in water, stabilizing settlement by filling depressions and voids. an open caisson may fill with water during sinking. the material is excavated by clamshell excavator bucket on crane. the formation level subsoil may still not be suitable for excavation or bearing capacity. the water in the caisson ( due to a high water table ) balances the upthrust forces of the soft soils underneath. if dewatered, the base may " pipe " or " boil ", causing the caisson to sink. to combat this problem, piles may be driven from the surface to act as : load - bearing walls, in that they transmit loads to deeper soils. anchors, in that they resist flotation because of the friction at the interface between their surfaces and the surrounding earth into which they have been driven. h - beam sections ( typical column sections, due to resistance to bending in all axis ) may be driven at angles " raked " to rock or other firmer soils ; the h - beams are left extended above the base. a reinforced concrete plug may be placed under the water, a process known as tremie concrete placement. when the caisson is dewatered, this plug acts as a pile cap, resisting the upward forces of the subsoil. = = = monolithic = = = a monolithic caisson ( or simply a monolith ) is larger than the other types of caisson, but similar to open caissons. such caissons are often found in quay walls, where resistance to impact from ships is required. = = = pneumatic = = = shallow caissons may be open to the air, whereas pneumatic caissons ( sometimes called pressurized caissons ), which penetrate soft mud, are bottomless boxes sealed at the top and filled with compressed air to keep water and mud out at depth. an airlock allows access to the chamber. workers, called sandhogs in american english, move mud and rock debris ( called muck ) from the edge of the workspace to a water - filled pit, connected by a tube ( called the muck tube ) to the surface. a crane at the surface removes the soil with a clamshell bucket. the water pressure in the tube balances the air pressure, with excess air escaping up
be the more significant to modern soil theory than fallou ' s. previously, soil had been considered a product of chemical transformations of rocks, a dead substrate from which plants derive nutritious elements. soil and bedrock were in fact equated. dokuchaev considers the soil as a natural body having its own genesis and its own history of development, a body with complex and multiform processes taking place within it. the soil is considered as different from bedrock. the latter becomes soil under the influence of a series of soil - formation factors ( climate, vegetation, country, relief and age ). according to him, soil should be called the " daily " or outward horizons of rocks regardless of the type ; they are changed naturally by the common effect of water, air and various kinds of living and dead organisms. a 1914 encyclopedic definition : " the different forms of earth on the surface of the rocks, formed by the breaking down or weathering of rocks ". serves to illustrate the historic view of soil which persisted from the 19th century. dokuchaev ' s late 19th century soil concept developed in the 20th century to one of soil as earthy material that has been altered by living processes. a corollary concept is that soil without a living component is simply a part of earth ' s outer layer. further refinement of the soil concept is occurring in view of an appreciation of energy transport and transformation within soil. the term is popularly applied to the material on the surface of the earth ' s moon and mars, a usage acceptable within a portion of the scientific community. accurate to this modern understanding of soil is nikiforoff ' s 1959 definition of soil as the " excited skin of the sub aerial part of the earth ' s crust ". = = areas of practice = = academically, soil scientists tend to be drawn to one of five areas of specialization : microbiology, pedology, edaphology, physics, or chemistry. yet the work specifics are very much dictated by the challenges facing our civilization ' s desire to sustain the land that supports it, and the distinctions between the sub - disciplines of soil science often blur in the process. soil science professionals commonly stay current in soil chemistry, soil physics, soil microbiology, pedology, and applied soil science in related disciplines. one exciting effort drawing in soil scientists in the u. s. as of 2004 is the soil quality initiative. central to the soil quality initiative is developing indices of soil health and then monitoring them in a way
= = = = = = environmental remediation = = = environmental remediation is the process through which contaminants or pollutants in soil, water and other media are removed to improve environmental quality. the main focus is the reduction of hazardous substances within the environment. some of the areas involved in environmental remediation include ; soil contamination, hazardous waste, groundwater contamination, oil, gas and chemical spills. there are three most common types of environmental remediation. these include soil, water, and sediment remediation. soil remediation consists of removing contaminants in soil, as these pose great risks to humans and the ecosystem. some examples of this are heavy metals, pesticides, and radioactive materials. depending on the contaminant the remedial processes can be physical, chemical, thermal, or biological. water remediation is one of the most important considering water is an essential natural resource. depending on the source of water there will be different contaminants. surface water contamination mainly consists of agricultural, animal, and industrial waste, as well as acid mine drainage. there has been a rise in the need for water remediation due to the increased discharge of industrial waste, leading to a demand for sustainable water solutions. the market for water remediation is expected to consistently increase to $ 19. 6 billion by 2030. sediment remediation consists of removing contaminated sediments. is it almost similar to soil remediation except it is often more sophisticated as it involves additional contaminants. to reduce the contaminants it is likely to use physical, chemical, and biological processes that help with source control, but if these processes are executed correctly, there ' s a risk of contamination resurfacing. = = = solid waste management = = = solid waste management is the purification, consumption, reuse, disposal, and treatment of solid waste that is undertaken by the government or the ruling bodies of a city / town. it refers to the collection, treatment, and disposal of non - soluble, solid waste material. solid waste is associated with both industrial, institutional, commercial and residential activities. hazardous solid waste, when improperly disposed can encourage the infestation of insects and rodents, contributing to the spread of diseases. some of the most common types of solid waste management include ; landfills, vermicomposting, composting, recycling, and incineration. however, a major barrier for solid waste management practices is the high costs associated with recycling
equalizing the flow at these places, produces a lowering of the river above the rapids by facilitating the efflux, which may result in the appearance of fresh shoals at the low stage of the river. where, however, narrow rocky reefs or other hard shoals stretch across the bottom of a river and present obstacles to the erosion by the current of the soft materials forming the bed of the river above and below, their removal may result in permanent improvement by enabling the river to deepen its bed by natural scour. the capability of a river to provide a waterway for navigation during the summer or throughout the dry season depends on the depth that can be secured in the channel at the lowest stage. the problem in the dry season is the small discharge and deficiency in scour during this period. a typical solution is to restrict the width of the low - water channel, concentrate all of the flow in it, and also to fix its position so that it is scoured out every year by the floods which follow the deepest part of the bed along the line of the strongest current. this can be effected by closing subsidiary low - water channels with dikes across them, and narrowing the channel at the low stage by low - dipping cross dikes extending from the river banks down the slope and pointing slightly up - stream so as to direct the water flowing over them into a central channel. = = estuarine works = = the needs of navigation may also require that a stable, continuous, navigable channel is prolonged from the navigable river to deep water at the mouth of the estuary. the interaction of river flow and tide needs to be modeled by computer or using scale models, moulded to the configuration of the estuary under consideration and reproducing in miniature the tidal ebb and flow and fresh - water discharge over a bed of fine sand, in which various lines of training walls can be successively inserted. the models should be capable of furnishing valuable indications of the respective effects and comparative merits of the different schemes proposed for works. = = see also = = bridge scour flood control = = references = = = = external links = = u. s. army corps of engineers β civil works program river morphology and stream restoration references - wildland hydrology at the library of congress web archives ( archived 2002 - 08 - 13 )
genesis and its own history of development, a body with complex and multiform processes taking place within it. the soil is considered as different from bedrock. the latter becomes soil under the influence of a series of soil - formation factors ( climate, vegetation, country, relief and age ). according to him, soil should be called the " daily " or outward horizons of rocks regardless of the type ; they are changed naturally by the common effect of water, air and various kinds of living and dead organisms. a 1914 encyclopedic definition : " the different forms of earth on the surface of the rocks, formed by the breaking down or weathering of rocks ". serves to illustrate the historic view of soil which persisted from the 19th century. dokuchaev ' s late 19th century soil concept developed in the 20th century to one of soil as earthy material that has been altered by living processes. a corollary concept is that soil without a living component is simply a part of earth ' s outer layer. further refinement of the soil concept is occurring in view of an appreciation of energy transport and transformation within soil. the term is popularly applied to the material on the surface of the earth ' s moon and mars, a usage acceptable within a portion of the scientific community. accurate to this modern understanding of soil is nikiforoff ' s 1959 definition of soil as the " excited skin of the sub aerial part of the earth ' s crust ". = = areas of practice = = academically, soil scientists tend to be drawn to one of five areas of specialization : microbiology, pedology, edaphology, physics, or chemistry. yet the work specifics are very much dictated by the challenges facing our civilization ' s desire to sustain the land that supports it, and the distinctions between the sub - disciplines of soil science often blur in the process. soil science professionals commonly stay current in soil chemistry, soil physics, soil microbiology, pedology, and applied soil science in related disciplines. one exciting effort drawing in soil scientists in the u. s. as of 2004 is the soil quality initiative. central to the soil quality initiative is developing indices of soil health and then monitoring them in a way that gives us long - term ( decade - to - decade ) feedback on our performance as stewards of the planet. the effort includes understanding the functions of soil microbiotic crusts and exploring the potential to sequester atmospheric carbon in soil organic matter. relating the concept of agriculture to soil quality, however, has not
to be separated conceptually from geology and crop production and treated as a whole. as a founding father of soil science, fallou has primacy in time. fallou was working on the origins of soil before dokuchaev was born ; however dokuchaev ' s work was more extensive and is considered to be the more significant to modern soil theory than fallou ' s. previously, soil had been considered a product of chemical transformations of rocks, a dead substrate from which plants derive nutritious elements. soil and bedrock were in fact equated. dokuchaev considers the soil as a natural body having its own genesis and its own history of development, a body with complex and multiform processes taking place within it. the soil is considered as different from bedrock. the latter becomes soil under the influence of a series of soil - formation factors ( climate, vegetation, country, relief and age ). according to him, soil should be called the " daily " or outward horizons of rocks regardless of the type ; they are changed naturally by the common effect of water, air and various kinds of living and dead organisms. a 1914 encyclopedic definition : " the different forms of earth on the surface of the rocks, formed by the breaking down or weathering of rocks ". serves to illustrate the historic view of soil which persisted from the 19th century. dokuchaev ' s late 19th century soil concept developed in the 20th century to one of soil as earthy material that has been altered by living processes. a corollary concept is that soil without a living component is simply a part of earth ' s outer layer. further refinement of the soil concept is occurring in view of an appreciation of energy transport and transformation within soil. the term is popularly applied to the material on the surface of the earth ' s moon and mars, a usage acceptable within a portion of the scientific community. accurate to this modern understanding of soil is nikiforoff ' s 1959 definition of soil as the " excited skin of the sub aerial part of the earth ' s crust ". = = areas of practice = = academically, soil scientists tend to be drawn to one of five areas of specialization : microbiology, pedology, edaphology, physics, or chemistry. yet the work specifics are very much dictated by the challenges facing our civilization ' s desire to sustain the land that supports it, and the distinctions between the sub - disciplines of soil science often blur in the process. soil science professionals commonly stay current
on a large scale provided protection from insect pests or tolerance to herbicides. fungal and virus resistant crops have also been developed or are in development. this makes the insect and weed management of crops easier and can indirectly increase crop yield. gm crops that directly improve yield by accelerating growth or making the plant more hardy ( by improving salt, cold or drought tolerance ) are also under development. in 2016 salmon have been genetically modified with growth hormones to reach normal adult size much faster. gmos have been developed that modify the quality of produce by increasing the nutritional value or providing more industrially useful qualities or quantities. the amflora potato produces a more industrially useful blend of starches. soybeans and canola have been genetically modified to produce more healthy oils. the first commercialised gm food was a tomato that had delayed ripening, increasing its shelf life. plants and animals have been engineered to produce materials they do not normally make. pharming uses crops and animals as bioreactors to produce vaccines, drug intermediates, or the drugs themselves ; the useful product is purified from the harvest and then used in the standard pharmaceutical production process. cows and goats have been engineered to express drugs and other proteins in their milk, and in 2009 the fda approved a drug produced in goat milk. = = = other applications = = = genetic engineering has potential applications in conservation and natural area management. gene transfer through viral vectors has been proposed as a means of controlling invasive species as well as vaccinating threatened fauna from disease. transgenic trees have been suggested as a way to confer resistance to pathogens in wild populations. with the increasing risks of maladaptation in organisms as a result of climate change and other perturbations, facilitated adaptation through gene tweaking could be one solution to reducing extinction risks. applications of genetic engineering in conservation are thus far mostly theoretical and have yet to be put into practice. genetic engineering is also being used to create microbial art. some bacteria have been genetically engineered to create black and white photographs. novelty items such as lavender - colored carnations, blue roses, and glowing fish, have also been produced through genetic engineering. = = regulation = = the regulation of genetic engineering concerns the approaches taken by governments to assess and manage the risks associated with the development and release of gmos. the development of a regulatory framework began in 1975, at asilomar, california. the asilomar meeting recommended a set of voluntary guidelines regarding the use of recombinant technology. as the technology improved
##nts from the air to reduce the potential adverse effects on humans and the environment. the process of air purification may be performed using methods such as mechanical filtration, ionization, activated carbon adsorption, photocatalytic oxidation, and ultraviolet light germicidal irradiation. = = = sewage treatment = = = = = = environmental remediation = = = environmental remediation is the process through which contaminants or pollutants in soil, water and other media are removed to improve environmental quality. the main focus is the reduction of hazardous substances within the environment. some of the areas involved in environmental remediation include ; soil contamination, hazardous waste, groundwater contamination, oil, gas and chemical spills. there are three most common types of environmental remediation. these include soil, water, and sediment remediation. soil remediation consists of removing contaminants in soil, as these pose great risks to humans and the ecosystem. some examples of this are heavy metals, pesticides, and radioactive materials. depending on the contaminant the remedial processes can be physical, chemical, thermal, or biological. water remediation is one of the most important considering water is an essential natural resource. depending on the source of water there will be different contaminants. surface water contamination mainly consists of agricultural, animal, and industrial waste, as well as acid mine drainage. there has been a rise in the need for water remediation due to the increased discharge of industrial waste, leading to a demand for sustainable water solutions. the market for water remediation is expected to consistently increase to $ 19. 6 billion by 2030. sediment remediation consists of removing contaminated sediments. is it almost similar to soil remediation except it is often more sophisticated as it involves additional contaminants. to reduce the contaminants it is likely to use physical, chemical, and biological processes that help with source control, but if these processes are executed correctly, there ' s a risk of contamination resurfacing. = = = solid waste management = = = solid waste management is the purification, consumption, reuse, disposal, and treatment of solid waste that is undertaken by the government or the ruling bodies of a city / town. it refers to the collection, treatment, and disposal of non - soluble, solid waste material. solid waste is associated with both industrial, institutional, commercial and residential activities. hazardous solid waste, when improperly disposed can encourage the
to increase as far as practicable the navigable depth at the lowest stage of the water level. engineering works to increase the navigability of rivers can only be advantageously undertaken in large rivers with a moderate fall and a fair discharge at their lowest stage, for with a large fall the current presents a great impediment to up - stream navigation, and there are generally variations in water level, and when the discharge becomes small in the dry season. it is impossible to maintain a sufficient depth of water in the low - water channel. the possibility to secure uniformity of depth in a river by lowering the shoals obstructing the channel depends on the nature of the shoals. a soft shoal in the bed of a river is due to deposit from a diminution in velocity of flow, produced by a reduction in fall and by a widening of the channel, or to a loss in concentration of the scour of the main current in passing over from one concave bank to the next on the opposite side. the lowering of such a shoal by dredging merely effects a temporary deepening, for it soon forms again from the causes which produced it. the removal, moreover, of the rocky obstructions at rapids, though increasing the depth and equalizing the flow at these places, produces a lowering of the river above the rapids by facilitating the efflux, which may result in the appearance of fresh shoals at the low stage of the river. where, however, narrow rocky reefs or other hard shoals stretch across the bottom of a river and present obstacles to the erosion by the current of the soft materials forming the bed of the river above and below, their removal may result in permanent improvement by enabling the river to deepen its bed by natural scour. the capability of a river to provide a waterway for navigation during the summer or throughout the dry season depends on the depth that can be secured in the channel at the lowest stage. the problem in the dry season is the small discharge and deficiency in scour during this period. a typical solution is to restrict the width of the low - water channel, concentrate all of the flow in it, and also to fix its position so that it is scoured out every year by the floods which follow the deepest part of the bed along the line of the strongest current. this can be effected by closing subsidiary low - water channels with dikes across them, and narrowing the channel at the low stage by low - dipping cross dikes extending from the river banks down the
and nine classes, depending on its color, texture and hydrology. contemporaries friedrich albert fallou ( the german founder of modern soil science ) and vasily dokuchaev ( the russian founder of modern soil science ) are both credited with being among the first to identify soil as a resource whose distinctness and complexity deserved to be separated conceptually from geology and crop production and treated as a whole. as a founding father of soil science, fallou has primacy in time. fallou was working on the origins of soil before dokuchaev was born ; however dokuchaev ' s work was more extensive and is considered to be the more significant to modern soil theory than fallou ' s. previously, soil had been considered a product of chemical transformations of rocks, a dead substrate from which plants derive nutritious elements. soil and bedrock were in fact equated. dokuchaev considers the soil as a natural body having its own genesis and its own history of development, a body with complex and multiform processes taking place within it. the soil is considered as different from bedrock. the latter becomes soil under the influence of a series of soil - formation factors ( climate, vegetation, country, relief and age ). according to him, soil should be called the " daily " or outward horizons of rocks regardless of the type ; they are changed naturally by the common effect of water, air and various kinds of living and dead organisms. a 1914 encyclopedic definition : " the different forms of earth on the surface of the rocks, formed by the breaking down or weathering of rocks ". serves to illustrate the historic view of soil which persisted from the 19th century. dokuchaev ' s late 19th century soil concept developed in the 20th century to one of soil as earthy material that has been altered by living processes. a corollary concept is that soil without a living component is simply a part of earth ' s outer layer. further refinement of the soil concept is occurring in view of an appreciation of energy transport and transformation within soil. the term is popularly applied to the material on the surface of the earth ' s moon and mars, a usage acceptable within a portion of the scientific community. accurate to this modern understanding of soil is nikiforoff ' s 1959 definition of soil as the " excited skin of the sub aerial part of the earth ' s crust ". = = areas of practice = = academically, soil scientists tend to be drawn to one of five areas of specialization : microbiology
Question: Michelle is interested in improving the quality of the soil in her yard. She wants more water and air to penetrate the soil. What should she add to the soil to help her achieve this goal?
A) bacteria
B) fertilizers
C) pesticides
D) earthworms
|
D) earthworms
|
Context:
becomes quite gentle. accordingly, in large basins, rivers in most cases begin as torrents with a variable flow, and end as gently flowing rivers with a comparatively regular discharge. the irregular flow of rivers throughout their course forms one of the main difficulties in devising works for mitigating inundations or for increasing the navigable capabilities of rivers. in tropical countries subject to periodical rains, the rivers are in flood during the rainy season and have hardly any flow during the rest of the year, while in temperate regions, where the rainfall is more evenly distributed throughout the year, evaporation causes the available rainfall to be much less in hot summer weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under
navigable capabilities of rivers. in tropical countries subject to periodical rains, the rivers are in flood during the rainy season and have hardly any flow during the rest of the year, while in temperate regions, where the rainfall is more evenly distributed throughout the year, evaporation causes the available rainfall to be much less in hot summer weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under ordinary conditions, most of the materials brought down from the high lands by torrential water courses are carried forward by the main river to the sea, or partially strewn over flat alluvial plains during floods ; the size of the materials forming the bed of the river or borne along by the stream is gradually reduced on proceeding sea
superheated droplets are proven to be excelent detectors for neutrons and could be used as a neutron dosimeter. to detect accurately the volume of the vapour formed upon nucleation and hence to observe the nucleation quantitatively an air displacement system has been developed.
approximately corresponds to the slope of the country it traverses ; as rivers rise close to the highest part of their basins, generally in hilly regions, their fall is rapid near their source and gradually diminishes, with occasional irregularities, until, in traversing plains along the latter part of their course, their fall usually becomes quite gentle. accordingly, in large basins, rivers in most cases begin as torrents with a variable flow, and end as gently flowing rivers with a comparatively regular discharge. the irregular flow of rivers throughout their course forms one of the main difficulties in devising works for mitigating inundations or for increasing the navigable capabilities of rivers. in tropical countries subject to periodical rains, the rivers are in flood during the rainy season and have hardly any flow during the rest of the year, while in temperate regions, where the rainfall is more evenly distributed throughout the year, evaporation causes the available rainfall to be much less in hot summer weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with
using only lidar or radar an accurate cloud boundary height estimate is often not possible. the combination of lidar and radar can give a reliable cloud boundary estimate in a much broader range of cases. however, also this combination with standard methods still can not measure the cloud boundaries in all cases. this will be illustrated with data from the clouds and radiation measurement campaigns, clara. rain is a problem : the radar has problems to measure the small cloud droplets in the presence of raindrops. similarly, few large particles below cloud base can obscure the cloud base in radar measurements. and the radar reflectivity can be very low at the cloud base of water clouds or in large regions of ice clouds, due to small particles. multiple cloud layers and clouds with specular reflections can pose problems for lidar. more advanced measurement techniques are suggested to solve these problems. an angle scanning lidar can, for example, detect specular reflections, while using information from the radars doppler velocity spectrum may help to detect clouds during rain.
weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under ordinary conditions, most of the materials brought down from the high lands by torrential water courses are carried forward by the main river to the sea, or partially strewn over flat alluvial plains during floods ; the size of the materials forming the bed of the river or borne along by the stream is gradually reduced on proceeding seawards, so that in the po river in italy, for instance, pebbles and gravel are found for about 140 miles below turin, sand along the next 100 miles, and silt and mud in the last 110 miles ( 176 km ). = = channelization = = the removal of obstructions, natural or artificial
floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under ordinary conditions, most of the materials brought down from the high lands by torrential water courses are carried forward by the main river to the sea, or partially strewn over flat alluvial plains during floods ; the size of the materials forming the bed of the river or borne along by the stream is gradually reduced on proceeding seawards, so that in the po river in italy, for instance, pebbles and gravel are found for about 140 miles below turin, sand along the next 100 miles, and silt and mud in the last 110 miles ( 176 km ). = = channelization = = the removal of obstructions, natural or artificial ( e. g., trunks of trees, boulders and accumulations of gravel ) from a river bed furnishes a simple and efficient means of increasing the discharging capacity of its channel. such removals will consequently lower the height of floods upstream. every impediment to the flow, in proportion to its extent, raises the level of the river above it so as to produce the additional artificial fall necessary to convey the flow through the restricted channel, thereby reducing the total available fall. reducing the length of the channel by substituting straight cuts for a winding course is the only way in which the effective fall can be increased.
discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under ordinary conditions, most of the materials brought down from the high lands by torrential water courses are carried forward by the main river to the sea, or partially strewn over flat alluvial plains during floods ; the size of the materials forming the bed of the river or borne along by the stream is gradually reduced on proceeding seawards, so that in the po river in italy, for instance, pebbles and gravel are found for about 140 miles below turin, sand along the next 100 miles, and silt and mud in the last 110 miles ( 176 km ). = = channelization = = the removal of obstructions, natural or artificial ( e. g., trunks of trees, boulders and accumulations of gravel ) from a river bed furnishes a simple and efficient means of increasing the discharging capacity of its channel. such removals will consequently lower the height of floods upstream. every impediment to the flow, in proportion to its extent, raises the level of the river above it so as to produce the additional artificial fall necessary to convey the flow through the restricted channel, thereby reducing the total available fall. reducing the length of the channel by substituting straight cuts for a winding course is the only way in which the effective fall can be increased. this involves some loss of capacity in the channel as a whole, and in the case of a large river with a considerable flow it is difficult to maintain a straight cut owing to the tendency of the current to erode the banks and form again a sinuous channel. even if the cut is preserved by protecting the banks,
hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under ordinary conditions, most of the materials brought down from the high lands by torrential water courses are carried forward by the main river to the sea, or partially strewn over flat alluvial plains during floods ; the size of the materials forming the bed of the river or borne along by the stream is gradually reduced on proceeding seawards, so that in the po river in italy, for instance, pebbles and gravel are found for about 140 miles below turin, sand along the next 100 miles, and silt and mud in the last 110 miles ( 176 km ). = = channelization = = the removal of obstructions, natural or artificial ( e. g., trunks of trees, boulders and accumulations of gravel ) from a river bed furnishes a simple and efficient means of increasing the discharging capacity of its channel. such removals will consequently lower the height of floods upstream. every impediment to the flow, in proportion to
a minimum atmospheric temperature, or tropopause, occurs at a pressure of around 0. 1 bar in the atmospheres of earth, titan, jupiter, saturn, uranus and neptune, despite great differences in atmospheric composition, gravity, internal heat and sunlight. in all these bodies, the tropopause separates a stratosphere with a temperature profile that is controlled by the absorption of shortwave solar radiation, from a region below characterised by convection, weather, and clouds. however, it is not obvious why the tropopause occurs at the specific pressure near 0. 1 bar. here we use a physically - based model to demonstrate that, at atmospheric pressures lower than 0. 1 bar, transparency to thermal radiation allows shortwave heating to dominate, creating a stratosphere. at higher pressures, atmospheres become opaque to thermal radiation, causing temperatures to increase with depth and convection to ensue. a common dependence of infrared opacity on pressure, arising from the shared physics of molecular absorption, sets the 0. 1 bar tropopause. we hypothesize that a tropopause at a pressure of approximately 0. 1 bar is characteristic of many thick atmospheres, including exoplanets and exomoons in our galaxy and beyond. judicious use of this rule could help constrain the atmospheric structure, and thus the surface environments and habitability, of exoplanets.
Question: Which of these could cause water vapor to condense into rain droplets?
A) sunlight warming the surface of oceans
B) air rising into colder regions of the atmosphere
C) lightning heating the air in the atmosphere
D) winds blowing along the surface of the ground
|
B) air rising into colder regions of the atmosphere
|
Context:
a process of fermentation. the pyruvate is not transported into the mitochondrion but remains in the cytoplasm, where it is converted to waste products that may be removed from the cell. this serves the purpose of oxidizing the electron carriers so that they can perform glycolysis again and removing the excess pyruvate. fermentation oxidizes nadh to nad + so it can be re - used in glycolysis. in the absence of oxygen, fermentation prevents the buildup of nadh in the cytoplasm and provides nad + for glycolysis. this waste product varies depending on the organism. in skeletal muscles, the waste product is lactic acid. this type of fermentation is called lactic acid fermentation. in strenuous exercise, when energy demands exceed energy supply, the respiratory chain cannot process all of the hydrogen atoms joined by nadh. during anaerobic glycolysis, nad + regenerates when pairs of hydrogen combine with pyruvate to form lactate. lactate formation is catalyzed by lactate dehydrogenase in a reversible reaction. lactate can also be used as an indirect precursor for liver glycogen. during recovery, when oxygen becomes available, nad + attaches to hydrogen from lactate to form atp. in yeast, the waste products are ethanol and carbon dioxide. this type of fermentation is known as alcoholic or ethanol fermentation. the atp generated in this process is made by substrate - level phosphorylation, which does not require oxygen. = = = photosynthesis = = = photosynthesis is a process used by plants and other organisms to convert light energy into chemical energy that can later be released to fuel the organism ' s metabolic activities via cellular respiration. this chemical energy is stored in carbohydrate molecules, such as sugars, which are synthesized from carbon dioxide and water. in most cases, oxygen is released as a waste product. most plants, algae, and cyanobacteria perform photosynthesis, which is largely responsible for producing and maintaining the oxygen content of the earth ' s atmosphere, and supplies most of the energy necessary for life on earth. photosynthesis has four stages : light absorption, electron transport, atp synthesis, and carbon fixation. light absorption is the initial step of photosynthesis whereby light energy is absorbed by chlorophyll pigments attached to
organelles. in terms of their structural composition, the microtubules are made up of tubulin ( e. g., Ξ± - tubulin and Ξ² - tubulin ) whereas intermediate filaments are made up of fibrous proteins. microfilaments are made up of actin molecules that interact with other strands of proteins. = = = metabolism = = = all cells require energy to sustain cellular processes. metabolism is the set of chemical reactions in an organism. the three main purposes of metabolism are : the conversion of food to energy to run cellular processes ; the conversion of food / fuel to monomer building blocks ; and the elimination of metabolic wastes. these enzyme - catalyzed reactions allow organisms to grow and reproduce, maintain their structures, and respond to their environments. metabolic reactions may be categorized as catabolic β the breaking down of compounds ( for example, the breaking down of glucose to pyruvate by cellular respiration ) ; or anabolic β the building up ( synthesis ) of compounds ( such as proteins, carbohydrates, lipids, and nucleic acids ). usually, catabolism releases energy, and anabolism consumes energy. the chemical reactions of metabolism are organized into metabolic pathways, in which one chemical is transformed through a series of steps into another chemical, each step being facilitated by a specific enzyme. enzymes are crucial to metabolism because they allow organisms to drive desirable reactions that require energy that will not occur by themselves, by coupling them to spontaneous reactions that release energy. enzymes act as catalysts β they allow a reaction to proceed more rapidly without being consumed by it β by reducing the amount of activation energy needed to convert reactants into products. enzymes also allow the regulation of the rate of a metabolic reaction, for example in response to changes in the cell ' s environment or to signals from other cells. = = = cellular respiration = = = cellular respiration is a set of metabolic reactions and processes that take place in cells to convert chemical energy from nutrients into adenosine triphosphate ( atp ), and then release waste products. the reactions involved in respiration are catabolic reactions, which break large molecules into smaller ones, releasing energy. respiration is one of the key ways a cell releases chemical energy to fuel cellular activity. the overall reaction occurs in a series of biochemical steps, some of which are redox reactions. although cellular respiration is technically a combustion reaction, it clearly does not resemble one when it occurs in a cell because of the
the structural components of cells. as a by - product of photosynthesis, plants release oxygen into the atmosphere, a gas that is required by nearly all living things to carry out cellular respiration. in addition, they are influential in the global carbon and water cycles and plant roots bind and stabilise soils, preventing soil erosion. plants are crucial to the future of human society as they provide food, oxygen, biochemicals, and products for people, as well as creating and preserving soil. historically, all living things were classified as either animals or plants and botany covered the study of all organisms not considered animals. botanists examine both the internal functions and processes within plant organelles, cells, tissues, whole plants, plant populations and plant communities. at each of these levels, a botanist may be concerned with the classification ( taxonomy ), phylogeny and evolution, structure ( anatomy and morphology ), or function ( physiology ) of plant life. the strictest definition of " plant " includes only the " land plants " or embryophytes, which include seed plants ( gymnosperms, including the pines, and flowering plants ) and the free - sporing cryptogams including ferns, clubmosses, liverworts, hornworts and mosses. embryophytes are multicellular eukaryotes descended from an ancestor that obtained its energy from sunlight by photosynthesis. they have life cycles with alternating haploid and diploid phases. the sexual haploid phase of embryophytes, known as the gametophyte, nurtures the developing diploid embryo sporophyte within its tissues for at least part of its life, even in the seed plants, where the gametophyte itself is nurtured by its parent sporophyte. other groups of organisms that were previously studied by botanists include bacteria ( now studied in bacteriology ), fungi ( mycology ) β including lichen - forming fungi ( lichenology ), non - chlorophyte algae ( phycology ), and viruses ( virology ). however, attention is still given to these groups by botanists, and fungi ( including lichens ) and photosynthetic protists are usually covered in introductory botany courses. palaeobotanists study ancient plants in the fossil record to provide information about the evolutionary history of plants. cyanobacteria, the first oxygen - releasing photosynthetic organisms on earth, are thought to have given rise to the
= = = = = = environmental remediation = = = environmental remediation is the process through which contaminants or pollutants in soil, water and other media are removed to improve environmental quality. the main focus is the reduction of hazardous substances within the environment. some of the areas involved in environmental remediation include ; soil contamination, hazardous waste, groundwater contamination, oil, gas and chemical spills. there are three most common types of environmental remediation. these include soil, water, and sediment remediation. soil remediation consists of removing contaminants in soil, as these pose great risks to humans and the ecosystem. some examples of this are heavy metals, pesticides, and radioactive materials. depending on the contaminant the remedial processes can be physical, chemical, thermal, or biological. water remediation is one of the most important considering water is an essential natural resource. depending on the source of water there will be different contaminants. surface water contamination mainly consists of agricultural, animal, and industrial waste, as well as acid mine drainage. there has been a rise in the need for water remediation due to the increased discharge of industrial waste, leading to a demand for sustainable water solutions. the market for water remediation is expected to consistently increase to $ 19. 6 billion by 2030. sediment remediation consists of removing contaminated sediments. is it almost similar to soil remediation except it is often more sophisticated as it involves additional contaminants. to reduce the contaminants it is likely to use physical, chemical, and biological processes that help with source control, but if these processes are executed correctly, there ' s a risk of contamination resurfacing. = = = solid waste management = = = solid waste management is the purification, consumption, reuse, disposal, and treatment of solid waste that is undertaken by the government or the ruling bodies of a city / town. it refers to the collection, treatment, and disposal of non - soluble, solid waste material. solid waste is associated with both industrial, institutional, commercial and residential activities. hazardous solid waste, when improperly disposed can encourage the infestation of insects and rodents, contributing to the spread of diseases. some of the most common types of solid waste management include ; landfills, vermicomposting, composting, recycling, and incineration. however, a major barrier for solid waste management practices is the high costs associated with recycling
generally, dead - end filtration is used for feasibility studies on a laboratory scale. the dead - end membranes are relatively easy to fabricate which reduces the cost of the separation process. the dead - end membrane separation process is easy to implement and the process is usually cheaper than cross - flow membrane filtration. the dead - end filtration process is usually a batch - type process, where the filtering solution is loaded ( or slowly fed ) into the membrane device, which then allows passage of some particles subject to the driving force. the main disadvantage of dead - end filtration is the extensive membrane fouling and concentration polarization. the fouling is usually induced faster at higher driving forces. membrane fouling and particle retention in a feed solution also builds up a concentration gradients and particle backflow ( concentration polarization ). the tangential flow devices are more cost and labor - intensive, but they are less susceptible to fouling due to the sweeping effects and high shear rates of the passing flow. the most commonly used synthetic membrane devices ( modules ) are flat sheets / plates, spiral wounds, and hollow fibers. flat membranes used in filtration and separation processes can be enhanced with surface patterning, where microscopic structures are introduced to improve performance. these patterns increase surface area, optimize water flow, and reduce fouling, leading to higher permeability and longer membrane lifespan. research has shown that such modifications can significantly enhance efficiency in water purification, energy applications, and industrial separations. flat plates are usually constructed as circular thin flat membrane surfaces to be used in dead - end geometry modules. spiral wounds are constructed from similar flat membranes but in the form of a " pocket " containing two membrane sheets separated by a highly porous support plate. several such pockets are then wound around a tube to create a tangential flow geometry and to reduce membrane fouling. hollow fiber modules consist of an assembly of self - supporting fibers with dense skin separation layers, and a more open matrix helping to withstand pressure gradients and maintain structural integrity. the hollow fiber modules can contain up to 10, 000 fibers ranging from 200 to 2500 ΞΌm in diameter ; the main advantage of hollow fiber modules is the very large surface area within an enclosed volume, increasing the efficiency of the separation process. the disc tube module uses a cross - flow geometry and consists of a pressure tube and hydraulic discs, which are held by a central tension rod, and membrane cushions that lie between two discs. = = membrane performance and governing equations = = the selection of synthetic membranes
shaping the cell. cell membranes are involved in various cellular processes such as cell adhesion, storing electrical energy, and cell signalling and serve as the attachment surface for several extracellular structures such as a cell wall, glycocalyx, and cytoskeleton. within the cytoplasm of a cell, there are many biomolecules such as proteins and nucleic acids. in addition to biomolecules, eukaryotic cells have specialized structures called organelles that have their own lipid bilayers or are spatially units. these organelles include the cell nucleus, which contains most of the cell ' s dna, or mitochondria, which generate adenosine triphosphate ( atp ) to power cellular processes. other organelles such as endoplasmic reticulum and golgi apparatus play a role in the synthesis and packaging of proteins, respectively. biomolecules such as proteins can be engulfed by lysosomes, another specialized organelle. plant cells have additional organelles that distinguish them from animal cells such as a cell wall that provides support for the plant cell, chloroplasts that harvest sunlight energy to produce sugar, and vacuoles that provide storage and structural support as well as being involved in reproduction and breakdown of plant seeds. eukaryotic cells also have cytoskeleton that is made up of microtubules, intermediate filaments, and microfilaments, all of which provide support for the cell and are involved in the movement of the cell and its organelles. in terms of their structural composition, the microtubules are made up of tubulin ( e. g., Ξ± - tubulin and Ξ² - tubulin ) whereas intermediate filaments are made up of fibrous proteins. microfilaments are made up of actin molecules that interact with other strands of proteins. = = = metabolism = = = all cells require energy to sustain cellular processes. metabolism is the set of chemical reactions in an organism. the three main purposes of metabolism are : the conversion of food to energy to run cellular processes ; the conversion of food / fuel to monomer building blocks ; and the elimination of metabolic wastes. these enzyme - catalyzed reactions allow organisms to grow and reproduce, maintain their structures, and respond to their environments. metabolic reactions may be categorized as catabolic β the breaking down of compounds ( for example, the breaking down of glucose to pyruvate by cellular respiration
from the proton motive force drives the enzyme atp synthase to synthesize more atps by phosphorylating adps. the transfer of electrons terminates with molecular oxygen being the final electron acceptor. if oxygen were not present, pyruvate would not be metabolized by cellular respiration but undergoes a process of fermentation. the pyruvate is not transported into the mitochondrion but remains in the cytoplasm, where it is converted to waste products that may be removed from the cell. this serves the purpose of oxidizing the electron carriers so that they can perform glycolysis again and removing the excess pyruvate. fermentation oxidizes nadh to nad + so it can be re - used in glycolysis. in the absence of oxygen, fermentation prevents the buildup of nadh in the cytoplasm and provides nad + for glycolysis. this waste product varies depending on the organism. in skeletal muscles, the waste product is lactic acid. this type of fermentation is called lactic acid fermentation. in strenuous exercise, when energy demands exceed energy supply, the respiratory chain cannot process all of the hydrogen atoms joined by nadh. during anaerobic glycolysis, nad + regenerates when pairs of hydrogen combine with pyruvate to form lactate. lactate formation is catalyzed by lactate dehydrogenase in a reversible reaction. lactate can also be used as an indirect precursor for liver glycogen. during recovery, when oxygen becomes available, nad + attaches to hydrogen from lactate to form atp. in yeast, the waste products are ethanol and carbon dioxide. this type of fermentation is known as alcoholic or ethanol fermentation. the atp generated in this process is made by substrate - level phosphorylation, which does not require oxygen. = = = photosynthesis = = = photosynthesis is a process used by plants and other organisms to convert light energy into chemical energy that can later be released to fuel the organism ' s metabolic activities via cellular respiration. this chemical energy is stored in carbohydrate molecules, such as sugars, which are synthesized from carbon dioxide and water. in most cases, oxygen is released as a waste product. most plants, algae, and cyanobacteria perform photosynthesis, which is largely responsible for producing and maintaining the oxygen content of
, there are many biomolecules such as proteins and nucleic acids. in addition to biomolecules, eukaryotic cells have specialized structures called organelles that have their own lipid bilayers or are spatially units. these organelles include the cell nucleus, which contains most of the cell ' s dna, or mitochondria, which generate adenosine triphosphate ( atp ) to power cellular processes. other organelles such as endoplasmic reticulum and golgi apparatus play a role in the synthesis and packaging of proteins, respectively. biomolecules such as proteins can be engulfed by lysosomes, another specialized organelle. plant cells have additional organelles that distinguish them from animal cells such as a cell wall that provides support for the plant cell, chloroplasts that harvest sunlight energy to produce sugar, and vacuoles that provide storage and structural support as well as being involved in reproduction and breakdown of plant seeds. eukaryotic cells also have cytoskeleton that is made up of microtubules, intermediate filaments, and microfilaments, all of which provide support for the cell and are involved in the movement of the cell and its organelles. in terms of their structural composition, the microtubules are made up of tubulin ( e. g., Ξ± - tubulin and Ξ² - tubulin ) whereas intermediate filaments are made up of fibrous proteins. microfilaments are made up of actin molecules that interact with other strands of proteins. = = = metabolism = = = all cells require energy to sustain cellular processes. metabolism is the set of chemical reactions in an organism. the three main purposes of metabolism are : the conversion of food to energy to run cellular processes ; the conversion of food / fuel to monomer building blocks ; and the elimination of metabolic wastes. these enzyme - catalyzed reactions allow organisms to grow and reproduce, maintain their structures, and respond to their environments. metabolic reactions may be categorized as catabolic β the breaking down of compounds ( for example, the breaking down of glucose to pyruvate by cellular respiration ) ; or anabolic β the building up ( synthesis ) of compounds ( such as proteins, carbohydrates, lipids, and nucleic acids ). usually, catabolism releases energy, and anabolism consumes energy. the chemical reactions of metabolism are organized into metabolic pathways, in which
##nts from the air to reduce the potential adverse effects on humans and the environment. the process of air purification may be performed using methods such as mechanical filtration, ionization, activated carbon adsorption, photocatalytic oxidation, and ultraviolet light germicidal irradiation. = = = sewage treatment = = = = = = environmental remediation = = = environmental remediation is the process through which contaminants or pollutants in soil, water and other media are removed to improve environmental quality. the main focus is the reduction of hazardous substances within the environment. some of the areas involved in environmental remediation include ; soil contamination, hazardous waste, groundwater contamination, oil, gas and chemical spills. there are three most common types of environmental remediation. these include soil, water, and sediment remediation. soil remediation consists of removing contaminants in soil, as these pose great risks to humans and the ecosystem. some examples of this are heavy metals, pesticides, and radioactive materials. depending on the contaminant the remedial processes can be physical, chemical, thermal, or biological. water remediation is one of the most important considering water is an essential natural resource. depending on the source of water there will be different contaminants. surface water contamination mainly consists of agricultural, animal, and industrial waste, as well as acid mine drainage. there has been a rise in the need for water remediation due to the increased discharge of industrial waste, leading to a demand for sustainable water solutions. the market for water remediation is expected to consistently increase to $ 19. 6 billion by 2030. sediment remediation consists of removing contaminated sediments. is it almost similar to soil remediation except it is often more sophisticated as it involves additional contaminants. to reduce the contaminants it is likely to use physical, chemical, and biological processes that help with source control, but if these processes are executed correctly, there ' s a risk of contamination resurfacing. = = = solid waste management = = = solid waste management is the purification, consumption, reuse, disposal, and treatment of solid waste that is undertaken by the government or the ruling bodies of a city / town. it refers to the collection, treatment, and disposal of non - soluble, solid waste material. solid waste is associated with both industrial, institutional, commercial and residential activities. hazardous solid waste, when improperly disposed can encourage the
the surface of the membrane, retentate is removed from the same side further downstream, whereas the permeate flow is tracked on the other side. in dead - end filtration, the direction of the fluid flow is normal to the membrane surface. both flow geometries offer some advantages and disadvantages. generally, dead - end filtration is used for feasibility studies on a laboratory scale. the dead - end membranes are relatively easy to fabricate which reduces the cost of the separation process. the dead - end membrane separation process is easy to implement and the process is usually cheaper than cross - flow membrane filtration. the dead - end filtration process is usually a batch - type process, where the filtering solution is loaded ( or slowly fed ) into the membrane device, which then allows passage of some particles subject to the driving force. the main disadvantage of dead - end filtration is the extensive membrane fouling and concentration polarization. the fouling is usually induced faster at higher driving forces. membrane fouling and particle retention in a feed solution also builds up a concentration gradients and particle backflow ( concentration polarization ). the tangential flow devices are more cost and labor - intensive, but they are less susceptible to fouling due to the sweeping effects and high shear rates of the passing flow. the most commonly used synthetic membrane devices ( modules ) are flat sheets / plates, spiral wounds, and hollow fibers. flat membranes used in filtration and separation processes can be enhanced with surface patterning, where microscopic structures are introduced to improve performance. these patterns increase surface area, optimize water flow, and reduce fouling, leading to higher permeability and longer membrane lifespan. research has shown that such modifications can significantly enhance efficiency in water purification, energy applications, and industrial separations. flat plates are usually constructed as circular thin flat membrane surfaces to be used in dead - end geometry modules. spiral wounds are constructed from similar flat membranes but in the form of a " pocket " containing two membrane sheets separated by a highly porous support plate. several such pockets are then wound around a tube to create a tangential flow geometry and to reduce membrane fouling. hollow fiber modules consist of an assembly of self - supporting fibers with dense skin separation layers, and a more open matrix helping to withstand pressure gradients and maintain structural integrity. the hollow fiber modules can contain up to 10, 000 fibers ranging from 200 to 2500 ΞΌm in diameter ; the main advantage of hollow fiber modules is the very large surface area within
Question: The human body processes and eliminates food waste using the organs of the excretory system. Which organelle performs a similar function in humans at the cellular level?
A) mitochondrion
B) endoplasmic reticulum
C) lysosome
D) Golgi complex
|
C) lysosome
|
Context:
##morphology studies the origin of landscapes. structural geology studies the deformation of rocks to produce mountains and lowlands. resource geology studies how energy resources can be obtained from minerals. environmental geology studies how pollution and contaminants affect soil and rock. mineralogy is the study of minerals and includes the study of mineral formation, crystal structure, hazards associated with minerals, and the physical and chemical properties of minerals. petrology is the study of rocks, including the formation and composition of rocks. petrography is a branch of petrology that studies the typology and classification of rocks. = = earth ' s interior = = plate tectonics, mountain ranges, volcanoes, and earthquakes are geological phenomena that can be explained in terms of physical and chemical processes in the earth ' s crust. beneath the earth ' s crust lies the mantle which is heated by the radioactive decay of heavy elements. the mantle is not quite solid and consists of magma which is in a state of semi - perpetual convection. this convection process causes the lithospheric plates to move, albeit slowly. the resulting process is known as plate tectonics. areas of the crust where new crust is created are called divergent boundaries, those where it is brought back into the earth are convergent boundaries and those where plates slide past each other, but no new lithospheric material is created or destroyed, are referred to as transform ( or conservative ) boundaries. earthquakes result from the movement of the lithospheric plates, and they often occur near convergent boundaries where parts of the crust are forced into the earth as part of subduction. plate tectonics might be thought of as the process by which the earth is resurfaced. as the result of seafloor spreading, new crust and lithosphere is created by the flow of magma from the mantle to the near surface, through fissures, where it cools and solidifies. through subduction, oceanic crust and lithosphere vehemently returns to the convecting mantle. volcanoes result primarily from the melting of subducted crust material. crust material that is forced into the asthenosphere melts, and some portion of the melted material becomes light enough to rise to the surface β giving birth to volcanoes. = = atmospheric science = = atmospheric science initially developed in the late - 19th century as a means to forecast the weather through meteorology, the study of weather. atmospheric chemistry was developed in the 20th century to measure air pollution and expanded in the 1970s in response to
be the more significant to modern soil theory than fallou ' s. previously, soil had been considered a product of chemical transformations of rocks, a dead substrate from which plants derive nutritious elements. soil and bedrock were in fact equated. dokuchaev considers the soil as a natural body having its own genesis and its own history of development, a body with complex and multiform processes taking place within it. the soil is considered as different from bedrock. the latter becomes soil under the influence of a series of soil - formation factors ( climate, vegetation, country, relief and age ). according to him, soil should be called the " daily " or outward horizons of rocks regardless of the type ; they are changed naturally by the common effect of water, air and various kinds of living and dead organisms. a 1914 encyclopedic definition : " the different forms of earth on the surface of the rocks, formed by the breaking down or weathering of rocks ". serves to illustrate the historic view of soil which persisted from the 19th century. dokuchaev ' s late 19th century soil concept developed in the 20th century to one of soil as earthy material that has been altered by living processes. a corollary concept is that soil without a living component is simply a part of earth ' s outer layer. further refinement of the soil concept is occurring in view of an appreciation of energy transport and transformation within soil. the term is popularly applied to the material on the surface of the earth ' s moon and mars, a usage acceptable within a portion of the scientific community. accurate to this modern understanding of soil is nikiforoff ' s 1959 definition of soil as the " excited skin of the sub aerial part of the earth ' s crust ". = = areas of practice = = academically, soil scientists tend to be drawn to one of five areas of specialization : microbiology, pedology, edaphology, physics, or chemistry. yet the work specifics are very much dictated by the challenges facing our civilization ' s desire to sustain the land that supports it, and the distinctions between the sub - disciplines of soil science often blur in the process. soil science professionals commonly stay current in soil chemistry, soil physics, soil microbiology, pedology, and applied soil science in related disciplines. one exciting effort drawing in soil scientists in the u. s. as of 2004 is the soil quality initiative. central to the soil quality initiative is developing indices of soil health and then monitoring them in a way
to be separated conceptually from geology and crop production and treated as a whole. as a founding father of soil science, fallou has primacy in time. fallou was working on the origins of soil before dokuchaev was born ; however dokuchaev ' s work was more extensive and is considered to be the more significant to modern soil theory than fallou ' s. previously, soil had been considered a product of chemical transformations of rocks, a dead substrate from which plants derive nutritious elements. soil and bedrock were in fact equated. dokuchaev considers the soil as a natural body having its own genesis and its own history of development, a body with complex and multiform processes taking place within it. the soil is considered as different from bedrock. the latter becomes soil under the influence of a series of soil - formation factors ( climate, vegetation, country, relief and age ). according to him, soil should be called the " daily " or outward horizons of rocks regardless of the type ; they are changed naturally by the common effect of water, air and various kinds of living and dead organisms. a 1914 encyclopedic definition : " the different forms of earth on the surface of the rocks, formed by the breaking down or weathering of rocks ". serves to illustrate the historic view of soil which persisted from the 19th century. dokuchaev ' s late 19th century soil concept developed in the 20th century to one of soil as earthy material that has been altered by living processes. a corollary concept is that soil without a living component is simply a part of earth ' s outer layer. further refinement of the soil concept is occurring in view of an appreciation of energy transport and transformation within soil. the term is popularly applied to the material on the surface of the earth ' s moon and mars, a usage acceptable within a portion of the scientific community. accurate to this modern understanding of soil is nikiforoff ' s 1959 definition of soil as the " excited skin of the sub aerial part of the earth ' s crust ". = = areas of practice = = academically, soil scientists tend to be drawn to one of five areas of specialization : microbiology, pedology, edaphology, physics, or chemistry. yet the work specifics are very much dictated by the challenges facing our civilization ' s desire to sustain the land that supports it, and the distinctions between the sub - disciplines of soil science often blur in the process. soil science professionals commonly stay current
remediation include ; soil contamination, hazardous waste, groundwater contamination, oil, gas and chemical spills. there are three most common types of environmental remediation. these include soil, water, and sediment remediation. soil remediation consists of removing contaminants in soil, as these pose great risks to humans and the ecosystem. some examples of this are heavy metals, pesticides, and radioactive materials. depending on the contaminant the remedial processes can be physical, chemical, thermal, or biological. water remediation is one of the most important considering water is an essential natural resource. depending on the source of water there will be different contaminants. surface water contamination mainly consists of agricultural, animal, and industrial waste, as well as acid mine drainage. there has been a rise in the need for water remediation due to the increased discharge of industrial waste, leading to a demand for sustainable water solutions. the market for water remediation is expected to consistently increase to $ 19. 6 billion by 2030. sediment remediation consists of removing contaminated sediments. is it almost similar to soil remediation except it is often more sophisticated as it involves additional contaminants. to reduce the contaminants it is likely to use physical, chemical, and biological processes that help with source control, but if these processes are executed correctly, there ' s a risk of contamination resurfacing. = = = solid waste management = = = solid waste management is the purification, consumption, reuse, disposal, and treatment of solid waste that is undertaken by the government or the ruling bodies of a city / town. it refers to the collection, treatment, and disposal of non - soluble, solid waste material. solid waste is associated with both industrial, institutional, commercial and residential activities. hazardous solid waste, when improperly disposed can encourage the infestation of insects and rodents, contributing to the spread of diseases. some of the most common types of solid waste management include ; landfills, vermicomposting, composting, recycling, and incineration. however, a major barrier for solid waste management practices is the high costs associated with recycling and the risks of creating more pollution. = = = e - waste recycling = = = the recycling of electronic waste ( e - waste ) has seen significant technological advancements due to increasing environmental concerns and the growing volume of electronic product disposals. traditional e - waste recycling methods, which often involve manual disassemb
##sphere ( or lithosphere ). earth science can be considered to be a branch of planetary science but with a much older history. = = geology = = geology is broadly the study of earth ' s structure, substance, and processes. geology is largely the study of the lithosphere, or earth ' s surface, including the crust and rocks. it includes the physical characteristics and processes that occur in the lithosphere as well as how they are affected by geothermal energy. it incorporates aspects of chemistry, physics, and biology as elements of geology interact. historical geology is the application of geology to interpret earth history and how it has changed over time. geochemistry studies the chemical components and processes of the earth. geophysics studies the physical properties of the earth. paleontology studies fossilized biological material in the lithosphere. planetary geology studies geoscience as it pertains to extraterrestrial bodies. geomorphology studies the origin of landscapes. structural geology studies the deformation of rocks to produce mountains and lowlands. resource geology studies how energy resources can be obtained from minerals. environmental geology studies how pollution and contaminants affect soil and rock. mineralogy is the study of minerals and includes the study of mineral formation, crystal structure, hazards associated with minerals, and the physical and chemical properties of minerals. petrology is the study of rocks, including the formation and composition of rocks. petrography is a branch of petrology that studies the typology and classification of rocks. = = earth ' s interior = = plate tectonics, mountain ranges, volcanoes, and earthquakes are geological phenomena that can be explained in terms of physical and chemical processes in the earth ' s crust. beneath the earth ' s crust lies the mantle which is heated by the radioactive decay of heavy elements. the mantle is not quite solid and consists of magma which is in a state of semi - perpetual convection. this convection process causes the lithospheric plates to move, albeit slowly. the resulting process is known as plate tectonics. areas of the crust where new crust is created are called divergent boundaries, those where it is brought back into the earth are convergent boundaries and those where plates slide past each other, but no new lithospheric material is created or destroyed, are referred to as transform ( or conservative ) boundaries. earthquakes result from the movement of the lithospheric plates, and they often occur near convergent boundaries where parts of the crust are forced into the earth as
genesis and its own history of development, a body with complex and multiform processes taking place within it. the soil is considered as different from bedrock. the latter becomes soil under the influence of a series of soil - formation factors ( climate, vegetation, country, relief and age ). according to him, soil should be called the " daily " or outward horizons of rocks regardless of the type ; they are changed naturally by the common effect of water, air and various kinds of living and dead organisms. a 1914 encyclopedic definition : " the different forms of earth on the surface of the rocks, formed by the breaking down or weathering of rocks ". serves to illustrate the historic view of soil which persisted from the 19th century. dokuchaev ' s late 19th century soil concept developed in the 20th century to one of soil as earthy material that has been altered by living processes. a corollary concept is that soil without a living component is simply a part of earth ' s outer layer. further refinement of the soil concept is occurring in view of an appreciation of energy transport and transformation within soil. the term is popularly applied to the material on the surface of the earth ' s moon and mars, a usage acceptable within a portion of the scientific community. accurate to this modern understanding of soil is nikiforoff ' s 1959 definition of soil as the " excited skin of the sub aerial part of the earth ' s crust ". = = areas of practice = = academically, soil scientists tend to be drawn to one of five areas of specialization : microbiology, pedology, edaphology, physics, or chemistry. yet the work specifics are very much dictated by the challenges facing our civilization ' s desire to sustain the land that supports it, and the distinctions between the sub - disciplines of soil science often blur in the process. soil science professionals commonly stay current in soil chemistry, soil physics, soil microbiology, pedology, and applied soil science in related disciplines. one exciting effort drawing in soil scientists in the u. s. as of 2004 is the soil quality initiative. central to the soil quality initiative is developing indices of soil health and then monitoring them in a way that gives us long - term ( decade - to - decade ) feedback on our performance as stewards of the planet. the effort includes understanding the functions of soil microbiotic crusts and exploring the potential to sequester atmospheric carbon in soil organic matter. relating the concept of agriculture to soil quality, however, has not
and nine classes, depending on its color, texture and hydrology. contemporaries friedrich albert fallou ( the german founder of modern soil science ) and vasily dokuchaev ( the russian founder of modern soil science ) are both credited with being among the first to identify soil as a resource whose distinctness and complexity deserved to be separated conceptually from geology and crop production and treated as a whole. as a founding father of soil science, fallou has primacy in time. fallou was working on the origins of soil before dokuchaev was born ; however dokuchaev ' s work was more extensive and is considered to be the more significant to modern soil theory than fallou ' s. previously, soil had been considered a product of chemical transformations of rocks, a dead substrate from which plants derive nutritious elements. soil and bedrock were in fact equated. dokuchaev considers the soil as a natural body having its own genesis and its own history of development, a body with complex and multiform processes taking place within it. the soil is considered as different from bedrock. the latter becomes soil under the influence of a series of soil - formation factors ( climate, vegetation, country, relief and age ). according to him, soil should be called the " daily " or outward horizons of rocks regardless of the type ; they are changed naturally by the common effect of water, air and various kinds of living and dead organisms. a 1914 encyclopedic definition : " the different forms of earth on the surface of the rocks, formed by the breaking down or weathering of rocks ". serves to illustrate the historic view of soil which persisted from the 19th century. dokuchaev ' s late 19th century soil concept developed in the 20th century to one of soil as earthy material that has been altered by living processes. a corollary concept is that soil without a living component is simply a part of earth ' s outer layer. further refinement of the soil concept is occurring in view of an appreciation of energy transport and transformation within soil. the term is popularly applied to the material on the surface of the earth ' s moon and mars, a usage acceptable within a portion of the scientific community. accurate to this modern understanding of soil is nikiforoff ' s 1959 definition of soil as the " excited skin of the sub aerial part of the earth ' s crust ". = = areas of practice = = academically, soil scientists tend to be drawn to one of five areas of specialization : microbiology
made of steel. the shoe is generally wider than the caisson to reduce friction, and the leading edge may be supplied with pressurised bentonite slurry, which swells in water, stabilizing settlement by filling depressions and voids. an open caisson may fill with water during sinking. the material is excavated by clamshell excavator bucket on crane. the formation level subsoil may still not be suitable for excavation or bearing capacity. the water in the caisson ( due to a high water table ) balances the upthrust forces of the soft soils underneath. if dewatered, the base may " pipe " or " boil ", causing the caisson to sink. to combat this problem, piles may be driven from the surface to act as : load - bearing walls, in that they transmit loads to deeper soils. anchors, in that they resist flotation because of the friction at the interface between their surfaces and the surrounding earth into which they have been driven. h - beam sections ( typical column sections, due to resistance to bending in all axis ) may be driven at angles " raked " to rock or other firmer soils ; the h - beams are left extended above the base. a reinforced concrete plug may be placed under the water, a process known as tremie concrete placement. when the caisson is dewatered, this plug acts as a pile cap, resisting the upward forces of the subsoil. = = = monolithic = = = a monolithic caisson ( or simply a monolith ) is larger than the other types of caisson, but similar to open caissons. such caissons are often found in quay walls, where resistance to impact from ships is required. = = = pneumatic = = = shallow caissons may be open to the air, whereas pneumatic caissons ( sometimes called pressurized caissons ), which penetrate soft mud, are bottomless boxes sealed at the top and filled with compressed air to keep water and mud out at depth. an airlock allows access to the chamber. workers, called sandhogs in american english, move mud and rock debris ( called muck ) from the edge of the workspace to a water - filled pit, connected by a tube ( called the muck tube ) to the surface. a crane at the surface removes the soil with a clamshell bucket. the water pressure in the tube balances the air pressure, with excess air escaping up
how it has changed over time. geochemistry studies the chemical components and processes of the earth. geophysics studies the physical properties of the earth. paleontology studies fossilized biological material in the lithosphere. planetary geology studies geoscience as it pertains to extraterrestrial bodies. geomorphology studies the origin of landscapes. structural geology studies the deformation of rocks to produce mountains and lowlands. resource geology studies how energy resources can be obtained from minerals. environmental geology studies how pollution and contaminants affect soil and rock. mineralogy is the study of minerals and includes the study of mineral formation, crystal structure, hazards associated with minerals, and the physical and chemical properties of minerals. petrology is the study of rocks, including the formation and composition of rocks. petrography is a branch of petrology that studies the typology and classification of rocks. = = earth ' s interior = = plate tectonics, mountain ranges, volcanoes, and earthquakes are geological phenomena that can be explained in terms of physical and chemical processes in the earth ' s crust. beneath the earth ' s crust lies the mantle which is heated by the radioactive decay of heavy elements. the mantle is not quite solid and consists of magma which is in a state of semi - perpetual convection. this convection process causes the lithospheric plates to move, albeit slowly. the resulting process is known as plate tectonics. areas of the crust where new crust is created are called divergent boundaries, those where it is brought back into the earth are convergent boundaries and those where plates slide past each other, but no new lithospheric material is created or destroyed, are referred to as transform ( or conservative ) boundaries. earthquakes result from the movement of the lithospheric plates, and they often occur near convergent boundaries where parts of the crust are forced into the earth as part of subduction. plate tectonics might be thought of as the process by which the earth is resurfaced. as the result of seafloor spreading, new crust and lithosphere is created by the flow of magma from the mantle to the near surface, through fissures, where it cools and solidifies. through subduction, oceanic crust and lithosphere vehemently returns to the convecting mantle. volcanoes result primarily from the melting of subducted crust material. crust material that is forced into the asthenosphere melts, and some portion of the melted material becomes light
of tool usage was found in ethiopia within the great rift valley, dating back to 2. 5 million years ago. the earliest methods of stone tool making, known as the oldowan " industry ", date back to at least 2. 3 million years ago. this era of stone tool use is called the paleolithic, or " old stone age ", and spans all of human history up to the development of agriculture approximately 12, 000 years ago. to make a stone tool, a " core " of hard stone with specific flaking properties ( such as flint ) was struck with a hammerstone. this flaking produced sharp edges which could be used as tools, primarily in the form of choppers or scrapers. these tools greatly aided the early humans in their hunter - gatherer lifestyle to perform a variety of tasks including butchering carcasses ( and breaking bones to get at the marrow ) ; chopping wood ; cracking open nuts ; skinning an animal for its hide, and even forming other tools out of softer materials such as bone and wood. the earliest stone tools were irrelevant, being little more than a fractured rock. in the acheulian era, beginning approximately 1. 65 million years ago, methods of working these stones into specific shapes, such as hand axes emerged. this early stone age is described as the lower paleolithic. the middle paleolithic, approximately 300, 000 years ago, saw the introduction of the prepared - core technique, where multiple blades could be rapidly formed from a single core stone. the upper paleolithic, beginning approximately 40, 000 years ago, saw the introduction of pressure flaking, where a wood, bone, or antler punch could be used to shape a stone very finely. the end of the last ice age about 10, 000 years ago is taken as the end point of the upper paleolithic and the beginning of the epipaleolithic / mesolithic. the mesolithic technology included the use of microliths as composite stone tools, along with wood, bone, and antler tools. the later stone age, during which the rudiments of agricultural technology were developed, is called the neolithic period. during this period, polished stone tools were made from a variety of hard rocks such as flint, jade, jadeite, and greenstone, largely by working exposures as quarries, but later the valuable rocks were pursued by tunneling underground, the first steps in mining technology. the polished axes were used for forest clearance and the establishment of crop
Question: What causes the most soil and rock erosion?
A) sun
B) wind
C) water
D) gravity
|
C) water
|
Context:
have evolved from the earliest emergence of life to present day. earth formed about 4. 5 billion years ago and all life on earth, both living and extinct, descended from a last universal common ancestor that lived about 3. 5 billion years ago. geologists have developed a geologic time scale that divides the history of the earth into major divisions, starting with four eons ( hadean, archean, proterozoic, and phanerozoic ), the first three of which are collectively known as the precambrian, which lasted approximately 4 billion years. each eon can be divided into eras, with the phanerozoic eon that began 539 million years ago being subdivided into paleozoic, mesozoic, and cenozoic eras. these three eras together comprise eleven periods ( cambrian, ordovician, silurian, devonian, carboniferous, permian, triassic, jurassic, cretaceous, tertiary, and quaternary ). the similarities among all known present - day species indicate that they have diverged through the process of evolution from their common ancestor. biologists regard the ubiquity of the genetic code as evidence of universal common descent for all bacteria, archaea, and eukaryotes. microbial mats of coexisting bacteria and archaea were the dominant form of life in the early archean eon and many of the major steps in early evolution are thought to have taken place in this environment. the earliest evidence of eukaryotes dates from 1. 85 billion years ago, and while they may have been present earlier, their diversification accelerated when they started using oxygen in their metabolism. later, around 1. 7 billion years ago, multicellular organisms began to appear, with differentiated cells performing specialised functions. algae - like multicellular land plants are dated back to about 1 billion years ago, although evidence suggests that microorganisms formed the earliest terrestrial ecosystems, at least 2. 7 billion years ago. microorganisms are thought to have paved the way for the inception of land plants in the ordovician period. land plants were so successful that they are thought to have contributed to the late devonian extinction event. ediacara biota appear during the ediacaran period, while vertebrates, along with most other modern phyla originated about 525 million years ago during the cambrian explosion. during the permian period, synapsids, including the ancestors of mammals, dominated the land, but most of this group became
, tertiary, and quaternary ). the similarities among all known present - day species indicate that they have diverged through the process of evolution from their common ancestor. biologists regard the ubiquity of the genetic code as evidence of universal common descent for all bacteria, archaea, and eukaryotes. microbial mats of coexisting bacteria and archaea were the dominant form of life in the early archean eon and many of the major steps in early evolution are thought to have taken place in this environment. the earliest evidence of eukaryotes dates from 1. 85 billion years ago, and while they may have been present earlier, their diversification accelerated when they started using oxygen in their metabolism. later, around 1. 7 billion years ago, multicellular organisms began to appear, with differentiated cells performing specialised functions. algae - like multicellular land plants are dated back to about 1 billion years ago, although evidence suggests that microorganisms formed the earliest terrestrial ecosystems, at least 2. 7 billion years ago. microorganisms are thought to have paved the way for the inception of land plants in the ordovician period. land plants were so successful that they are thought to have contributed to the late devonian extinction event. ediacara biota appear during the ediacaran period, while vertebrates, along with most other modern phyla originated about 525 million years ago during the cambrian explosion. during the permian period, synapsids, including the ancestors of mammals, dominated the land, but most of this group became extinct in the permian β triassic extinction event 252 million years ago. during the recovery from this catastrophe, archosaurs became the most abundant land vertebrates ; one archosaur group, the dinosaurs, dominated the jurassic and cretaceous periods. after the cretaceous β paleogene extinction event 66 million years ago killed off the non - avian dinosaurs, mammals increased rapidly in size and diversity. such mass extinctions may have accelerated evolution by providing opportunities for new groups of organisms to diversify. = = diversity = = = = = bacteria and archaea = = = bacteria are a type of cell that constitute a large domain of prokaryotic microorganisms. typically a few micrometers in length, bacteria have a number of shapes, ranging from spheres to rods and spirals. bacteria were among the first life forms to appear on earth, and are present in most of its habitats. bacteria inhabit soil, water, acidic hot springs, radioactive
##rozoic eon that began 539 million years ago being subdivided into paleozoic, mesozoic, and cenozoic eras. these three eras together comprise eleven periods ( cambrian, ordovician, silurian, devonian, carboniferous, permian, triassic, jurassic, cretaceous, tertiary, and quaternary ). the similarities among all known present - day species indicate that they have diverged through the process of evolution from their common ancestor. biologists regard the ubiquity of the genetic code as evidence of universal common descent for all bacteria, archaea, and eukaryotes. microbial mats of coexisting bacteria and archaea were the dominant form of life in the early archean eon and many of the major steps in early evolution are thought to have taken place in this environment. the earliest evidence of eukaryotes dates from 1. 85 billion years ago, and while they may have been present earlier, their diversification accelerated when they started using oxygen in their metabolism. later, around 1. 7 billion years ago, multicellular organisms began to appear, with differentiated cells performing specialised functions. algae - like multicellular land plants are dated back to about 1 billion years ago, although evidence suggests that microorganisms formed the earliest terrestrial ecosystems, at least 2. 7 billion years ago. microorganisms are thought to have paved the way for the inception of land plants in the ordovician period. land plants were so successful that they are thought to have contributed to the late devonian extinction event. ediacara biota appear during the ediacaran period, while vertebrates, along with most other modern phyla originated about 525 million years ago during the cambrian explosion. during the permian period, synapsids, including the ancestors of mammals, dominated the land, but most of this group became extinct in the permian β triassic extinction event 252 million years ago. during the recovery from this catastrophe, archosaurs became the most abundant land vertebrates ; one archosaur group, the dinosaurs, dominated the jurassic and cretaceous periods. after the cretaceous β paleogene extinction event 66 million years ago killed off the non - avian dinosaurs, mammals increased rapidly in size and diversity. such mass extinctions may have accelerated evolution by providing opportunities for new groups of organisms to diversify. = = diversity = = = = = bacteria and archaea = = = bacteria are a type of cell that constitute a large domain of prokar
one of the greatest discoveries of modern times is that of the expanding universe, almost invariably attributed to hubble ( 1929 ). what is not widely known is that the original treatise by lemaitre ( 1927 ) contained a rich fusion of both theory and of observation. stiglers law of eponymy is yet again affirmed : no scientific discovery is named after its original discoverer ( merton, 1957 ). an appeal is made for a lemaitre telescope, to honour the discoverer of the expanding universe.
the theory outright... lakatos sought to reconcile the rationalism of popperian falsificationism with what seemed to be its own refutation by history ". many philosophers have tried to solve the problem of demarcation in the following terms : a statement constitutes knowledge if sufficiently many people believe it sufficiently strongly. but the history of thought shows us that many people were totally committed to absurd beliefs. if the strengths of beliefs were a hallmark of knowledge, we should have to rank some tales about demons, angels, devils, and of heaven and hell as knowledge. scientists, on the other hand, are very sceptical even of their best theories. newton ' s is the most powerful theory science has yet produced, but newton himself never believed that bodies attract each other at a distance. so no degree of commitment to beliefs makes them knowledge. indeed, the hallmark of scientific behaviour is a certain scepticism even towards one ' s most cherished theories. blind commitment to a theory is not an intellectual virtue : it is an intellectual crime. thus a statement may be pseudoscientific even if it is eminently ' plausible ' and everybody believes in it, and it may be scientifically valuable even if it is unbelievable and nobody believes in it. a theory may even be of supreme scientific value even if no one understands it, let alone believes in it. the boundary between science and pseudoscience is disputed and difficult to determine analytically, even after more than a century of study by philosophers of science and scientists, and despite some basic agreements on the fundamentals of the scientific method. the concept of pseudoscience rests on an understanding that the scientific method has been misrepresented or misapplied with respect to a given theory, but many philosophers of science maintain that different kinds of methods are held as appropriate across different fields and different eras of human history. according to lakatos, the typical descriptive unit of great scientific achievements is not an isolated hypothesis but " a powerful problem - solving machinery, which, with the help of sophisticated mathematical techniques, digests anomalies and even turns them into positive evidence ". to popper, pseudoscience uses induction to generate theories, and only performs experiments to seek to verify them. to popper, falsifiability is what determines the scientific status of a theory. taking a historical approach, kuhn observed that scientists did not follow popper ' s rule, and might ignore falsifying data, unless overwhelming. to kuhn, puzzle - solving within
into major divisions, starting with four eons ( hadean, archean, proterozoic, and phanerozoic ), the first three of which are collectively known as the precambrian, which lasted approximately 4 billion years. each eon can be divided into eras, with the phanerozoic eon that began 539 million years ago being subdivided into paleozoic, mesozoic, and cenozoic eras. these three eras together comprise eleven periods ( cambrian, ordovician, silurian, devonian, carboniferous, permian, triassic, jurassic, cretaceous, tertiary, and quaternary ). the similarities among all known present - day species indicate that they have diverged through the process of evolution from their common ancestor. biologists regard the ubiquity of the genetic code as evidence of universal common descent for all bacteria, archaea, and eukaryotes. microbial mats of coexisting bacteria and archaea were the dominant form of life in the early archean eon and many of the major steps in early evolution are thought to have taken place in this environment. the earliest evidence of eukaryotes dates from 1. 85 billion years ago, and while they may have been present earlier, their diversification accelerated when they started using oxygen in their metabolism. later, around 1. 7 billion years ago, multicellular organisms began to appear, with differentiated cells performing specialised functions. algae - like multicellular land plants are dated back to about 1 billion years ago, although evidence suggests that microorganisms formed the earliest terrestrial ecosystems, at least 2. 7 billion years ago. microorganisms are thought to have paved the way for the inception of land plants in the ordovician period. land plants were so successful that they are thought to have contributed to the late devonian extinction event. ediacara biota appear during the ediacaran period, while vertebrates, along with most other modern phyla originated about 525 million years ago during the cambrian explosion. during the permian period, synapsids, including the ancestors of mammals, dominated the land, but most of this group became extinct in the permian β triassic extinction event 252 million years ago. during the recovery from this catastrophe, archosaurs became the most abundant land vertebrates ; one archosaur group, the dinosaurs, dominated the jurassic and cretaceous periods. after the cretaceous β paleogene extinction event 66 million years ago killed off
##sphere ( or lithosphere ). earth science can be considered to be a branch of planetary science but with a much older history. = = geology = = geology is broadly the study of earth ' s structure, substance, and processes. geology is largely the study of the lithosphere, or earth ' s surface, including the crust and rocks. it includes the physical characteristics and processes that occur in the lithosphere as well as how they are affected by geothermal energy. it incorporates aspects of chemistry, physics, and biology as elements of geology interact. historical geology is the application of geology to interpret earth history and how it has changed over time. geochemistry studies the chemical components and processes of the earth. geophysics studies the physical properties of the earth. paleontology studies fossilized biological material in the lithosphere. planetary geology studies geoscience as it pertains to extraterrestrial bodies. geomorphology studies the origin of landscapes. structural geology studies the deformation of rocks to produce mountains and lowlands. resource geology studies how energy resources can be obtained from minerals. environmental geology studies how pollution and contaminants affect soil and rock. mineralogy is the study of minerals and includes the study of mineral formation, crystal structure, hazards associated with minerals, and the physical and chemical properties of minerals. petrology is the study of rocks, including the formation and composition of rocks. petrography is a branch of petrology that studies the typology and classification of rocks. = = earth ' s interior = = plate tectonics, mountain ranges, volcanoes, and earthquakes are geological phenomena that can be explained in terms of physical and chemical processes in the earth ' s crust. beneath the earth ' s crust lies the mantle which is heated by the radioactive decay of heavy elements. the mantle is not quite solid and consists of magma which is in a state of semi - perpetual convection. this convection process causes the lithospheric plates to move, albeit slowly. the resulting process is known as plate tectonics. areas of the crust where new crust is created are called divergent boundaries, those where it is brought back into the earth are convergent boundaries and those where plates slide past each other, but no new lithospheric material is created or destroyed, are referred to as transform ( or conservative ) boundaries. earthquakes result from the movement of the lithospheric plates, and they often occur near convergent boundaries where parts of the crust are forced into the earth as
##hosphere ) and its historic development. major subdisciplines are mineralogy and petrology, geomorphology, paleontology, stratigraphy, structural geology, engineering geology, and sedimentology. physical geography focuses on geography as an earth science. physical geography is the study of earth ' s seasons, climate, atmosphere, soil, streams, landforms, and oceans. physical geography can be divided into several branches or related fields, as follows : geomorphology, biogeography, environmental geography, palaeogeography, climatology, meteorology, coastal geography, hydrology, ecology, glaciology. geophysics and geodesy investigate the shape of the earth, its reaction to forces and its magnetic and gravity fields. geophysicists explore the earth ' s core and mantle as well as the tectonic and seismic activity of the lithosphere. geophysics is commonly used to supplement the work of geologists in developing a comprehensive understanding of crustal geology, particularly in mineral and petroleum exploration. seismologists use geophysics to understand plate tectonic movement, as well as predict seismic activity. geochemistry studies the processes that control the abundance, composition, and distribution of chemical compounds and isotopes in geologic environments. geochemists use the tools and principles of chemistry to study the earth ' s composition, structure, processes, and other physical aspects. major subdisciplines are aqueous geochemistry, cosmochemistry, isotope geochemistry and biogeochemistry. soil science covers the outermost layer of the earth ' s crust that is subject to soil formation processes ( or pedosphere ). major subdivisions in this field of study include edaphology and pedology. ecology covers the interactions between organisms and their environment. this field of study differentiates the study of earth from other planets in the solar system, earth being the only planet teeming with life. hydrology, oceanography and limnology are studies which focus on the movement, distribution, and quality of the water and involve all the components of the hydrologic cycle on the earth and its atmosphere ( or hydrosphere ). " sub - disciplines of hydrology include hydrometeorology, surface water hydrology, hydrogeology, watershed science, forest hydrology, and water chemistry. " glaciology covers the icy parts of the earth ( or cryosphere ). atmospheric sciences cover the gaseous parts of the earth ( or atmosphere
##wi, turkana, dating from 3. 3 million years ago. stone tools diversified through the pleistocene period, which ended ~ 12, 000 years ago. the earliest evidence of warfare between two groups is recorded at the site of nataruk in turkana, kenya, where human skeletons with major traumatic injuries to the head, neck, ribs, knees and hands, including an embedded obsidian bladelet on a skull, are evidence of inter - group conflict between groups of nomadic hunter - gatherers 10, 000 years ago. humans entered the bronze age as they learned to smelt copper into an alloy with tin to make weapons. in asia where copper - tin ores are rare, this development was delayed until trading in bronze began in the third millennium bce. in the middle east and southern european regions, the bronze age follows the neolithic period, but in other parts of the world, the copper age is a transition from neolithic to the bronze age. although the iron age generally follows the bronze age, in some areas the iron age intrudes directly on the neolithic from outside the region, with the exception of sub - saharan africa where it was developed independently. the first large - scale use of iron weapons began in asia minor around the 14th century bce and in central europe around the 11th century bce followed by the middle east ( about 1000 bce ) and india and china. the assyrians are credited with the introduction of horse cavalry in warfare and the extensive use of iron weapons by 1100 bce. assyrians were also the first to use iron - tipped arrows. = = = post - classical technology = = = the wujing zongyao ( essentials of the military arts ), written by zeng gongliang, ding du, and others at the order of emperor renzong around 1043 during the song dynasty illustrate the eras focus on advancing intellectual issues and military technology due to the significance of warfare between the song and the liao, jin, and yuan to their north. the book covers topics of military strategy, training, and the production and employment of advanced weaponry. advances in military technology aided the song dynasty in its defense against hostile neighbors to the north. the flamethrower found its origins in byzantine - era greece, employing greek fire ( a chemically complex, highly flammable petrol fluid ) in a device with a siphon hose by the 7th century. : 77 the earliest reference to greek fire in china was made in 917, written by wu renchen in his spring and autumn annals of the ten kingdoms. : 80 in 91
##morphology studies the origin of landscapes. structural geology studies the deformation of rocks to produce mountains and lowlands. resource geology studies how energy resources can be obtained from minerals. environmental geology studies how pollution and contaminants affect soil and rock. mineralogy is the study of minerals and includes the study of mineral formation, crystal structure, hazards associated with minerals, and the physical and chemical properties of minerals. petrology is the study of rocks, including the formation and composition of rocks. petrography is a branch of petrology that studies the typology and classification of rocks. = = earth ' s interior = = plate tectonics, mountain ranges, volcanoes, and earthquakes are geological phenomena that can be explained in terms of physical and chemical processes in the earth ' s crust. beneath the earth ' s crust lies the mantle which is heated by the radioactive decay of heavy elements. the mantle is not quite solid and consists of magma which is in a state of semi - perpetual convection. this convection process causes the lithospheric plates to move, albeit slowly. the resulting process is known as plate tectonics. areas of the crust where new crust is created are called divergent boundaries, those where it is brought back into the earth are convergent boundaries and those where plates slide past each other, but no new lithospheric material is created or destroyed, are referred to as transform ( or conservative ) boundaries. earthquakes result from the movement of the lithospheric plates, and they often occur near convergent boundaries where parts of the crust are forced into the earth as part of subduction. plate tectonics might be thought of as the process by which the earth is resurfaced. as the result of seafloor spreading, new crust and lithosphere is created by the flow of magma from the mantle to the near surface, through fissures, where it cools and solidifies. through subduction, oceanic crust and lithosphere vehemently returns to the convecting mantle. volcanoes result primarily from the melting of subducted crust material. crust material that is forced into the asthenosphere melts, and some portion of the melted material becomes light enough to rise to the surface β giving birth to volcanoes. = = atmospheric science = = atmospheric science initially developed in the late - 19th century as a means to forecast the weather through meteorology, the study of weather. atmospheric chemistry was developed in the 20th century to measure air pollution and expanded in the 1970s in response to
Question: The theory of continental drift states that the land masses called continents were once closer together and have drifted apart over time. Which circumstance is the strongest evidence in support of this theory?
A) the migration patterns of birds
B) the periodic climate changes of Earth
C) the large variety of animals on each continent
D) the similarity of fossils on different continents
|
D) the similarity of fossils on different continents
|
Context:
c. 4000 bc, associated with the maadi culture. this represents the earliest evidence for smelting in africa. the varna necropolis, bulgaria, is a burial site located in the western industrial zone of varna, approximately 4 km from the city centre, internationally considered one of the key archaeological sites in world prehistory. the oldest gold treasure in the world, dating from 4, 600 bc to 4, 200 bc, was discovered at the site. the gold piece dating from 4, 500 bc, found in 2019 in durankulak, near varna is another important example. other signs of early metals are found from the third millennium bc in palmela, portugal, los millares, spain, and stonehenge, united kingdom. the precise beginnings, however, have not be clearly ascertained and new discoveries are both continuous and ongoing. in approximately 1900 bc, ancient iron smelting sites existed in tamil nadu. in the near east, about 3, 500 bc, it was discovered that by combining copper and tin, a superior metal could be made, an alloy called bronze. this represented a major technological shift known as the bronze age. the extraction of iron from its ore into a workable metal is much more difficult than for copper or tin. the process appears to have been invented by the hittites in about 1200 bc, beginning the iron age. the secret of extracting and working iron was a key factor in the success of the philistines. historical developments in ferrous metallurgy can be found in a wide variety of past cultures and civilizations. this includes the ancient and medieval kingdoms and empires of the middle east and near east, ancient iran, ancient egypt, ancient nubia, and anatolia in present - day turkey, ancient nok, carthage, the celts, greeks and romans of ancient europe, medieval europe, ancient and medieval china, ancient and medieval india, ancient and medieval japan, amongst others. a 16th century book by georg agricola, de re metallica, describes the highly developed and complex processes of mining metal ores, metal extraction, and metallurgy of the time. agricola has been described as the " father of metallurgy ". = = extraction = = extractive metallurgy is the practice of removing valuable metals from an ore and refining the extracted raw metals into a purer form. in order to convert a metal oxide or sulphide to a purer metal, the ore must be reduced physically, chemically, or electroly
prehistory. the oldest gold treasure in the world, dating from 4, 600 bc to 4, 200 bc, was discovered at the site. the gold piece dating from 4, 500 bc, found in 2019 in durankulak, near varna is another important example. other signs of early metals are found from the third millennium bc in palmela, portugal, los millares, spain, and stonehenge, united kingdom. the precise beginnings, however, have not be clearly ascertained and new discoveries are both continuous and ongoing. in approximately 1900 bc, ancient iron smelting sites existed in tamil nadu. in the near east, about 3, 500 bc, it was discovered that by combining copper and tin, a superior metal could be made, an alloy called bronze. this represented a major technological shift known as the bronze age. the extraction of iron from its ore into a workable metal is much more difficult than for copper or tin. the process appears to have been invented by the hittites in about 1200 bc, beginning the iron age. the secret of extracting and working iron was a key factor in the success of the philistines. historical developments in ferrous metallurgy can be found in a wide variety of past cultures and civilizations. this includes the ancient and medieval kingdoms and empires of the middle east and near east, ancient iran, ancient egypt, ancient nubia, and anatolia in present - day turkey, ancient nok, carthage, the celts, greeks and romans of ancient europe, medieval europe, ancient and medieval china, ancient and medieval india, ancient and medieval japan, amongst others. a 16th century book by georg agricola, de re metallica, describes the highly developed and complex processes of mining metal ores, metal extraction, and metallurgy of the time. agricola has been described as the " father of metallurgy ". = = extraction = = extractive metallurgy is the practice of removing valuable metals from an ore and refining the extracted raw metals into a purer form. in order to convert a metal oxide or sulphide to a purer metal, the ore must be reduced physically, chemically, or electrolytically. extractive metallurgists are interested in three primary streams : feed, concentrate ( metal oxide / sulphide ) and tailings ( waste ). after mining, large pieces of the ore feed are broken through crushing or grinding in order to obtain particles small enough, where each particle is either mostly valuable or
other contemporary production centre. the earliest documented use of lead ( possibly native or smelted ) in the near east dates from the 6th millennium bc, is from the late neolithic settlements of yarim tepe and arpachiyah in iraq. the artifacts suggest that lead smelting may have predated copper smelting. metallurgy of lead has also been found in the balkans during the same period. copper smelting is documented at sites in anatolia and at the site of tal - i iblis in southeastern iran from c. 5000 bc. copper smelting is first documented in the delta region of northern egypt in c. 4000 bc, associated with the maadi culture. this represents the earliest evidence for smelting in africa. the varna necropolis, bulgaria, is a burial site located in the western industrial zone of varna, approximately 4 km from the city centre, internationally considered one of the key archaeological sites in world prehistory. the oldest gold treasure in the world, dating from 4, 600 bc to 4, 200 bc, was discovered at the site. the gold piece dating from 4, 500 bc, found in 2019 in durankulak, near varna is another important example. other signs of early metals are found from the third millennium bc in palmela, portugal, los millares, spain, and stonehenge, united kingdom. the precise beginnings, however, have not be clearly ascertained and new discoveries are both continuous and ongoing. in approximately 1900 bc, ancient iron smelting sites existed in tamil nadu. in the near east, about 3, 500 bc, it was discovered that by combining copper and tin, a superior metal could be made, an alloy called bronze. this represented a major technological shift known as the bronze age. the extraction of iron from its ore into a workable metal is much more difficult than for copper or tin. the process appears to have been invented by the hittites in about 1200 bc, beginning the iron age. the secret of extracting and working iron was a key factor in the success of the philistines. historical developments in ferrous metallurgy can be found in a wide variety of past cultures and civilizations. this includes the ancient and medieval kingdoms and empires of the middle east and near east, ancient iran, ancient egypt, ancient nubia, and anatolia in present - day turkey, ancient nok, carthage, the celts, greeks and romans of ancient europe, medieval europe, ancient and medieval china, ancient and
##rozoic eon that began 539 million years ago being subdivided into paleozoic, mesozoic, and cenozoic eras. these three eras together comprise eleven periods ( cambrian, ordovician, silurian, devonian, carboniferous, permian, triassic, jurassic, cretaceous, tertiary, and quaternary ). the similarities among all known present - day species indicate that they have diverged through the process of evolution from their common ancestor. biologists regard the ubiquity of the genetic code as evidence of universal common descent for all bacteria, archaea, and eukaryotes. microbial mats of coexisting bacteria and archaea were the dominant form of life in the early archean eon and many of the major steps in early evolution are thought to have taken place in this environment. the earliest evidence of eukaryotes dates from 1. 85 billion years ago, and while they may have been present earlier, their diversification accelerated when they started using oxygen in their metabolism. later, around 1. 7 billion years ago, multicellular organisms began to appear, with differentiated cells performing specialised functions. algae - like multicellular land plants are dated back to about 1 billion years ago, although evidence suggests that microorganisms formed the earliest terrestrial ecosystems, at least 2. 7 billion years ago. microorganisms are thought to have paved the way for the inception of land plants in the ordovician period. land plants were so successful that they are thought to have contributed to the late devonian extinction event. ediacara biota appear during the ediacaran period, while vertebrates, along with most other modern phyla originated about 525 million years ago during the cambrian explosion. during the permian period, synapsids, including the ancestors of mammals, dominated the land, but most of this group became extinct in the permian β triassic extinction event 252 million years ago. during the recovery from this catastrophe, archosaurs became the most abundant land vertebrates ; one archosaur group, the dinosaurs, dominated the jurassic and cretaceous periods. after the cretaceous β paleogene extinction event 66 million years ago killed off the non - avian dinosaurs, mammals increased rapidly in size and diversity. such mass extinctions may have accelerated evolution by providing opportunities for new groups of organisms to diversify. = = diversity = = = = = bacteria and archaea = = = bacteria are a type of cell that constitute a large domain of prokar
##elting. metallurgy of lead has also been found in the balkans during the same period. copper smelting is documented at sites in anatolia and at the site of tal - i iblis in southeastern iran from c. 5000 bc. copper smelting is first documented in the delta region of northern egypt in c. 4000 bc, associated with the maadi culture. this represents the earliest evidence for smelting in africa. the varna necropolis, bulgaria, is a burial site located in the western industrial zone of varna, approximately 4 km from the city centre, internationally considered one of the key archaeological sites in world prehistory. the oldest gold treasure in the world, dating from 4, 600 bc to 4, 200 bc, was discovered at the site. the gold piece dating from 4, 500 bc, found in 2019 in durankulak, near varna is another important example. other signs of early metals are found from the third millennium bc in palmela, portugal, los millares, spain, and stonehenge, united kingdom. the precise beginnings, however, have not be clearly ascertained and new discoveries are both continuous and ongoing. in approximately 1900 bc, ancient iron smelting sites existed in tamil nadu. in the near east, about 3, 500 bc, it was discovered that by combining copper and tin, a superior metal could be made, an alloy called bronze. this represented a major technological shift known as the bronze age. the extraction of iron from its ore into a workable metal is much more difficult than for copper or tin. the process appears to have been invented by the hittites in about 1200 bc, beginning the iron age. the secret of extracting and working iron was a key factor in the success of the philistines. historical developments in ferrous metallurgy can be found in a wide variety of past cultures and civilizations. this includes the ancient and medieval kingdoms and empires of the middle east and near east, ancient iran, ancient egypt, ancient nubia, and anatolia in present - day turkey, ancient nok, carthage, the celts, greeks and romans of ancient europe, medieval europe, ancient and medieval china, ancient and medieval india, ancient and medieval japan, amongst others. a 16th century book by georg agricola, de re metallica, describes the highly developed and complex processes of mining metal ores, metal extraction, and metallurgy of the time. agricola has been described as the " father of metallurgy
the third millennium bc in palmela, portugal, los millares, spain, and stonehenge, united kingdom. the precise beginnings, however, have not be clearly ascertained and new discoveries are both continuous and ongoing. in approximately 1900 bc, ancient iron smelting sites existed in tamil nadu. in the near east, about 3, 500 bc, it was discovered that by combining copper and tin, a superior metal could be made, an alloy called bronze. this represented a major technological shift known as the bronze age. the extraction of iron from its ore into a workable metal is much more difficult than for copper or tin. the process appears to have been invented by the hittites in about 1200 bc, beginning the iron age. the secret of extracting and working iron was a key factor in the success of the philistines. historical developments in ferrous metallurgy can be found in a wide variety of past cultures and civilizations. this includes the ancient and medieval kingdoms and empires of the middle east and near east, ancient iran, ancient egypt, ancient nubia, and anatolia in present - day turkey, ancient nok, carthage, the celts, greeks and romans of ancient europe, medieval europe, ancient and medieval china, ancient and medieval india, ancient and medieval japan, amongst others. a 16th century book by georg agricola, de re metallica, describes the highly developed and complex processes of mining metal ores, metal extraction, and metallurgy of the time. agricola has been described as the " father of metallurgy ". = = extraction = = extractive metallurgy is the practice of removing valuable metals from an ore and refining the extracted raw metals into a purer form. in order to convert a metal oxide or sulphide to a purer metal, the ore must be reduced physically, chemically, or electrolytically. extractive metallurgists are interested in three primary streams : feed, concentrate ( metal oxide / sulphide ) and tailings ( waste ). after mining, large pieces of the ore feed are broken through crushing or grinding in order to obtain particles small enough, where each particle is either mostly valuable or mostly waste. concentrating the particles of value in a form supporting separation enables the desired metal to be removed from waste products. mining may not be necessary, if the ore body and physical environment are conducive to leaching. leaching dissolves minerals in an ore body and results in an enriched solution. the solution
, tertiary, and quaternary ). the similarities among all known present - day species indicate that they have diverged through the process of evolution from their common ancestor. biologists regard the ubiquity of the genetic code as evidence of universal common descent for all bacteria, archaea, and eukaryotes. microbial mats of coexisting bacteria and archaea were the dominant form of life in the early archean eon and many of the major steps in early evolution are thought to have taken place in this environment. the earliest evidence of eukaryotes dates from 1. 85 billion years ago, and while they may have been present earlier, their diversification accelerated when they started using oxygen in their metabolism. later, around 1. 7 billion years ago, multicellular organisms began to appear, with differentiated cells performing specialised functions. algae - like multicellular land plants are dated back to about 1 billion years ago, although evidence suggests that microorganisms formed the earliest terrestrial ecosystems, at least 2. 7 billion years ago. microorganisms are thought to have paved the way for the inception of land plants in the ordovician period. land plants were so successful that they are thought to have contributed to the late devonian extinction event. ediacara biota appear during the ediacaran period, while vertebrates, along with most other modern phyla originated about 525 million years ago during the cambrian explosion. during the permian period, synapsids, including the ancestors of mammals, dominated the land, but most of this group became extinct in the permian β triassic extinction event 252 million years ago. during the recovery from this catastrophe, archosaurs became the most abundant land vertebrates ; one archosaur group, the dinosaurs, dominated the jurassic and cretaceous periods. after the cretaceous β paleogene extinction event 66 million years ago killed off the non - avian dinosaurs, mammals increased rapidly in size and diversity. such mass extinctions may have accelerated evolution by providing opportunities for new groups of organisms to diversify. = = diversity = = = = = bacteria and archaea = = = bacteria are a type of cell that constitute a large domain of prokaryotic microorganisms. typically a few micrometers in length, bacteria have a number of shapes, ranging from spheres to rods and spirals. bacteria were among the first life forms to appear on earth, and are present in most of its habitats. bacteria inhabit soil, water, acidic hot springs, radioactive
. microbial mats of coexisting bacteria and archaea were the dominant form of life in the early archean eon and many of the major steps in early evolution are thought to have taken place in this environment. the earliest evidence of eukaryotes dates from 1. 85 billion years ago, and while they may have been present earlier, their diversification accelerated when they started using oxygen in their metabolism. later, around 1. 7 billion years ago, multicellular organisms began to appear, with differentiated cells performing specialised functions. algae - like multicellular land plants are dated back to about 1 billion years ago, although evidence suggests that microorganisms formed the earliest terrestrial ecosystems, at least 2. 7 billion years ago. microorganisms are thought to have paved the way for the inception of land plants in the ordovician period. land plants were so successful that they are thought to have contributed to the late devonian extinction event. ediacara biota appear during the ediacaran period, while vertebrates, along with most other modern phyla originated about 525 million years ago during the cambrian explosion. during the permian period, synapsids, including the ancestors of mammals, dominated the land, but most of this group became extinct in the permian β triassic extinction event 252 million years ago. during the recovery from this catastrophe, archosaurs became the most abundant land vertebrates ; one archosaur group, the dinosaurs, dominated the jurassic and cretaceous periods. after the cretaceous β paleogene extinction event 66 million years ago killed off the non - avian dinosaurs, mammals increased rapidly in size and diversity. such mass extinctions may have accelerated evolution by providing opportunities for new groups of organisms to diversify. = = diversity = = = = = bacteria and archaea = = = bacteria are a type of cell that constitute a large domain of prokaryotic microorganisms. typically a few micrometers in length, bacteria have a number of shapes, ranging from spheres to rods and spirals. bacteria were among the first life forms to appear on earth, and are present in most of its habitats. bacteria inhabit soil, water, acidic hot springs, radioactive waste, and the deep biosphere of the earth ' s crust. bacteria also live in symbiotic and parasitic relationships with plants and animals. most bacteria have not been characterised, and only about 27 percent of the bacterial phyla have species that can be grown in the laboratory. archaea constitute the other domain of
they may have been present earlier, their diversification accelerated when they started using oxygen in their metabolism. later, around 1. 7 billion years ago, multicellular organisms began to appear, with differentiated cells performing specialised functions. algae - like multicellular land plants are dated back to about 1 billion years ago, although evidence suggests that microorganisms formed the earliest terrestrial ecosystems, at least 2. 7 billion years ago. microorganisms are thought to have paved the way for the inception of land plants in the ordovician period. land plants were so successful that they are thought to have contributed to the late devonian extinction event. ediacara biota appear during the ediacaran period, while vertebrates, along with most other modern phyla originated about 525 million years ago during the cambrian explosion. during the permian period, synapsids, including the ancestors of mammals, dominated the land, but most of this group became extinct in the permian β triassic extinction event 252 million years ago. during the recovery from this catastrophe, archosaurs became the most abundant land vertebrates ; one archosaur group, the dinosaurs, dominated the jurassic and cretaceous periods. after the cretaceous β paleogene extinction event 66 million years ago killed off the non - avian dinosaurs, mammals increased rapidly in size and diversity. such mass extinctions may have accelerated evolution by providing opportunities for new groups of organisms to diversify. = = diversity = = = = = bacteria and archaea = = = bacteria are a type of cell that constitute a large domain of prokaryotic microorganisms. typically a few micrometers in length, bacteria have a number of shapes, ranging from spheres to rods and spirals. bacteria were among the first life forms to appear on earth, and are present in most of its habitats. bacteria inhabit soil, water, acidic hot springs, radioactive waste, and the deep biosphere of the earth ' s crust. bacteria also live in symbiotic and parasitic relationships with plants and animals. most bacteria have not been characterised, and only about 27 percent of the bacterial phyla have species that can be grown in the laboratory. archaea constitute the other domain of prokaryotic cells and were initially classified as bacteria, receiving the name archaebacteria ( in the archaebacteria kingdom ), a term that has fallen out of use. archaeal cells have unique properties separating them from the other two domains, bacteria and eukaryota. archaea
have evolved from the earliest emergence of life to present day. earth formed about 4. 5 billion years ago and all life on earth, both living and extinct, descended from a last universal common ancestor that lived about 3. 5 billion years ago. geologists have developed a geologic time scale that divides the history of the earth into major divisions, starting with four eons ( hadean, archean, proterozoic, and phanerozoic ), the first three of which are collectively known as the precambrian, which lasted approximately 4 billion years. each eon can be divided into eras, with the phanerozoic eon that began 539 million years ago being subdivided into paleozoic, mesozoic, and cenozoic eras. these three eras together comprise eleven periods ( cambrian, ordovician, silurian, devonian, carboniferous, permian, triassic, jurassic, cretaceous, tertiary, and quaternary ). the similarities among all known present - day species indicate that they have diverged through the process of evolution from their common ancestor. biologists regard the ubiquity of the genetic code as evidence of universal common descent for all bacteria, archaea, and eukaryotes. microbial mats of coexisting bacteria and archaea were the dominant form of life in the early archean eon and many of the major steps in early evolution are thought to have taken place in this environment. the earliest evidence of eukaryotes dates from 1. 85 billion years ago, and while they may have been present earlier, their diversification accelerated when they started using oxygen in their metabolism. later, around 1. 7 billion years ago, multicellular organisms began to appear, with differentiated cells performing specialised functions. algae - like multicellular land plants are dated back to about 1 billion years ago, although evidence suggests that microorganisms formed the earliest terrestrial ecosystems, at least 2. 7 billion years ago. microorganisms are thought to have paved the way for the inception of land plants in the ordovician period. land plants were so successful that they are thought to have contributed to the late devonian extinction event. ediacara biota appear during the ediacaran period, while vertebrates, along with most other modern phyla originated about 525 million years ago during the cambrian explosion. during the permian period, synapsids, including the ancestors of mammals, dominated the land, but most of this group became
Question: Rich beds of marine fossils have been found in Florida. By dating these fossils, scientists can best determine
A) when Florida collided with the North American continent.
B) the date when organisms moved onto dry land.
C) the period during which Florida was covered in water.
D) when the atmosphere of Earth was more humid than it is today.
|
C) the period during which Florida was covered in water.
|
Context:
stems mainly provide support to the leaves and reproductive structures, but can store water in succulent plants such as cacti, food as in potato tubers, or reproduce vegetatively as in the stolons of strawberry plants or in the process of layering. leaves gather sunlight and carry out photosynthesis. large, flat, flexible, green leaves are called foliage leaves. gymnosperms, such as conifers, cycads, ginkgo, and gnetophytes are seed - producing plants with open seeds. angiosperms are seed - producing plants that produce flowers and have enclosed seeds. woody plants, such as azaleas and oaks, undergo a secondary growth phase resulting in two additional types of tissues : wood ( secondary xylem ) and bark ( secondary phloem and cork ). all gymnosperms and many angiosperms are woody plants. some plants reproduce sexually, some asexually, and some via both means. although reference to major morphological categories such as root, stem, leaf, and trichome are useful, one has to keep in mind that these categories are linked through intermediate forms so that a continuum between the categories results. furthermore, structures can be seen as processes, that is, process combinations. = = systematic botany = = systematic botany is part of systematic biology, which is concerned with the range and diversity of organisms and their relationships, particularly as determined by their evolutionary history. it involves, or is related to, biological classification, scientific taxonomy and phylogenetics. biological classification is the method by which botanists group organisms into categories such as genera or species. biological classification is a form of scientific taxonomy. modern taxonomy is rooted in the work of carl linnaeus, who grouped species according to shared physical characteristics. these groupings have since been revised to align better with the darwinian principle of common descent β grouping organisms by ancestry rather than superficial characteristics. while scientists do not always agree on how to classify organisms, molecular phylogenetics, which uses dna sequences as data, has driven many recent revisions along evolutionary lines and is likely to continue to do so. the dominant classification system is called linnaean taxonomy. it includes ranks and binomial nomenclature. the nomenclature of botanical organisms is codified in the international code of nomenclature for algae, fungi, and plants ( icn ) and administered by the international botanical congress. kingdom plantae belongs to domain eukaryota and is broken down recursively until each species is separately classified. the order is :
unspecialised cells ) that can grow into a new plant. in vascular plants, the xylem and phloem are the conductive tissues that transport resources between shoots and roots. roots are often adapted to store food such as sugars or starch, as in sugar beets and carrots. stems mainly provide support to the leaves and reproductive structures, but can store water in succulent plants such as cacti, food as in potato tubers, or reproduce vegetatively as in the stolons of strawberry plants or in the process of layering. leaves gather sunlight and carry out photosynthesis. large, flat, flexible, green leaves are called foliage leaves. gymnosperms, such as conifers, cycads, ginkgo, and gnetophytes are seed - producing plants with open seeds. angiosperms are seed - producing plants that produce flowers and have enclosed seeds. woody plants, such as azaleas and oaks, undergo a secondary growth phase resulting in two additional types of tissues : wood ( secondary xylem ) and bark ( secondary phloem and cork ). all gymnosperms and many angiosperms are woody plants. some plants reproduce sexually, some asexually, and some via both means. although reference to major morphological categories such as root, stem, leaf, and trichome are useful, one has to keep in mind that these categories are linked through intermediate forms so that a continuum between the categories results. furthermore, structures can be seen as processes, that is, process combinations. = = systematic botany = = systematic botany is part of systematic biology, which is concerned with the range and diversity of organisms and their relationships, particularly as determined by their evolutionary history. it involves, or is related to, biological classification, scientific taxonomy and phylogenetics. biological classification is the method by which botanists group organisms into categories such as genera or species. biological classification is a form of scientific taxonomy. modern taxonomy is rooted in the work of carl linnaeus, who grouped species according to shared physical characteristics. these groupings have since been revised to align better with the darwinian principle of common descent β grouping organisms by ancestry rather than superficial characteristics. while scientists do not always agree on how to classify organisms, molecular phylogenetics, which uses dna sequences as data, has driven many recent revisions along evolutionary lines and is likely to continue to do so. the dominant classification system is called linnaean taxonomy. it includes ranks and binomi
the best - suited crops ( e. g., those with the highest yields ) to produce enough food to support a growing population. as crops and fields became increasingly large and difficult to maintain, it was discovered that specific organisms and their by - products could effectively fertilize, restore nitrogen, and control pests. throughout the history of agriculture, farmers have inadvertently altered the genetics of their crops through introducing them to new environments and breeding them with other plants β one of the first forms of biotechnology. these processes also were included in early fermentation of beer. these processes were introduced in early mesopotamia, egypt, china and india, and still use the same basic biological methods. in brewing, malted grains ( containing enzymes ) convert starch from grains into sugar and then adding specific yeasts to produce beer. in this process, carbohydrates in the grains broke down into alcohols, such as ethanol. later, other cultures produced the process of lactic acid fermentation, which produced other preserved foods, such as soy sauce. fermentation was also used in this time period to produce leavened bread. although the process of fermentation was not fully understood until louis pasteur ' s work in 1857, it is still the first use of biotechnology to convert a food source into another form. before the time of charles darwin ' s work and life, animal and plant scientists had already used selective breeding. darwin added to that body of work with his scientific observations about the ability of science to change species. these accounts contributed to darwin ' s theory of natural selection. for thousands of years, humans have used selective breeding to improve the production of crops and livestock to use them for food. in selective breeding, organisms with desirable characteristics are mated to produce offspring with the same characteristics. for example, this technique was used with corn to produce the largest and sweetest crops. in the early twentieth century scientists gained a greater understanding of microbiology and explored ways of manufacturing specific products. in 1917, chaim weizmann first used a pure microbiological culture in an industrial process, that of manufacturing corn starch using clostridium acetobutylicum, to produce acetone, which the united kingdom desperately needed to manufacture explosives during world war i. biotechnology has also led to the development of antibiotics. in 1928, alexander fleming discovered the mold penicillium. his work led to the purification of the antibiotic formed by the mold by howard florey, ernst boris chain and norman heatley β to form
masculinity and warmth. the five phases β fire, earth, metal, wood, and water β described a cycle of transformations in nature. the water turned into wood, which turned into the fire when it burned. the ashes left by fire were earth. using these principles, chinese philosophers and doctors explored human anatomy, characterizing organs as predominantly yin or yang, and understood the relationship between the pulse, the heart, and the flow of blood in the body centuries before it became accepted in the west. little evidence survives of how ancient indian cultures around the indus river understood nature, but some of their perspectives may be reflected in the vedas, a set of sacred hindu texts. they reveal a conception of the universe as ever - expanding and constantly being recycled and reformed. surgeons in the ayurvedic tradition saw health and illness as a combination of three humors : wind, bile and phlegm. a healthy life resulted from a balance among these humors. in ayurvedic thought, the body consisted of five elements : earth, water, fire, wind, and space. ayurvedic surgeons performed complex surgeries and developed a detailed understanding of human anatomy. pre - socratic philosophers in ancient greek culture brought natural philosophy a step closer to direct inquiry about cause and effect in nature between 600 and 400 bc. however, an element of magic and mythology remained. natural phenomena such as earthquakes and eclipses were explained increasingly in the context of nature itself instead of being attributed to angry gods. thales of miletus, an early philosopher who lived from 625 to 546 bc, explained earthquakes by theorizing that the world floated on water and that water was the fundamental element in nature. in the 5th century bc, leucippus was an early exponent of atomism, the idea that the world is made up of fundamental indivisible particles. pythagoras applied greek innovations in mathematics to astronomy and suggested that the earth was spherical. = = = aristotelian natural philosophy ( 400 bc β 1100 ad ) = = = later socratic and platonic thought focused on ethics, morals, and art and did not attempt an investigation of the physical world ; plato criticized pre - socratic thinkers as materialists and anti - religionists. aristotle, however, a student of plato who lived from 384 to 322 bc, paid closer attention to the natural world in his philosophy. in his history of animals, he described the inner workings of 110 species, including the stingray, catfish and
. throughout the history of agriculture, farmers have inadvertently altered the genetics of their crops through introducing them to new environments and breeding them with other plants β one of the first forms of biotechnology. these processes also were included in early fermentation of beer. these processes were introduced in early mesopotamia, egypt, china and india, and still use the same basic biological methods. in brewing, malted grains ( containing enzymes ) convert starch from grains into sugar and then adding specific yeasts to produce beer. in this process, carbohydrates in the grains broke down into alcohols, such as ethanol. later, other cultures produced the process of lactic acid fermentation, which produced other preserved foods, such as soy sauce. fermentation was also used in this time period to produce leavened bread. although the process of fermentation was not fully understood until louis pasteur ' s work in 1857, it is still the first use of biotechnology to convert a food source into another form. before the time of charles darwin ' s work and life, animal and plant scientists had already used selective breeding. darwin added to that body of work with his scientific observations about the ability of science to change species. these accounts contributed to darwin ' s theory of natural selection. for thousands of years, humans have used selective breeding to improve the production of crops and livestock to use them for food. in selective breeding, organisms with desirable characteristics are mated to produce offspring with the same characteristics. for example, this technique was used with corn to produce the largest and sweetest crops. in the early twentieth century scientists gained a greater understanding of microbiology and explored ways of manufacturing specific products. in 1917, chaim weizmann first used a pure microbiological culture in an industrial process, that of manufacturing corn starch using clostridium acetobutylicum, to produce acetone, which the united kingdom desperately needed to manufacture explosives during world war i. biotechnology has also led to the development of antibiotics. in 1928, alexander fleming discovered the mold penicillium. his work led to the purification of the antibiotic formed by the mold by howard florey, ernst boris chain and norman heatley β to form what we today know as penicillin. in 1940, penicillin became available for medicinal use to treat bacterial infections in humans. the field of modern biotechnology is generally thought of as having been born in 1971 when paul berg ' s ( stanford ) experiments in gene splicing had early success. herbert w. boyer
the broad definition of " utilizing a biotechnological system to make products ". indeed, the cultivation of plants may be viewed as the earliest biotechnological enterprise. agriculture has been theorized to have become the dominant way of producing food since the neolithic revolution. through early biotechnology, the earliest farmers selected and bred the best - suited crops ( e. g., those with the highest yields ) to produce enough food to support a growing population. as crops and fields became increasingly large and difficult to maintain, it was discovered that specific organisms and their by - products could effectively fertilize, restore nitrogen, and control pests. throughout the history of agriculture, farmers have inadvertently altered the genetics of their crops through introducing them to new environments and breeding them with other plants β one of the first forms of biotechnology. these processes also were included in early fermentation of beer. these processes were introduced in early mesopotamia, egypt, china and india, and still use the same basic biological methods. in brewing, malted grains ( containing enzymes ) convert starch from grains into sugar and then adding specific yeasts to produce beer. in this process, carbohydrates in the grains broke down into alcohols, such as ethanol. later, other cultures produced the process of lactic acid fermentation, which produced other preserved foods, such as soy sauce. fermentation was also used in this time period to produce leavened bread. although the process of fermentation was not fully understood until louis pasteur ' s work in 1857, it is still the first use of biotechnology to convert a food source into another form. before the time of charles darwin ' s work and life, animal and plant scientists had already used selective breeding. darwin added to that body of work with his scientific observations about the ability of science to change species. these accounts contributed to darwin ' s theory of natural selection. for thousands of years, humans have used selective breeding to improve the production of crops and livestock to use them for food. in selective breeding, organisms with desirable characteristics are mated to produce offspring with the same characteristics. for example, this technique was used with corn to produce the largest and sweetest crops. in the early twentieth century scientists gained a greater understanding of microbiology and explored ways of manufacturing specific products. in 1917, chaim weizmann first used a pure microbiological culture in an industrial process, that of manufacturing corn starch using clostridium acetobutylicum, to produce acetone, which the united
the strictest definition of " plant " includes only the " land plants " or embryophytes, which include seed plants ( gymnosperms, including the pines, and flowering plants ) and the free - sporing cryptogams including ferns, clubmosses, liverworts, hornworts and mosses. embryophytes are multicellular eukaryotes descended from an ancestor that obtained its energy from sunlight by photosynthesis. they have life cycles with alternating haploid and diploid phases. the sexual haploid phase of embryophytes, known as the gametophyte, nurtures the developing diploid embryo sporophyte within its tissues for at least part of its life, even in the seed plants, where the gametophyte itself is nurtured by its parent sporophyte. other groups of organisms that were previously studied by botanists include bacteria ( now studied in bacteriology ), fungi ( mycology ) β including lichen - forming fungi ( lichenology ), non - chlorophyte algae ( phycology ), and viruses ( virology ). however, attention is still given to these groups by botanists, and fungi ( including lichens ) and photosynthetic protists are usually covered in introductory botany courses. palaeobotanists study ancient plants in the fossil record to provide information about the evolutionary history of plants. cyanobacteria, the first oxygen - releasing photosynthetic organisms on earth, are thought to have given rise to the ancestor of plants by entering into an endosymbiotic relationship with an early eukaryote, ultimately becoming the chloroplasts in plant cells. the new photosynthetic plants ( along with their algal relatives ) accelerated the rise in atmospheric oxygen started by the cyanobacteria, changing the ancient oxygen - free, reducing, atmosphere to one in which free oxygen has been abundant for more than 2 billion years. among the important botanical questions of the 21st century are the role of plants as primary producers in the global cycling of life ' s basic ingredients : energy, carbon, oxygen, nitrogen and water, and ways that our plant stewardship can help address the global environmental issues of resource management, conservation, human food security, biologically invasive organisms, carbon sequestration, climate change, and sustainability. = = = human nutrition = = = virtually all staple foods come either directly from primary production by plants, or indirectly from animals that
##ian period, several groups, including the lycopods, sphenophylls and progymnosperms, had independently evolved " megaspory " β their spores were of two distinct sizes, larger megaspores and smaller microspores. their reduced gametophytes developed from megaspores retained within the spore - producing organs ( megasporangia ) of the sporophyte, a condition known as endospory. seeds consist of an endosporic megasporangium surrounded by one or two sheathing layers ( integuments ). the young sporophyte develops within the seed, which on germination splits to release it. the earliest known seed plants date from the latest devonian famennian stage. following the evolution of the seed habit, seed plants diversified, giving rise to a number of now - extinct groups, including seed ferns, as well as the modern gymnosperms and angiosperms. gymnosperms produce " naked seeds " not fully enclosed in an ovary ; modern representatives include conifers, cycads, ginkgo, and gnetales. angiosperms produce seeds enclosed in a structure such as a carpel or an ovary. ongoing research on the molecular phylogenetics of living plants appears to show that the angiosperms are a sister clade to the gymnosperms. = = plant physiology = = plant physiology encompasses all the internal chemical and physical activities of plants associated with life. chemicals obtained from the air, soil and water form the basis of all plant metabolism. the energy of sunlight, captured by oxygenic photosynthesis and released by cellular respiration, is the basis of almost all life. photoautotrophs, including all green plants, algae and cyanobacteria gather energy directly from sunlight by photosynthesis. heterotrophs including all animals, all fungi, all completely parasitic plants, and non - photosynthetic bacteria take in organic molecules produced by photoautotrophs and respire them or use them in the construction of cells and tissues. respiration is the oxidation of carbon compounds by breaking them down into simpler structures to release the energy they contain, essentially the opposite of photosynthesis. molecules are moved within plants by transport processes that operate at a variety of spatial scales. subcellular transport of ions, electrons and molecules such as water and enzymes occurs across cell membranes. minerals and water are transported from roots to other parts of the plant in
##l ( active ingredient in cannabis ), caffeine, morphine and nicotine come directly from plants. others are simple derivatives of botanical natural products. for example, the pain killer aspirin is the acetyl ester of salicylic acid, originally isolated from the bark of willow trees, and a wide range of opiate painkillers like heroin are obtained by chemical modification of morphine obtained from the opium poppy. popular stimulants come from plants, such as caffeine from coffee, tea and chocolate, and nicotine from tobacco. most alcoholic beverages come from fermentation of carbohydrate - rich plant products such as barley ( beer ), rice ( sake ) and grapes ( wine ). native americans have used various plants as ways of treating illness or disease for thousands of years. this knowledge native americans have on plants has been recorded by enthnobotanists and then in turn has been used by pharmaceutical companies as a way of drug discovery. plants can synthesise coloured dyes and pigments such as the anthocyanins responsible for the red colour of red wine, yellow weld and blue woad used together to produce lincoln green, indoxyl, source of the blue dye indigo traditionally used to dye denim and the artist ' s pigments gamboge and rose madder. sugar, starch, cotton, linen, hemp, some types of rope, wood and particle boards, papyrus and paper, vegetable oils, wax, and natural rubber are examples of commercially important materials made from plant tissues or their secondary products. charcoal, a pure form of carbon made by pyrolysis of wood, has a long history as a metal - smelting fuel, as a filter material and adsorbent and as an artist ' s material and is one of the three ingredients of gunpowder. cellulose, the world ' s most abundant organic polymer, can be converted into energy, fuels, materials and chemical feedstock. products made from cellulose include rayon and cellophane, wallpaper paste, biobutanol and gun cotton. sugarcane, rapeseed and soy are some of the plants with a highly fermentable sugar or oil content that are used as sources of biofuels, important alternatives to fossil fuels, such as biodiesel. sweetgrass was used by native americans to ward off bugs like mosquitoes. these bug repelling properties of sweetgrass were later found by the american chemical society in the molecules phytol
a wide range of opiate painkillers like heroin are obtained by chemical modification of morphine obtained from the opium poppy. popular stimulants come from plants, such as caffeine from coffee, tea and chocolate, and nicotine from tobacco. most alcoholic beverages come from fermentation of carbohydrate - rich plant products such as barley ( beer ), rice ( sake ) and grapes ( wine ). native americans have used various plants as ways of treating illness or disease for thousands of years. this knowledge native americans have on plants has been recorded by enthnobotanists and then in turn has been used by pharmaceutical companies as a way of drug discovery. plants can synthesise coloured dyes and pigments such as the anthocyanins responsible for the red colour of red wine, yellow weld and blue woad used together to produce lincoln green, indoxyl, source of the blue dye indigo traditionally used to dye denim and the artist ' s pigments gamboge and rose madder. sugar, starch, cotton, linen, hemp, some types of rope, wood and particle boards, papyrus and paper, vegetable oils, wax, and natural rubber are examples of commercially important materials made from plant tissues or their secondary products. charcoal, a pure form of carbon made by pyrolysis of wood, has a long history as a metal - smelting fuel, as a filter material and adsorbent and as an artist ' s material and is one of the three ingredients of gunpowder. cellulose, the world ' s most abundant organic polymer, can be converted into energy, fuels, materials and chemical feedstock. products made from cellulose include rayon and cellophane, wallpaper paste, biobutanol and gun cotton. sugarcane, rapeseed and soy are some of the plants with a highly fermentable sugar or oil content that are used as sources of biofuels, important alternatives to fossil fuels, such as biodiesel. sweetgrass was used by native americans to ward off bugs like mosquitoes. these bug repelling properties of sweetgrass were later found by the american chemical society in the molecules phytol and coumarin. = = plant ecology = = plant ecology is the science of the functional relationships between plants and their habitats β the environments where they complete their life cycles. plant ecologists study the composition of local and regional floras, their biodiversity, genetic diversity and fitness, the adaptation of plants to their environment,
Question: In the forest, one type of tree produces special seeds. These seeds start to grow only after going through a fire. In the fire, the adult trees are destroyed. Which resources, needed for growth, are now available to the newly growing seeds?
A) sunlight and wind
B) sunlight and space
C) soil and pollen producers
D) pollen producers and space
|
B) sunlight and space
|
Context:
the decomposition theorem is deduced from local purity.
oil umbrella ) ; for calculating the time of death ( allowing for weather and insect activity ) ; described how to wash and examine the dead body to ascertain the reason for death. at that time the book had described methods for distinguishing between suicide and faked suicide. he wrote the book on forensics stating that all wounds or dead bodies should be examined, not avoided. the book became the first form of literature to help determine the cause of death. in one of song ci ' s accounts ( washing away of wrongs ), the case of a person murdered with a sickle was solved by an investigator who instructed each suspect to bring his sickle to one location. ( he realized it was a sickle by testing various blades on an animal carcass and comparing the wounds. ) flies, attracted by the smell of blood, eventually gathered on a single sickle. in light of this, the owner of that sickle confessed to the murder. the book also described how to distinguish between a drowning ( water in the lungs ) and strangulation ( broken neck cartilage ), and described evidence from examining corpses to determine if a death was caused by murder, suicide or accident. methods from around the world involved saliva and examination of the mouth and tongue to determine innocence or guilt, as a precursor to the polygraph test. in ancient india, some suspects were made to fill their mouths with dried rice and spit it back out. similarly, in ancient china, those accused of a crime would have rice powder placed in their mouths. in ancient middle - eastern cultures, the accused were made to lick hot metal rods briefly. it is thought that these tests had some validity since a guilty person would produce less saliva and thus have a drier mouth ; the accused would be considered guilty if rice was sticking to their mouths in abundance or if their tongues were severely burned due to lack of shielding from saliva. = = education and training = = initial glance, forensic intelligence may appear as a nascent facet of forensic science facilitated by advancements in information technologies such as computers, databases, and data - flow management software. however, a more profound examination reveals that forensic intelligence represents a genuine and emerging inclination among forensic practitioners to actively participate in investigative and policing strategies. in doing so, it elucidates existing practices within scientific literature, advocating for a paradigm shift from the prevailing conception of forensic science as a conglomerate of disciplines merely aiding the criminal justice system. instead, it urges a perspective that views forensic science as a discipline studying the informative potential of
the surface of the membrane, retentate is removed from the same side further downstream, whereas the permeate flow is tracked on the other side. in dead - end filtration, the direction of the fluid flow is normal to the membrane surface. both flow geometries offer some advantages and disadvantages. generally, dead - end filtration is used for feasibility studies on a laboratory scale. the dead - end membranes are relatively easy to fabricate which reduces the cost of the separation process. the dead - end membrane separation process is easy to implement and the process is usually cheaper than cross - flow membrane filtration. the dead - end filtration process is usually a batch - type process, where the filtering solution is loaded ( or slowly fed ) into the membrane device, which then allows passage of some particles subject to the driving force. the main disadvantage of dead - end filtration is the extensive membrane fouling and concentration polarization. the fouling is usually induced faster at higher driving forces. membrane fouling and particle retention in a feed solution also builds up a concentration gradients and particle backflow ( concentration polarization ). the tangential flow devices are more cost and labor - intensive, but they are less susceptible to fouling due to the sweeping effects and high shear rates of the passing flow. the most commonly used synthetic membrane devices ( modules ) are flat sheets / plates, spiral wounds, and hollow fibers. flat membranes used in filtration and separation processes can be enhanced with surface patterning, where microscopic structures are introduced to improve performance. these patterns increase surface area, optimize water flow, and reduce fouling, leading to higher permeability and longer membrane lifespan. research has shown that such modifications can significantly enhance efficiency in water purification, energy applications, and industrial separations. flat plates are usually constructed as circular thin flat membrane surfaces to be used in dead - end geometry modules. spiral wounds are constructed from similar flat membranes but in the form of a " pocket " containing two membrane sheets separated by a highly porous support plate. several such pockets are then wound around a tube to create a tangential flow geometry and to reduce membrane fouling. hollow fiber modules consist of an assembly of self - supporting fibers with dense skin separation layers, and a more open matrix helping to withstand pressure gradients and maintain structural integrity. the hollow fiber modules can contain up to 10, 000 fibers ranging from 200 to 2500 ΞΌm in diameter ; the main advantage of hollow fiber modules is the very large surface area within
use less energy than conventional thermal separation processes such as distillation, sublimation or crystallization. the separation process is purely physical and both fractions ( permeate and retentate ) can be obtained as useful products. cold separation using membrane technology is widely used in the food technology, biotechnology and pharmaceutical industries. furthermore, using membranes enables separations to take place that would be impossible using thermal separation methods. for example, it is impossible to separate the constituents of azeotropic liquids or solutes which form isomorphic crystals by distillation or recrystallization but such separations can be achieved using membrane technology. depending on the type of membrane, the selective separation of certain individual substances or substance mixtures is possible. important technical applications include the production of drinking water by reverse osmosis. in waste water treatment, membrane technology is becoming increasingly important. ultra / microfiltration can be very effective in removing colloids and macromolecules from wastewater. this is needed if wastewater is discharged into sensitive waters especially those designated for contact water sports and recreation. about half of the market is in medical applications such as artificial kidneys to remove toxic substances by hemodialysis and as artificial lung for bubble - free supply of oxygen in the blood. the importance of membrane technology is growing in the field of environmental protection ( nano - mem - pro ippc database ). even in modern energy recovery techniques, membranes are increasingly used, for example in fuel cells and in osmotic power plants. = = mass transfer = = two basic models can be distinguished for mass transfer through the membrane : the solution - diffusion model and the hydrodynamic model. in real membranes, these two transport mechanisms certainly occur side by side, especially during ultra - filtration. = = = solution - diffusion model = = = in the solution - diffusion model, transport occurs only by diffusion. the component that needs to be transported must first be dissolved in the membrane. the general approach of the solution - diffusion model is to assume that the chemical potential of the feed and permeate fluids are in equilibrium with the adjacent membrane surfaces such that appropriate expressions for the chemical potential in the fluid and membrane phases can be equated at the solution - membrane interface. this principle is more important for dense membranes without natural pores such as those used for reverse osmosis and in fuel cells. during the filtration process a boundary layer forms on the membrane. this concentration gradient is created by molecules which cannot pass through the membrane. the
the elimination of metabolic wastes. these enzyme - catalyzed reactions allow organisms to grow and reproduce, maintain their structures, and respond to their environments. metabolic reactions may be categorized as catabolic β the breaking down of compounds ( for example, the breaking down of glucose to pyruvate by cellular respiration ) ; or anabolic β the building up ( synthesis ) of compounds ( such as proteins, carbohydrates, lipids, and nucleic acids ). usually, catabolism releases energy, and anabolism consumes energy. the chemical reactions of metabolism are organized into metabolic pathways, in which one chemical is transformed through a series of steps into another chemical, each step being facilitated by a specific enzyme. enzymes are crucial to metabolism because they allow organisms to drive desirable reactions that require energy that will not occur by themselves, by coupling them to spontaneous reactions that release energy. enzymes act as catalysts β they allow a reaction to proceed more rapidly without being consumed by it β by reducing the amount of activation energy needed to convert reactants into products. enzymes also allow the regulation of the rate of a metabolic reaction, for example in response to changes in the cell ' s environment or to signals from other cells. = = = cellular respiration = = = cellular respiration is a set of metabolic reactions and processes that take place in cells to convert chemical energy from nutrients into adenosine triphosphate ( atp ), and then release waste products. the reactions involved in respiration are catabolic reactions, which break large molecules into smaller ones, releasing energy. respiration is one of the key ways a cell releases chemical energy to fuel cellular activity. the overall reaction occurs in a series of biochemical steps, some of which are redox reactions. although cellular respiration is technically a combustion reaction, it clearly does not resemble one when it occurs in a cell because of the slow, controlled release of energy from the series of reactions. sugar in the form of glucose is the main nutrient used by animal and plant cells in respiration. cellular respiration involving oxygen is called aerobic respiration, which has four stages : glycolysis, citric acid cycle ( or krebs cycle ), electron transport chain, and oxidative phosphorylation. glycolysis is a metabolic process that occurs in the cytoplasm whereby glucose is converted into two pyruvates, with two net molecules of atp being produced at the same time. each pyruvate is then
generally, dead - end filtration is used for feasibility studies on a laboratory scale. the dead - end membranes are relatively easy to fabricate which reduces the cost of the separation process. the dead - end membrane separation process is easy to implement and the process is usually cheaper than cross - flow membrane filtration. the dead - end filtration process is usually a batch - type process, where the filtering solution is loaded ( or slowly fed ) into the membrane device, which then allows passage of some particles subject to the driving force. the main disadvantage of dead - end filtration is the extensive membrane fouling and concentration polarization. the fouling is usually induced faster at higher driving forces. membrane fouling and particle retention in a feed solution also builds up a concentration gradients and particle backflow ( concentration polarization ). the tangential flow devices are more cost and labor - intensive, but they are less susceptible to fouling due to the sweeping effects and high shear rates of the passing flow. the most commonly used synthetic membrane devices ( modules ) are flat sheets / plates, spiral wounds, and hollow fibers. flat membranes used in filtration and separation processes can be enhanced with surface patterning, where microscopic structures are introduced to improve performance. these patterns increase surface area, optimize water flow, and reduce fouling, leading to higher permeability and longer membrane lifespan. research has shown that such modifications can significantly enhance efficiency in water purification, energy applications, and industrial separations. flat plates are usually constructed as circular thin flat membrane surfaces to be used in dead - end geometry modules. spiral wounds are constructed from similar flat membranes but in the form of a " pocket " containing two membrane sheets separated by a highly porous support plate. several such pockets are then wound around a tube to create a tangential flow geometry and to reduce membrane fouling. hollow fiber modules consist of an assembly of self - supporting fibers with dense skin separation layers, and a more open matrix helping to withstand pressure gradients and maintain structural integrity. the hollow fiber modules can contain up to 10, 000 fibers ranging from 200 to 2500 ΞΌm in diameter ; the main advantage of hollow fiber modules is the very large surface area within an enclosed volume, increasing the efficiency of the separation process. the disc tube module uses a cross - flow geometry and consists of a pressure tube and hydraulic discs, which are held by a central tension rod, and membrane cushions that lie between two discs. = = membrane performance and governing equations = = the selection of synthetic membranes
reaction to proceed more rapidly without being consumed by it β by reducing the amount of activation energy needed to convert reactants into products. enzymes also allow the regulation of the rate of a metabolic reaction, for example in response to changes in the cell ' s environment or to signals from other cells. = = = cellular respiration = = = cellular respiration is a set of metabolic reactions and processes that take place in cells to convert chemical energy from nutrients into adenosine triphosphate ( atp ), and then release waste products. the reactions involved in respiration are catabolic reactions, which break large molecules into smaller ones, releasing energy. respiration is one of the key ways a cell releases chemical energy to fuel cellular activity. the overall reaction occurs in a series of biochemical steps, some of which are redox reactions. although cellular respiration is technically a combustion reaction, it clearly does not resemble one when it occurs in a cell because of the slow, controlled release of energy from the series of reactions. sugar in the form of glucose is the main nutrient used by animal and plant cells in respiration. cellular respiration involving oxygen is called aerobic respiration, which has four stages : glycolysis, citric acid cycle ( or krebs cycle ), electron transport chain, and oxidative phosphorylation. glycolysis is a metabolic process that occurs in the cytoplasm whereby glucose is converted into two pyruvates, with two net molecules of atp being produced at the same time. each pyruvate is then oxidized into acetyl - coa by the pyruvate dehydrogenase complex, which also generates nadh and carbon dioxide. acetyl - coa enters the citric acid cycle, which takes places inside the mitochondrial matrix. at the end of the cycle, the total yield from 1 glucose ( or 2 pyruvates ) is 6 nadh, 2 fadh2, and 2 atp molecules. finally, the next stage is oxidative phosphorylation, which in eukaryotes, occurs in the mitochondrial cristae. oxidative phosphorylation comprises the electron transport chain, which is a series of four protein complexes that transfer electrons from one complex to another, thereby releasing energy from nadh and fadh2 that is coupled to the pumping of protons ( hydrogen ions ) across the inner mitochondrial membrane ( chemiosmosis ), which generates a proton motive force. energy
##ration fuel cell operations in a temperature gradient membrane distillation = = membrane shapes and flow geometries = = there are two main flow configurations of membrane processes : cross - flow ( or tangential flow ) and dead - end filtrations. in cross - flow filtration the feed flow is tangential to the surface of the membrane, retentate is removed from the same side further downstream, whereas the permeate flow is tracked on the other side. in dead - end filtration, the direction of the fluid flow is normal to the membrane surface. both flow geometries offer some advantages and disadvantages. generally, dead - end filtration is used for feasibility studies on a laboratory scale. the dead - end membranes are relatively easy to fabricate which reduces the cost of the separation process. the dead - end membrane separation process is easy to implement and the process is usually cheaper than cross - flow membrane filtration. the dead - end filtration process is usually a batch - type process, where the filtering solution is loaded ( or slowly fed ) into the membrane device, which then allows passage of some particles subject to the driving force. the main disadvantage of dead - end filtration is the extensive membrane fouling and concentration polarization. the fouling is usually induced faster at higher driving forces. membrane fouling and particle retention in a feed solution also builds up a concentration gradients and particle backflow ( concentration polarization ). the tangential flow devices are more cost and labor - intensive, but they are less susceptible to fouling due to the sweeping effects and high shear rates of the passing flow. the most commonly used synthetic membrane devices ( modules ) are flat sheets / plates, spiral wounds, and hollow fibers. flat membranes used in filtration and separation processes can be enhanced with surface patterning, where microscopic structures are introduced to improve performance. these patterns increase surface area, optimize water flow, and reduce fouling, leading to higher permeability and longer membrane lifespan. research has shown that such modifications can significantly enhance efficiency in water purification, energy applications, and industrial separations. flat plates are usually constructed as circular thin flat membrane surfaces to be used in dead - end geometry modules. spiral wounds are constructed from similar flat membranes but in the form of a " pocket " containing two membrane sheets separated by a highly porous support plate. several such pockets are then wound around a tube to create a tangential flow geometry and to reduce membrane fouling. hollow fiber modules consist of an
liver glycogen. during recovery, when oxygen becomes available, nad + attaches to hydrogen from lactate to form atp. in yeast, the waste products are ethanol and carbon dioxide. this type of fermentation is known as alcoholic or ethanol fermentation. the atp generated in this process is made by substrate - level phosphorylation, which does not require oxygen. = = = photosynthesis = = = photosynthesis is a process used by plants and other organisms to convert light energy into chemical energy that can later be released to fuel the organism ' s metabolic activities via cellular respiration. this chemical energy is stored in carbohydrate molecules, such as sugars, which are synthesized from carbon dioxide and water. in most cases, oxygen is released as a waste product. most plants, algae, and cyanobacteria perform photosynthesis, which is largely responsible for producing and maintaining the oxygen content of the earth ' s atmosphere, and supplies most of the energy necessary for life on earth. photosynthesis has four stages : light absorption, electron transport, atp synthesis, and carbon fixation. light absorption is the initial step of photosynthesis whereby light energy is absorbed by chlorophyll pigments attached to proteins in the thylakoid membranes. the absorbed light energy is used to remove electrons from a donor ( water ) to a primary electron acceptor, a quinone designated as q. in the second stage, electrons move from the quinone primary electron acceptor through a series of electron carriers until they reach a final electron acceptor, which is usually the oxidized form of nadp +, which is reduced to nadph, a process that takes place in a protein complex called photosystem i ( psi ). the transport of electrons is coupled to the movement of protons ( or hydrogen ) from the stroma to the thylakoid membrane, which forms a ph gradient across the membrane as hydrogen becomes more concentrated in the lumen than in the stroma. this is analogous to the proton - motive force generated across the inner mitochondrial membrane in aerobic respiration. during the third stage of photosynthesis, the movement of protons down their concentration gradients from the thylakoid lumen to the stroma through the atp synthase is coupled to the synthesis of atp by that same atp synthase. the nadph and atps generated by the light - dependent reactions in the second and third stages, respectively, provide the energy and
a process of fermentation. the pyruvate is not transported into the mitochondrion but remains in the cytoplasm, where it is converted to waste products that may be removed from the cell. this serves the purpose of oxidizing the electron carriers so that they can perform glycolysis again and removing the excess pyruvate. fermentation oxidizes nadh to nad + so it can be re - used in glycolysis. in the absence of oxygen, fermentation prevents the buildup of nadh in the cytoplasm and provides nad + for glycolysis. this waste product varies depending on the organism. in skeletal muscles, the waste product is lactic acid. this type of fermentation is called lactic acid fermentation. in strenuous exercise, when energy demands exceed energy supply, the respiratory chain cannot process all of the hydrogen atoms joined by nadh. during anaerobic glycolysis, nad + regenerates when pairs of hydrogen combine with pyruvate to form lactate. lactate formation is catalyzed by lactate dehydrogenase in a reversible reaction. lactate can also be used as an indirect precursor for liver glycogen. during recovery, when oxygen becomes available, nad + attaches to hydrogen from lactate to form atp. in yeast, the waste products are ethanol and carbon dioxide. this type of fermentation is known as alcoholic or ethanol fermentation. the atp generated in this process is made by substrate - level phosphorylation, which does not require oxygen. = = = photosynthesis = = = photosynthesis is a process used by plants and other organisms to convert light energy into chemical energy that can later be released to fuel the organism ' s metabolic activities via cellular respiration. this chemical energy is stored in carbohydrate molecules, such as sugars, which are synthesized from carbon dioxide and water. in most cases, oxygen is released as a waste product. most plants, algae, and cyanobacteria perform photosynthesis, which is largely responsible for producing and maintaining the oxygen content of the earth ' s atmosphere, and supplies most of the energy necessary for life on earth. photosynthesis has four stages : light absorption, electron transport, atp synthesis, and carbon fixation. light absorption is the initial step of photosynthesis whereby light energy is absorbed by chlorophyll pigments attached to
Question: Which is an example of a decomposer?
A) a hawk
B) a mouse
C) a flower
D) a mushroom
|
D) a mushroom
|
Context:
radiation exposure. marie curie died from aplastic anemia which resulted from her high levels of exposure. two scientists, an american and canadian respectively, harry daghlian and louis slotin, died after mishandling the same plutonium mass. unlike conventional weapons, the intense light, heat, and explosive force is not the only deadly component to a nuclear weapon. approximately half of the deaths from hiroshima and nagasaki died two to five years afterward from radiation exposure. civilian nuclear and radiological accidents primarily involve nuclear power plants. most common are nuclear leaks that expose workers to hazardous material. a nuclear meltdown refers to the more serious hazard of releasing nuclear material into the surrounding environment. the most significant meltdowns occurred at three mile island in pennsylvania and chernobyl in the soviet ukraine. the earthquake and tsunami on march 11, 2011 caused serious damage to three nuclear reactors and a spent fuel storage pond at the fukushima daiichi nuclear power plant in japan. military reactors that experienced similar accidents were windscale in the united kingdom and sl - 1 in the united states. military accidents usually involve the loss or unexpected detonation of nuclear weapons. the castle bravo test in 1954 produced a larger yield than expected, which contaminated nearby islands, a japanese fishing boat ( with one fatality ), and raised concerns about contaminated fish in japan. in the 1950s through 1970s, several nuclear bombs were lost from submarines and aircraft, some of which have never been recovered. the last twenty years have seen a marked decline in such accidents. = = examples of environmental benefits = = proponents of nuclear energy note that annually, nuclear - generated electricity reduces 470 million metric tons of carbon dioxide emissions that would otherwise come from fossil fuels. additionally, the amount of comparatively low waste that nuclear energy does create is safely disposed of by the large scale nuclear energy production facilities or it is repurposed / recycled for other energy uses. proponents of nuclear energy also bring to attention the opportunity cost of utilizing other forms of electricity. for example, the environmental protection agency estimates that coal kills 30, 000 people a year, as a result of its environmental impact, while 60 people died in the chernobyl disaster. a real world example of impact provided by proponents of nuclear energy is the 650, 000 ton increase in carbon emissions in the two months following the closure of the vermont yankee nuclear plant. = = see also = = atomic age lists of nuclear disasters and radioactive incidents nuclear power debate outline of nuclear technology radiology = = references = = = = external links = = nuclear energy institute β beneficial uses
and by processing power using accelerators. food irradiation is only a niche application compared to medical supplies, plastic materials, raw materials, gemstones, cables and wires, etc. = = accidents = = nuclear accidents, because of the powerful forces involved, are often very dangerous. historically, the first incidents involved fatal radiation exposure. marie curie died from aplastic anemia which resulted from her high levels of exposure. two scientists, an american and canadian respectively, harry daghlian and louis slotin, died after mishandling the same plutonium mass. unlike conventional weapons, the intense light, heat, and explosive force is not the only deadly component to a nuclear weapon. approximately half of the deaths from hiroshima and nagasaki died two to five years afterward from radiation exposure. civilian nuclear and radiological accidents primarily involve nuclear power plants. most common are nuclear leaks that expose workers to hazardous material. a nuclear meltdown refers to the more serious hazard of releasing nuclear material into the surrounding environment. the most significant meltdowns occurred at three mile island in pennsylvania and chernobyl in the soviet ukraine. the earthquake and tsunami on march 11, 2011 caused serious damage to three nuclear reactors and a spent fuel storage pond at the fukushima daiichi nuclear power plant in japan. military reactors that experienced similar accidents were windscale in the united kingdom and sl - 1 in the united states. military accidents usually involve the loss or unexpected detonation of nuclear weapons. the castle bravo test in 1954 produced a larger yield than expected, which contaminated nearby islands, a japanese fishing boat ( with one fatality ), and raised concerns about contaminated fish in japan. in the 1950s through 1970s, several nuclear bombs were lost from submarines and aircraft, some of which have never been recovered. the last twenty years have seen a marked decline in such accidents. = = examples of environmental benefits = = proponents of nuclear energy note that annually, nuclear - generated electricity reduces 470 million metric tons of carbon dioxide emissions that would otherwise come from fossil fuels. additionally, the amount of comparatively low waste that nuclear energy does create is safely disposed of by the large scale nuclear energy production facilities or it is repurposed / recycled for other energy uses. proponents of nuclear energy also bring to attention the opportunity cost of utilizing other forms of electricity. for example, the environmental protection agency estimates that coal kills 30, 000 people a year, as a result of its environmental impact, while 60 people died in the chernobyl disaster. a real world example of impact provided by proponents of nuclear energy is
the walls of a victim ' s stomach. toxicology, a subfield of forensic chemistry, focuses on detecting and identifying drugs, poisons, and other toxic substances in biological samples. forensic toxicologists work on cases involving drug overdoses, poisoning, and substance abuse. their work is critical in determining whether harmful substances play a role in a person β s death or impairment. read more james marsh was the first to apply this new science to the art of forensics. he was called by the prosecution in a murder trial to give evidence as a chemist in 1832. the defendant, john bodle, was accused of poisoning his grandfather with arsenic - laced coffee. marsh performed the standard test by mixing a suspected sample with hydrogen sulfide and hydrochloric acid. while he was able to detect arsenic as yellow arsenic trisulfide, when it was shown to the jury it had deteriorated, allowing the suspect to be acquitted due to reasonable doubt. annoyed by that, marsh developed a much better test. he combined a sample containing arsenic with sulfuric acid and arsenic - free zinc, resulting in arsine gas. the gas was ignited, and it decomposed to pure metallic arsenic, which, when passed to a cold surface, would appear as a silvery - black deposit. so sensitive was the test, known formally as the marsh test, that it could detect as little as one - fiftieth of a milligram of arsenic. he first described this test in the edinburgh philosophical journal in 1836. = = = ballistics and firearms = = = ballistics is " the science of the motion of projectiles in flight ". in forensic science, analysts examine the patterns left on bullets and cartridge casings after being ejected from a weapon. when fired, a bullet is left with indentations and markings that are unique to the barrel and firing pin of the firearm that ejected the bullet. this examination can help scientists identify possible makes and models of weapons connected to a crime. henry goddard at scotland yard pioneered the use of bullet comparison in 1835. he noticed a flaw in the bullet that killed the victim and was able to trace this back to the mold that was used in the manufacturing process. = = = anthropometry = = = the french police officer alphonse bertillon was the first to apply the anthropological technique of anthropometry to law enforcement, thereby creating an identification system based on physical measurements. before that time, criminals could be identified only by name or photograph. dissatisfied with the ad hoc methods used to identify captured
the first three greek letters. some of these kinds of radiation could pass through ordinary matter, and all of them could be harmful in large amounts. all of the early researchers received various radiation burns, much like sunburn, and thought little of it. the new phenomenon of radioactivity was seized upon by the manufacturers of quack medicine ( as had the discoveries of electricity and magnetism, earlier ), and a number of patent medicines and treatments involving radioactivity were put forward. gradually it was realized that the radiation produced by radioactive decay was ionizing radiation, and that even quantities too small to burn could pose a severe long - term hazard. many of the scientists working on radioactivity died of cancer as a result of their exposure. radioactive patent medicines mostly disappeared, but other applications of radioactive materials persisted, such as the use of radium salts to produce glowing dials on meters. as the atom came to be better understood, the nature of radioactivity became clearer. some larger atomic nuclei are unstable, and so decay ( release matter or energy ) after a random interval. the three forms of radiation that becquerel and the curies discovered are also more fully understood. alpha decay is when a nucleus releases an alpha particle, which is two protons and two neutrons, equivalent to a helium nucleus. beta decay is the release of a beta particle, a high - energy electron. gamma decay releases gamma rays, which unlike alpha and beta radiation are not matter but electromagnetic radiation of very high frequency, and therefore energy. this type of radiation is the most dangerous and most difficult to block. all three types of radiation occur naturally in certain elements. it has also become clear that the ultimate source of most terrestrial energy is nuclear, either through radiation from the sun caused by stellar thermonuclear reactions or by radioactive decay of uranium within the earth, the principal source of geothermal energy. = = = nuclear fission = = = in natural nuclear radiation, the byproducts are very small compared to the nuclei from which they originate. nuclear fission is the process of splitting a nucleus into roughly equal parts, and releasing energy and neutrons in the process. if these neutrons are captured by another unstable nucleus, they can fission as well, leading to a chain reaction. the average number of neutrons released per nucleus that go on to fission another nucleus is referred to as k. values of k larger than 1 mean that the fission reaction is releasing more neutrons than it absorbs, and therefore is referred to as a self
which came to be called radioactivity. he, pierre curie and marie curie began investigating the phenomenon. in the process, they isolated the element radium, which is highly radioactive. they discovered that radioactive materials produce intense, penetrating rays of three distinct sorts, which they labeled alpha, beta, and gamma after the first three greek letters. some of these kinds of radiation could pass through ordinary matter, and all of them could be harmful in large amounts. all of the early researchers received various radiation burns, much like sunburn, and thought little of it. the new phenomenon of radioactivity was seized upon by the manufacturers of quack medicine ( as had the discoveries of electricity and magnetism, earlier ), and a number of patent medicines and treatments involving radioactivity were put forward. gradually it was realized that the radiation produced by radioactive decay was ionizing radiation, and that even quantities too small to burn could pose a severe long - term hazard. many of the scientists working on radioactivity died of cancer as a result of their exposure. radioactive patent medicines mostly disappeared, but other applications of radioactive materials persisted, such as the use of radium salts to produce glowing dials on meters. as the atom came to be better understood, the nature of radioactivity became clearer. some larger atomic nuclei are unstable, and so decay ( release matter or energy ) after a random interval. the three forms of radiation that becquerel and the curies discovered are also more fully understood. alpha decay is when a nucleus releases an alpha particle, which is two protons and two neutrons, equivalent to a helium nucleus. beta decay is the release of a beta particle, a high - energy electron. gamma decay releases gamma rays, which unlike alpha and beta radiation are not matter but electromagnetic radiation of very high frequency, and therefore energy. this type of radiation is the most dangerous and most difficult to block. all three types of radiation occur naturally in certain elements. it has also become clear that the ultimate source of most terrestrial energy is nuclear, either through radiation from the sun caused by stellar thermonuclear reactions or by radioactive decay of uranium within the earth, the principal source of geothermal energy. = = = nuclear fission = = = in natural nuclear radiation, the byproducts are very small compared to the nuclei from which they originate. nuclear fission is the process of splitting a nucleus into roughly equal parts, and releasing energy and neutrons in the process. if these neutrons are captured by another unstable nucleus
in space, can adversely affect the earth ' s environment. some hypergolic rocket propellants, such as hydrazine, are highly toxic prior to combustion, but decompose into less toxic compounds after burning. rockets using hydrocarbon fuels, such as kerosene, release carbon dioxide and soot in their exhaust. carbon dioxide emissions are insignificant compared to those from other sources ; on average, the united states consumed 803 million us gal ( 3. 0 million m3 ) of liquid fuels per day in 2014, while a single falcon 9 rocket first stage burns around 25, 000 us gallons ( 95 m3 ) of kerosene fuel per launch. even if a falcon 9 were launched every single day, it would only represent 0. 006 % of liquid fuel consumption ( and carbon dioxide emissions ) for that day. additionally, the exhaust from lox - and lh2 - fueled engines, like the ssme, is almost entirely water vapor. nasa addressed environmental concerns with its canceled constellation program in accordance with the national environmental policy act in 2011. in contrast, ion engines use harmless noble gases like xenon for propulsion. an example of nasa ' s environmental efforts is the nasa sustainability base. additionally, the exploration sciences building was awarded the leed gold rating in 2010. on may 8, 2003, the environmental protection agency recognized nasa as the first federal agency to directly use landfill gas to produce energy at one of its facilities β the goddard space flight center, greenbelt, maryland. in 2018, nasa along with other companies including sensor coating systems, pratt & whitney, monitor coating and utrc launched the project caution ( coatings for ultra high temperature detection ). this project aims to enhance the temperature range of the thermal history coating up to 1, 500 Β°c ( 2, 730 Β°f ) and beyond. the final goal of this project is improving the safety of jet engines as well as increasing efficiency and reducing co2 emissions. = = = climate change = = = nasa also researches and publishes on climate change. its statements concur with the global scientific consensus that the climate is warming. bob walker, who has advised former us president donald trump on space issues, has advocated that nasa should focus on space exploration and that its climate study operations should be transferred to other agencies such as noaa. former nasa atmospheric scientist j. marshall shepherd countered that earth science study was built into nasa ' s mission at its creation in the 1958 national aeronautics and space act. nasa won the 2020 webby people ' s voice award for green in the category
, lightning strikes, tornadoes, building fires, wildfires, and mass shootings disabling most of the system if not the entirety of it. geographic redundancy locations can be more than 621 miles ( 999 km ) continental, more than 62 miles apart and less than 93 miles ( 150 km ) apart, less than 62 miles apart, but not on the same campus, or different buildings that are more than 300 feet ( 91 m ) apart on the same campus. the following methods can reduce the risks of damage by a fire conflagration : large buildings at least 80 feet ( 24 m ) to 110 feet ( 34 m ) apart, but sometimes a minimum of 210 feet ( 64 m ) apart. : 9 high - rise buildings at least 82 feet ( 25 m ) apart : 12 open spaces clear of flammable vegetation within 200 feet ( 61 m ) on each side of objects different wings on the same building, in rooms that are separated by more than 300 feet ( 91 m ) different floors on the same wing of a building in rooms that are horizontally offset by a minimum of 70 feet ( 21 m ) with fire walls between the rooms that are on different floors two rooms separated by another room, leaving at least a 70 - foot gap between the two rooms there should be a minimum of two separated fire walls and on opposite sides of a corridor geographic redundancy is used by amazon web services ( aws ), google cloud platform ( gcp ), microsoft azure, netflix, dropbox, salesforce, linkedin, paypal, twitter, facebook, apple icloud, cisco meraki, and many others to provide geographic redundancy, high availability, fault tolerance and to ensure availability and reliability for their cloud services. as another example, to minimize risk of damage from severe windstorms or water damage, buildings can be located at least 2 miles ( 3. 2 km ) away from the shore, with an elevation of at least 5 feet ( 1. 5 m ) above sea level. for additional protection, they can be located at least 100 feet ( 30 m ) away from flood plain areas. = = functions of redundancy = = the two functions of redundancy are passive redundancy and active redundancy. both functions prevent performance decline from exceeding specification limits without human intervention using extra capacity. passive redundancy uses excess capacity to reduce the impact of component failures. one common form of passive redundancy is the extra strength of cabling and struts used in bridges.
safety security assurance framework applied to two standards iec 61508 and common criteria - iso 15408
, buses, trucks, etc. it includes branch study of mechanical, electronic, software and safety elements. some of the engineering attributes and disciplines that are of importance to the automotive engineer include : safety engineering : safety engineering is the assessment of various crash scenarios and their impact on the vehicle occupants. these are tested against very stringent governmental regulations. some of these requirements include : seat belt and air bag functionality testing, front and side - impact testing, and tests of rollover resistance. assessments are done with various methods and tools, including computer crash simulation ( typically finite element analysis ), crash - test dummy, and partial system sled and full vehicle crashes. fuel economy / emissions : fuel economy is the measured fuel efficiency of the vehicle in miles per gallon or kilometers per liter. emissions - testing covers the measurement of vehicle emissions, including hydrocarbons, nitrogen oxides ( nox ), carbon monoxide ( co ), carbon dioxide ( co2 ), and evaporative emissions. nvh engineering ( noise, vibration, and harshness ) : nvh involves customer feedback ( both tactile [ felt ] and audible [ heard ] ) concerning a vehicle. while sound can be interpreted as a rattle, squeal, or hot, a tactile response can be seat vibration or a buzz in the steering wheel. this feedback is generated by components either rubbing, vibrating, or rotating. nvh response can be classified in various ways : powertrain nvh, road noise, wind noise, component noise, and squeak and rattle. note, there are both good and bad nvh qualities. the nvh engineer works to either eliminate bad nvh or change the " bad nvh " to good ( i. e., exhaust tones ). vehicle electronics : automotive electronics is an increasingly important aspect of automotive engineering. modern vehicles employ dozens of electronic systems. these systems are responsible for operational controls such as the throttle, brake and steering controls ; as well as many comfort - and - convenience systems such as the hvac, infotainment, and lighting systems. it would not be possible for automobiles to meet modern safety and fuel - economy requirements without electronic controls. performance : performance is a measurable and testable value of a vehicle ' s ability to perform in various conditions. performance can be considered in a wide variety of tasks, but it generally considers how quickly a car can accelerate ( e. g. standing start 1 / 4 mile elapsed time, 0 β 60 mph, etc. )
is also higher at high temperature, as shown by carnot ' s theorem. in a conventional metallic engine, much of the energy released from the fuel must be dissipated as waste heat in order to prevent a meltdown of the metallic parts. despite all of these desirable properties, such engines are not in production because the manufacturing of ceramic parts in the requisite precision and durability is difficult. imperfection in the ceramic leads to cracks, which can lead to potentially dangerous equipment failure. such engines are possible in laboratory settings, but mass - production is not feasible with current technology. work is being done in developing ceramic parts for gas turbine engines. currently, even blades made of advanced metal alloys used in the engines ' hot section require cooling and careful limiting of operating temperatures. turbine engines made with ceramics could operate more efficiently, giving aircraft greater range and payload for a set amount of fuel. recently, there have been advances in ceramics which include bio - ceramics, such as dental implants and synthetic bones. hydroxyapatite, the natural mineral component of bone, has been made synthetically from a number of biological and chemical sources and can be formed into ceramic materials. orthopedic implants made from these materials bond readily to bone and other tissues in the body without rejection or inflammatory reactions. because of this, they are of great interest for gene delivery and tissue engineering scaffolds. most hydroxyapatite ceramics are very porous and lack mechanical strength and are used to coat metal orthopedic devices to aid in forming a bond to bone or as bone fillers. they are also used as fillers for orthopedic plastic screws to aid in reducing the inflammation and increase absorption of these plastic materials. work is being done to make strong, fully dense nano crystalline hydroxyapatite ceramic materials for orthopedic weight bearing devices, replacing foreign metal and plastic orthopedic materials with a synthetic, but naturally occurring, bone mineral. ultimately these ceramic materials may be used as bone replacements or with the incorporation of protein collagens, synthetic bones. durable actinide - containing ceramic materials have many applications such as in nuclear fuels for burning excess pu and in chemically - inert sources of alpha irradiation for power supply of unmanned space vehicles or to produce electricity for microelectronic devices. both use and disposal of radioactive actinides require their immobilization in a durable host material. nuclear waste long - lived radionuclides such as actinides are immobilized using chemical
Question: In the lab, students are working with substances that could cause burns. Which of these safety precautions is most important to follow?
A) Sinks are empty.
B) Glassware is clean.
C) Eye protection is worn.
D) Electrical equipment is off.
|
C) Eye protection is worn.
|
Context:
oxygen, carbon dioxide, and water to pass through while restricting the movement of larger molecules and charged particles such as ions. cell membranes also contain membrane proteins, including integral membrane proteins that go across the membrane serving as membrane transporters, and peripheral proteins that loosely attach to the outer side of the cell membrane, acting as enzymes shaping the cell. cell membranes are involved in various cellular processes such as cell adhesion, storing electrical energy, and cell signalling and serve as the attachment surface for several extracellular structures such as a cell wall, glycocalyx, and cytoskeleton. within the cytoplasm of a cell, there are many biomolecules such as proteins and nucleic acids. in addition to biomolecules, eukaryotic cells have specialized structures called organelles that have their own lipid bilayers or are spatially units. these organelles include the cell nucleus, which contains most of the cell ' s dna, or mitochondria, which generate adenosine triphosphate ( atp ) to power cellular processes. other organelles such as endoplasmic reticulum and golgi apparatus play a role in the synthesis and packaging of proteins, respectively. biomolecules such as proteins can be engulfed by lysosomes, another specialized organelle. plant cells have additional organelles that distinguish them from animal cells such as a cell wall that provides support for the plant cell, chloroplasts that harvest sunlight energy to produce sugar, and vacuoles that provide storage and structural support as well as being involved in reproduction and breakdown of plant seeds. eukaryotic cells also have cytoskeleton that is made up of microtubules, intermediate filaments, and microfilaments, all of which provide support for the cell and are involved in the movement of the cell and its organelles. in terms of their structural composition, the microtubules are made up of tubulin ( e. g., Ξ± - tubulin and Ξ² - tubulin ) whereas intermediate filaments are made up of fibrous proteins. microfilaments are made up of actin molecules that interact with other strands of proteins. = = = metabolism = = = all cells require energy to sustain cellular processes. metabolism is the set of chemical reactions in an organism. the three main purposes of metabolism are : the conversion of food to energy to run cellular processes ; the conversion of food / fuel to monomer building blocks ; and
within a cell membrane that separates its cytoplasm from the extracellular space. a cell membrane consists of a lipid bilayer, including cholesterols that sit between phospholipids to maintain their fluidity at various temperatures. cell membranes are semipermeable, allowing small molecules such as oxygen, carbon dioxide, and water to pass through while restricting the movement of larger molecules and charged particles such as ions. cell membranes also contain membrane proteins, including integral membrane proteins that go across the membrane serving as membrane transporters, and peripheral proteins that loosely attach to the outer side of the cell membrane, acting as enzymes shaping the cell. cell membranes are involved in various cellular processes such as cell adhesion, storing electrical energy, and cell signalling and serve as the attachment surface for several extracellular structures such as a cell wall, glycocalyx, and cytoskeleton. within the cytoplasm of a cell, there are many biomolecules such as proteins and nucleic acids. in addition to biomolecules, eukaryotic cells have specialized structures called organelles that have their own lipid bilayers or are spatially units. these organelles include the cell nucleus, which contains most of the cell ' s dna, or mitochondria, which generate adenosine triphosphate ( atp ) to power cellular processes. other organelles such as endoplasmic reticulum and golgi apparatus play a role in the synthesis and packaging of proteins, respectively. biomolecules such as proteins can be engulfed by lysosomes, another specialized organelle. plant cells have additional organelles that distinguish them from animal cells such as a cell wall that provides support for the plant cell, chloroplasts that harvest sunlight energy to produce sugar, and vacuoles that provide storage and structural support as well as being involved in reproduction and breakdown of plant seeds. eukaryotic cells also have cytoskeleton that is made up of microtubules, intermediate filaments, and microfilaments, all of which provide support for the cell and are involved in the movement of the cell and its organelles. in terms of their structural composition, the microtubules are made up of tubulin ( e. g., Ξ± - tubulin and Ξ² - tubulin ) whereas intermediate filaments are made up of fibrous proteins. microfilaments are made up of actin molecules that interact with
the surface of the membrane, retentate is removed from the same side further downstream, whereas the permeate flow is tracked on the other side. in dead - end filtration, the direction of the fluid flow is normal to the membrane surface. both flow geometries offer some advantages and disadvantages. generally, dead - end filtration is used for feasibility studies on a laboratory scale. the dead - end membranes are relatively easy to fabricate which reduces the cost of the separation process. the dead - end membrane separation process is easy to implement and the process is usually cheaper than cross - flow membrane filtration. the dead - end filtration process is usually a batch - type process, where the filtering solution is loaded ( or slowly fed ) into the membrane device, which then allows passage of some particles subject to the driving force. the main disadvantage of dead - end filtration is the extensive membrane fouling and concentration polarization. the fouling is usually induced faster at higher driving forces. membrane fouling and particle retention in a feed solution also builds up a concentration gradients and particle backflow ( concentration polarization ). the tangential flow devices are more cost and labor - intensive, but they are less susceptible to fouling due to the sweeping effects and high shear rates of the passing flow. the most commonly used synthetic membrane devices ( modules ) are flat sheets / plates, spiral wounds, and hollow fibers. flat membranes used in filtration and separation processes can be enhanced with surface patterning, where microscopic structures are introduced to improve performance. these patterns increase surface area, optimize water flow, and reduce fouling, leading to higher permeability and longer membrane lifespan. research has shown that such modifications can significantly enhance efficiency in water purification, energy applications, and industrial separations. flat plates are usually constructed as circular thin flat membrane surfaces to be used in dead - end geometry modules. spiral wounds are constructed from similar flat membranes but in the form of a " pocket " containing two membrane sheets separated by a highly porous support plate. several such pockets are then wound around a tube to create a tangential flow geometry and to reduce membrane fouling. hollow fiber modules consist of an assembly of self - supporting fibers with dense skin separation layers, and a more open matrix helping to withstand pressure gradients and maintain structural integrity. the hollow fiber modules can contain up to 10, 000 fibers ranging from 200 to 2500 ΞΌm in diameter ; the main advantage of hollow fiber modules is the very large surface area within
are single - celled organisms such as bacteria, whereas eukaryotes can be single - celled or multicellular. in multicellular organisms, every cell in the organism ' s body is derived ultimately from a single cell in a fertilized egg. = = = cell structure = = = every cell is enclosed within a cell membrane that separates its cytoplasm from the extracellular space. a cell membrane consists of a lipid bilayer, including cholesterols that sit between phospholipids to maintain their fluidity at various temperatures. cell membranes are semipermeable, allowing small molecules such as oxygen, carbon dioxide, and water to pass through while restricting the movement of larger molecules and charged particles such as ions. cell membranes also contain membrane proteins, including integral membrane proteins that go across the membrane serving as membrane transporters, and peripheral proteins that loosely attach to the outer side of the cell membrane, acting as enzymes shaping the cell. cell membranes are involved in various cellular processes such as cell adhesion, storing electrical energy, and cell signalling and serve as the attachment surface for several extracellular structures such as a cell wall, glycocalyx, and cytoskeleton. within the cytoplasm of a cell, there are many biomolecules such as proteins and nucleic acids. in addition to biomolecules, eukaryotic cells have specialized structures called organelles that have their own lipid bilayers or are spatially units. these organelles include the cell nucleus, which contains most of the cell ' s dna, or mitochondria, which generate adenosine triphosphate ( atp ) to power cellular processes. other organelles such as endoplasmic reticulum and golgi apparatus play a role in the synthesis and packaging of proteins, respectively. biomolecules such as proteins can be engulfed by lysosomes, another specialized organelle. plant cells have additional organelles that distinguish them from animal cells such as a cell wall that provides support for the plant cell, chloroplasts that harvest sunlight energy to produce sugar, and vacuoles that provide storage and structural support as well as being involved in reproduction and breakdown of plant seeds. eukaryotic cells also have cytoskeleton that is made up of microtubules, intermediate filaments, and microfilaments, all of which provide support for the cell and are involved in the movement of the cell and its
generally, dead - end filtration is used for feasibility studies on a laboratory scale. the dead - end membranes are relatively easy to fabricate which reduces the cost of the separation process. the dead - end membrane separation process is easy to implement and the process is usually cheaper than cross - flow membrane filtration. the dead - end filtration process is usually a batch - type process, where the filtering solution is loaded ( or slowly fed ) into the membrane device, which then allows passage of some particles subject to the driving force. the main disadvantage of dead - end filtration is the extensive membrane fouling and concentration polarization. the fouling is usually induced faster at higher driving forces. membrane fouling and particle retention in a feed solution also builds up a concentration gradients and particle backflow ( concentration polarization ). the tangential flow devices are more cost and labor - intensive, but they are less susceptible to fouling due to the sweeping effects and high shear rates of the passing flow. the most commonly used synthetic membrane devices ( modules ) are flat sheets / plates, spiral wounds, and hollow fibers. flat membranes used in filtration and separation processes can be enhanced with surface patterning, where microscopic structures are introduced to improve performance. these patterns increase surface area, optimize water flow, and reduce fouling, leading to higher permeability and longer membrane lifespan. research has shown that such modifications can significantly enhance efficiency in water purification, energy applications, and industrial separations. flat plates are usually constructed as circular thin flat membrane surfaces to be used in dead - end geometry modules. spiral wounds are constructed from similar flat membranes but in the form of a " pocket " containing two membrane sheets separated by a highly porous support plate. several such pockets are then wound around a tube to create a tangential flow geometry and to reduce membrane fouling. hollow fiber modules consist of an assembly of self - supporting fibers with dense skin separation layers, and a more open matrix helping to withstand pressure gradients and maintain structural integrity. the hollow fiber modules can contain up to 10, 000 fibers ranging from 200 to 2500 ΞΌm in diameter ; the main advantage of hollow fiber modules is the very large surface area within an enclosed volume, increasing the efficiency of the separation process. the disc tube module uses a cross - flow geometry and consists of a pressure tube and hydraulic discs, which are held by a central tension rod, and membrane cushions that lie between two discs. = = membrane performance and governing equations = = the selection of synthetic membranes
##ization. the fouling is usually induced faster at higher driving forces. membrane fouling and particle retention in a feed solution also builds up a concentration gradients and particle backflow ( concentration polarization ). the tangential flow devices are more cost and labor - intensive, but they are less susceptible to fouling due to the sweeping effects and high shear rates of the passing flow. the most commonly used synthetic membrane devices ( modules ) are flat sheets / plates, spiral wounds, and hollow fibers. flat membranes used in filtration and separation processes can be enhanced with surface patterning, where microscopic structures are introduced to improve performance. these patterns increase surface area, optimize water flow, and reduce fouling, leading to higher permeability and longer membrane lifespan. research has shown that such modifications can significantly enhance efficiency in water purification, energy applications, and industrial separations. flat plates are usually constructed as circular thin flat membrane surfaces to be used in dead - end geometry modules. spiral wounds are constructed from similar flat membranes but in the form of a " pocket " containing two membrane sheets separated by a highly porous support plate. several such pockets are then wound around a tube to create a tangential flow geometry and to reduce membrane fouling. hollow fiber modules consist of an assembly of self - supporting fibers with dense skin separation layers, and a more open matrix helping to withstand pressure gradients and maintain structural integrity. the hollow fiber modules can contain up to 10, 000 fibers ranging from 200 to 2500 ΞΌm in diameter ; the main advantage of hollow fiber modules is the very large surface area within an enclosed volume, increasing the efficiency of the separation process. the disc tube module uses a cross - flow geometry and consists of a pressure tube and hydraulic discs, which are held by a central tension rod, and membrane cushions that lie between two discs. = = membrane performance and governing equations = = the selection of synthetic membranes for a targeted separation process is usually based on few requirements. membranes have to provide enough mass transfer area to process large amounts of feed stream. the selected membrane has to have high selectivity ( rejection ) properties for certain particles ; it has to resist fouling and to have high mechanical stability. it also needs to be reproducible and to have low manufacturing costs. the main modeling equation for the dead - end filtration at constant pressure drop is represented by darcy ' s law : d v p d t = q = Ξ΄ p ΞΌ a ( 1 r m + r ) { \ displaystyle { \ frac { dv _
shaping the cell. cell membranes are involved in various cellular processes such as cell adhesion, storing electrical energy, and cell signalling and serve as the attachment surface for several extracellular structures such as a cell wall, glycocalyx, and cytoskeleton. within the cytoplasm of a cell, there are many biomolecules such as proteins and nucleic acids. in addition to biomolecules, eukaryotic cells have specialized structures called organelles that have their own lipid bilayers or are spatially units. these organelles include the cell nucleus, which contains most of the cell ' s dna, or mitochondria, which generate adenosine triphosphate ( atp ) to power cellular processes. other organelles such as endoplasmic reticulum and golgi apparatus play a role in the synthesis and packaging of proteins, respectively. biomolecules such as proteins can be engulfed by lysosomes, another specialized organelle. plant cells have additional organelles that distinguish them from animal cells such as a cell wall that provides support for the plant cell, chloroplasts that harvest sunlight energy to produce sugar, and vacuoles that provide storage and structural support as well as being involved in reproduction and breakdown of plant seeds. eukaryotic cells also have cytoskeleton that is made up of microtubules, intermediate filaments, and microfilaments, all of which provide support for the cell and are involved in the movement of the cell and its organelles. in terms of their structural composition, the microtubules are made up of tubulin ( e. g., Ξ± - tubulin and Ξ² - tubulin ) whereas intermediate filaments are made up of fibrous proteins. microfilaments are made up of actin molecules that interact with other strands of proteins. = = = metabolism = = = all cells require energy to sustain cellular processes. metabolism is the set of chemical reactions in an organism. the three main purposes of metabolism are : the conversion of food to energy to run cellular processes ; the conversion of food / fuel to monomer building blocks ; and the elimination of metabolic wastes. these enzyme - catalyzed reactions allow organisms to grow and reproduce, maintain their structures, and respond to their environments. metabolic reactions may be categorized as catabolic β the breaking down of compounds ( for example, the breaking down of glucose to pyruvate by cellular respiration
ability of cells to receive, process, and transmit signals with its environment and with itself. signals can be non - chemical such as light, electrical impulses, and heat, or chemical signals ( or ligands ) that interact with receptors, which can be found embedded in the cell membrane of another cell or located deep inside a cell. there are generally four types of chemical signals : autocrine, paracrine, juxtacrine, and hormones. in autocrine signaling, the ligand affects the same cell that releases it. tumor cells, for example, can reproduce uncontrollably because they release signals that initiate their own self - division. in paracrine signaling, the ligand diffuses to nearby cells and affects them. for example, brain cells called neurons release ligands called neurotransmitters that diffuse across a synaptic cleft to bind with a receptor on an adjacent cell such as another neuron or muscle cell. in juxtacrine signaling, there is direct contact between the signaling and responding cells. finally, hormones are ligands that travel through the circulatory systems of animals or vascular systems of plants to reach their target cells. once a ligand binds with a receptor, it can influence the behavior of another cell, depending on the type of receptor. for instance, neurotransmitters that bind with an inotropic receptor can alter the excitability of a target cell. other types of receptors include protein kinase receptors ( e. g., receptor for the hormone insulin ) and g protein - coupled receptors. activation of g protein - coupled receptors can initiate second messenger cascades. the process by which a chemical or physical signal is transmitted through a cell as a series of molecular events is called signal transduction. = = = cell cycle = = = the cell cycle is a series of events that take place in a cell that cause it to divide into two daughter cells. these events include the duplication of its dna and some of its organelles, and the subsequent partitioning of its cytoplasm into two daughter cells in a process called cell division. in eukaryotes ( i. e., animal, plant, fungal, and protist cells ), there are two distinct types of cell division : mitosis and meiosis. mitosis is part of the cell cycle, in which replicated chromosomes are separated into two new nuclei. cell division gives rise to genetically identical cells in which the total number of chromosomes is maintained. in general, mitosis
. most cells are very small, with diameters ranging from 1 to 100 micrometers and are therefore only visible under a light or electron microscope. there are generally two types of cells : eukaryotic cells, which contain a nucleus, and prokaryotic cells, which do not. prokaryotes are single - celled organisms such as bacteria, whereas eukaryotes can be single - celled or multicellular. in multicellular organisms, every cell in the organism ' s body is derived ultimately from a single cell in a fertilized egg. = = = cell structure = = = every cell is enclosed within a cell membrane that separates its cytoplasm from the extracellular space. a cell membrane consists of a lipid bilayer, including cholesterols that sit between phospholipids to maintain their fluidity at various temperatures. cell membranes are semipermeable, allowing small molecules such as oxygen, carbon dioxide, and water to pass through while restricting the movement of larger molecules and charged particles such as ions. cell membranes also contain membrane proteins, including integral membrane proteins that go across the membrane serving as membrane transporters, and peripheral proteins that loosely attach to the outer side of the cell membrane, acting as enzymes shaping the cell. cell membranes are involved in various cellular processes such as cell adhesion, storing electrical energy, and cell signalling and serve as the attachment surface for several extracellular structures such as a cell wall, glycocalyx, and cytoskeleton. within the cytoplasm of a cell, there are many biomolecules such as proteins and nucleic acids. in addition to biomolecules, eukaryotic cells have specialized structures called organelles that have their own lipid bilayers or are spatially units. these organelles include the cell nucleus, which contains most of the cell ' s dna, or mitochondria, which generate adenosine triphosphate ( atp ) to power cellular processes. other organelles such as endoplasmic reticulum and golgi apparatus play a role in the synthesis and packaging of proteins, respectively. biomolecules such as proteins can be engulfed by lysosomes, another specialized organelle. plant cells have additional organelles that distinguish them from animal cells such as a cell wall that provides support for the plant cell, chloroplasts that harvest sunlight energy to produce sugar, and vacuoles that provide storage and structural support
the dead - end filtration process is usually a batch - type process, where the filtering solution is loaded ( or slowly fed ) into the membrane device, which then allows passage of some particles subject to the driving force. the main disadvantage of dead - end filtration is the extensive membrane fouling and concentration polarization. the fouling is usually induced faster at higher driving forces. membrane fouling and particle retention in a feed solution also builds up a concentration gradients and particle backflow ( concentration polarization ). the tangential flow devices are more cost and labor - intensive, but they are less susceptible to fouling due to the sweeping effects and high shear rates of the passing flow. the most commonly used synthetic membrane devices ( modules ) are flat sheets / plates, spiral wounds, and hollow fibers. flat membranes used in filtration and separation processes can be enhanced with surface patterning, where microscopic structures are introduced to improve performance. these patterns increase surface area, optimize water flow, and reduce fouling, leading to higher permeability and longer membrane lifespan. research has shown that such modifications can significantly enhance efficiency in water purification, energy applications, and industrial separations. flat plates are usually constructed as circular thin flat membrane surfaces to be used in dead - end geometry modules. spiral wounds are constructed from similar flat membranes but in the form of a " pocket " containing two membrane sheets separated by a highly porous support plate. several such pockets are then wound around a tube to create a tangential flow geometry and to reduce membrane fouling. hollow fiber modules consist of an assembly of self - supporting fibers with dense skin separation layers, and a more open matrix helping to withstand pressure gradients and maintain structural integrity. the hollow fiber modules can contain up to 10, 000 fibers ranging from 200 to 2500 ΞΌm in diameter ; the main advantage of hollow fiber modules is the very large surface area within an enclosed volume, increasing the efficiency of the separation process. the disc tube module uses a cross - flow geometry and consists of a pressure tube and hydraulic discs, which are held by a central tension rod, and membrane cushions that lie between two discs. = = membrane performance and governing equations = = the selection of synthetic membranes for a targeted separation process is usually based on few requirements. membranes have to provide enough mass transfer area to process large amounts of feed stream. the selected membrane has to have high selectivity ( rejection ) properties for certain particles ; it has to resist fouling and to have high mechanical stability. it also needs to be rep
Question: Which process moves an ion across a cell membrane against the concentration gradient?
A) diffusion
B) exocytosis
C) facilitated diffusion
D) active transport
|
D) active transport
|
Context:
intense research in the materials science community due to the unique properties that they exhibit. nanostructure deals with objects and structures that are in the 1 β 100 nm range. in many materials, atoms or molecules agglomerate to form objects at the nanoscale. this causes many interesting electrical, magnetic, optical, and mechanical properties. in describing nanostructures, it is necessary to differentiate between the number of dimensions on the nanoscale. nanotextured surfaces have one dimension on the nanoscale, i. e., only the thickness of the surface of an object is between 0. 1 and 100 nm. nanotubes have two dimensions on the nanoscale, i. e., the diameter of the tube is between 0. 1 and 100 nm ; its length could be much greater. finally, spherical nanoparticles have three dimensions on the nanoscale, i. e., the particle is between 0. 1 and 100 nm in each spatial dimension. the terms nanoparticles and ultrafine particles ( ufp ) often are used synonymously although ufp can reach into the micrometre range. the term ' nanostructure ' is often used, when referring to magnetic technology. nanoscale structure in biology is often called ultrastructure. = = = = microstructure = = = = microstructure is defined as the structure of a prepared surface or thin foil of material as revealed by a microscope above 25Γ magnification. it deals with objects from 100 nm to a few cm. the microstructure of a material ( which can be broadly classified into metallic, polymeric, ceramic and composite ) can strongly influence physical properties such as strength, toughness, ductility, hardness, corrosion resistance, high / low temperature behavior, wear resistance, and so on. most of the traditional materials ( such as metals and ceramics ) are microstructured. the manufacture of a perfect crystal of a material is physically impossible. for example, any crystalline material will contain defects such as precipitates, grain boundaries ( hall β petch relationship ), vacancies, interstitial atoms or substitutional atoms. the microstructure of materials reveals these larger defects and advances in simulation have allowed an increased understanding of how defects can be used to enhance material properties. = = = = macrostructure = = = = macrostructure is the appearance of a material in the scale millimeters to meters, it is the structure of
molecular diffusion processes give rise to significant changes in the primary microstructural features. this includes the gradual elimination of porosity, which is typically accompanied by a net shrinkage and overall densification of the component. thus, the pores in the object may close up, resulting in a denser product of significantly greater strength and fracture toughness. another major change in the body during the firing or sintering process will be the establishment of the polycrystalline nature of the solid. significant grain growth tends to occur during sintering, with this growth depending on temperature and duration of the sintering process. the growth of grains will result in some form of grain size distribution, which will have a significant impact on the ultimate physical properties of the material. in particular, abnormal grain growth in which certain grains grow very large in a matrix of finer grains will significantly alter the physical and mechanical properties of the obtained ceramic. in the sintered body, grain sizes are a product of the thermal processing parameters as well as the initial particle size, or possibly the sizes of aggregates or particle clusters which arise during the initial stages of processing. the ultimate microstructure ( and thus the physical properties ) of the final product will be limited by and subject to the form of the structural template or precursor which is created in the initial stages of chemical synthesis and physical forming. hence the importance of chemical powder and polymer processing as it pertains to the synthesis of industrial ceramics, glasses and glass - ceramics. there are numerous possible refinements of the sintering process. some of the most common involve pressing the green body to give the densification a head start and reduce the sintering time needed. sometimes organic binders such as polyvinyl alcohol are added to hold the green body together ; these burn out during the firing ( at 200 β 350 Β°c ). sometimes organic lubricants are added during pressing to increase densification. it is common to combine these, and add binders and lubricants to a powder, then press. ( the formulation of these organic chemical additives is an art in itself. this is particularly important in the manufacture of high performance ceramics such as those used by the billions for electronics, in capacitors, inductors, sensors, etc. ) a slurry can be used in place of a powder, and then cast into a desired shape, dried and then sintered. indeed, traditional pottery is done with this type of method, using a plastic mixture worked with the hands.
classes according to pore size : the form and shape of the membrane pores are highly dependent on the manufacturing process and are often difficult to specify. therefore, for characterization, test filtrations are carried out and the pore diameter refers to the diameter of the smallest particles which could not pass through the membrane. the rejection can be determined in various ways and provides an indirect measurement of the pore size. one possibility is the filtration of macromolecules ( often dextran, polyethylene glycol or albumin ), another is measurement of the cut - off by gel permeation chromatography. these methods are used mainly to measure membranes for ultrafiltration applications. another testing method is the filtration of particles with defined size and their measurement with a particle sizer or by laser induced breakdown spectroscopy ( libs ). a vivid characterization is to measure the rejection of dextran blue or other colored molecules. the retention of bacteriophage and bacteria, the so - called " bacteria challenge test ", can also provide information about the pore size. to determine the pore diameter, physical methods such as porosimeter ( mercury, liquid - liquid porosimeter and bubble point test ) are also used, but a certain form of the pores ( such as cylindrical or concatenated spherical holes ) is assumed. such methods are used for membranes whose pore geometry does not match the ideal, and we get " nominal " pore diameter, which characterizes the membrane, but does not necessarily reflect its actual filtration behavior and selectivity. the selectivity is highly dependent on the separation process, the composition of the membrane and its electrochemical properties in addition to the pore size. with high selectivity, isotopes can be enriched ( uranium enrichment ) in nuclear engineering or industrial gases like nitrogen can be recovered ( gas separation ). ideally, even racemics can be enriched with a suitable membrane. when choosing membranes selectivity has priority over a high permeability, as low flows can easily be offset by increasing the filter surface with a modular structure. in gas phase filtration different deposition mechanisms are operative, so that particles having sizes below the pore size of the membrane can be retained as well. = = membrane classification = = bio - membrane is classified in two categories, synthetic membrane and natural membrane. synthetic membranes further classified in organic and inorganic membranes. organic membrane sub classified polymeric membranes and inorganic membrane sub classified ceramic polymers. = = synthesis of biomass membrane
grains will result in some form of grain size distribution, which will have a significant impact on the ultimate physical properties of the material. in particular, abnormal grain growth in which certain grains grow very large in a matrix of finer grains will significantly alter the physical and mechanical properties of the obtained ceramic. in the sintered body, grain sizes are a product of the thermal processing parameters as well as the initial particle size, or possibly the sizes of aggregates or particle clusters which arise during the initial stages of processing. the ultimate microstructure ( and thus the physical properties ) of the final product will be limited by and subject to the form of the structural template or precursor which is created in the initial stages of chemical synthesis and physical forming. hence the importance of chemical powder and polymer processing as it pertains to the synthesis of industrial ceramics, glasses and glass - ceramics. there are numerous possible refinements of the sintering process. some of the most common involve pressing the green body to give the densification a head start and reduce the sintering time needed. sometimes organic binders such as polyvinyl alcohol are added to hold the green body together ; these burn out during the firing ( at 200 β 350 Β°c ). sometimes organic lubricants are added during pressing to increase densification. it is common to combine these, and add binders and lubricants to a powder, then press. ( the formulation of these organic chemical additives is an art in itself. this is particularly important in the manufacture of high performance ceramics such as those used by the billions for electronics, in capacitors, inductors, sensors, etc. ) a slurry can be used in place of a powder, and then cast into a desired shape, dried and then sintered. indeed, traditional pottery is done with this type of method, using a plastic mixture worked with the hands. if a mixture of different materials is used together in a ceramic, the sintering temperature is sometimes above the melting point of one minor component β a liquid phase sintering. this results in shorter sintering times compared to solid state sintering. such liquid phase sintering involves in faster diffusion processes and may result in abnormal grain growth. = = strength of ceramics = = a material ' s strength is dependent on its microstructure. the engineering processes to which a material is subjected can alter its microstructure. the variety of strengthening mechanisms that alter the strength of a material include the mechanism of grain boundary strengthening. thus, although yield
galactic collisions are normally modeled in a cdm model by assuming the dm consists of a small number of very massive objects. this note shows that the behaviour of a cdm halo during collisions depends critically on the mass of the particles that make it up, and in particular, all halo particles below a certain characteristic mass are likely to be lost.
small category. any ordinal number can be seen as a category when viewed as an ordered set. any monoid ( any algebraic structure with a single associative binary operation and an identity element ) forms a small category with a single object x. ( here, x is any fixed set. ) the morphisms from x to x are precisely the elements of the monoid, the identity morphism of x is the identity of the monoid, and the categorical composition of morphisms is given by the monoid operation. several definitions and theorems about monoids may be generalized for categories. similarly any group can be seen as a category with a single object in which every morphism is invertible, that is, for every morphism f there is a morphism g that is both left and right inverse to f under composition. a morphism that is invertible in this sense is called an isomorphism. a groupoid is a category in which every morphism is an isomorphism. groupoids are generalizations of groups, group actions and equivalence relations. actually, in the view of category the only difference between groupoid and group is that a groupoid may have more than one object but the group must have only one. consider a topological space x and fix a base point x 0 { \ displaystyle x _ { 0 } } of x, then Ο 1 ( x, x 0 ) { \ displaystyle \ pi _ { 1 } ( x, x _ { 0 } ) } is the fundamental group of the topological space x and the base point x 0 { \ displaystyle x _ { 0 } }, and as a set it has the structure of group ; if then let the base point x 0 { \ displaystyle x _ { 0 } } runs over all points of x, and take the union of all Ο 1 ( x, x 0 ) { \ displaystyle \ pi _ { 1 } ( x, x _ { 0 } ) }, then the set we get has only the structure of groupoid ( which is called as the fundamental groupoid of x ) : two loops ( under equivalence relation of homotopy ) may not have the same base point so they cannot multiply with each other. in the language of category, this means here two morphisms may not have the same source object ( or target object, because in this case for any morphism the source object and the target object are same : the base point ) so
significantly greater strength and fracture toughness. another major change in the body during the firing or sintering process will be the establishment of the polycrystalline nature of the solid. significant grain growth tends to occur during sintering, with this growth depending on temperature and duration of the sintering process. the growth of grains will result in some form of grain size distribution, which will have a significant impact on the ultimate physical properties of the material. in particular, abnormal grain growth in which certain grains grow very large in a matrix of finer grains will significantly alter the physical and mechanical properties of the obtained ceramic. in the sintered body, grain sizes are a product of the thermal processing parameters as well as the initial particle size, or possibly the sizes of aggregates or particle clusters which arise during the initial stages of processing. the ultimate microstructure ( and thus the physical properties ) of the final product will be limited by and subject to the form of the structural template or precursor which is created in the initial stages of chemical synthesis and physical forming. hence the importance of chemical powder and polymer processing as it pertains to the synthesis of industrial ceramics, glasses and glass - ceramics. there are numerous possible refinements of the sintering process. some of the most common involve pressing the green body to give the densification a head start and reduce the sintering time needed. sometimes organic binders such as polyvinyl alcohol are added to hold the green body together ; these burn out during the firing ( at 200 β 350 Β°c ). sometimes organic lubricants are added during pressing to increase densification. it is common to combine these, and add binders and lubricants to a powder, then press. ( the formulation of these organic chemical additives is an art in itself. this is particularly important in the manufacture of high performance ceramics such as those used by the billions for electronics, in capacitors, inductors, sensors, etc. ) a slurry can be used in place of a powder, and then cast into a desired shape, dried and then sintered. indeed, traditional pottery is done with this type of method, using a plastic mixture worked with the hands. if a mixture of different materials is used together in a ceramic, the sintering temperature is sometimes above the melting point of one minor component β a liquid phase sintering. this results in shorter sintering times compared to solid state sintering. such liquid phase sintering involves in faster diffusion processes and may result in abnormal grain
the manufacturer. one common distinction is by nominal pore size. it describes the maximum pore size distribution and gives only vague information about the retention capacity of a membrane. the exclusion limit or " cut - off " of the membrane is usually specified in the form of nmwc ( nominal molecular weight cut - off, or mwco, molecular weight cut off, with units in dalton ). it is defined as the minimum molecular weight of a globular molecule that is retained to 90 % by the membrane. the cut - off, depending on the method, can by converted to so - called d90, which is then expressed in a metric unit. in practice the mwco of the membrane should be at least 20 % lower than the molecular weight of the molecule that is to be separated. using track etched mica membranes beck and schultz demonstrated that hindered diffusion of molecules in pores can be described by the rankin equation. filter membranes are divided into four classes according to pore size : the form and shape of the membrane pores are highly dependent on the manufacturing process and are often difficult to specify. therefore, for characterization, test filtrations are carried out and the pore diameter refers to the diameter of the smallest particles which could not pass through the membrane. the rejection can be determined in various ways and provides an indirect measurement of the pore size. one possibility is the filtration of macromolecules ( often dextran, polyethylene glycol or albumin ), another is measurement of the cut - off by gel permeation chromatography. these methods are used mainly to measure membranes for ultrafiltration applications. another testing method is the filtration of particles with defined size and their measurement with a particle sizer or by laser induced breakdown spectroscopy ( libs ). a vivid characterization is to measure the rejection of dextran blue or other colored molecules. the retention of bacteriophage and bacteria, the so - called " bacteria challenge test ", can also provide information about the pore size. to determine the pore diameter, physical methods such as porosimeter ( mercury, liquid - liquid porosimeter and bubble point test ) are also used, but a certain form of the pores ( such as cylindrical or concatenated spherical holes ) is assumed. such methods are used for membranes whose pore geometry does not match the ideal, and we get " nominal " pore diameter, which characterizes the membrane, but does not necessarily reflect its actual filt
be at most one morphism between any two objects. the existence of identity morphisms and the composability of the morphisms are guaranteed by the reflexivity and the transitivity of the preorder. by the same argument, any partially ordered set and any equivalence relation can be seen as a small category. any ordinal number can be seen as a category when viewed as an ordered set. any monoid ( any algebraic structure with a single associative binary operation and an identity element ) forms a small category with a single object x. ( here, x is any fixed set. ) the morphisms from x to x are precisely the elements of the monoid, the identity morphism of x is the identity of the monoid, and the categorical composition of morphisms is given by the monoid operation. several definitions and theorems about monoids may be generalized for categories. similarly any group can be seen as a category with a single object in which every morphism is invertible, that is, for every morphism f there is a morphism g that is both left and right inverse to f under composition. a morphism that is invertible in this sense is called an isomorphism. a groupoid is a category in which every morphism is an isomorphism. groupoids are generalizations of groups, group actions and equivalence relations. actually, in the view of category the only difference between groupoid and group is that a groupoid may have more than one object but the group must have only one. consider a topological space x and fix a base point x 0 { \ displaystyle x _ { 0 } } of x, then Ο 1 ( x, x 0 ) { \ displaystyle \ pi _ { 1 } ( x, x _ { 0 } ) } is the fundamental group of the topological space x and the base point x 0 { \ displaystyle x _ { 0 } }, and as a set it has the structure of group ; if then let the base point x 0 { \ displaystyle x _ { 0 } } runs over all points of x, and take the union of all Ο 1 ( x, x 0 ) { \ displaystyle \ pi _ { 1 } ( x, x _ { 0 } ) }, then the set we get has only the structure of groupoid ( which is called as the fundamental groupoid of x ) : two loops ( under equivalence relation of homotopy ) may
0. 1 and 100 nm in each spatial dimension. the terms nanoparticles and ultrafine particles ( ufp ) often are used synonymously although ufp can reach into the micrometre range. the term ' nanostructure ' is often used, when referring to magnetic technology. nanoscale structure in biology is often called ultrastructure. = = = = microstructure = = = = microstructure is defined as the structure of a prepared surface or thin foil of material as revealed by a microscope above 25Γ magnification. it deals with objects from 100 nm to a few cm. the microstructure of a material ( which can be broadly classified into metallic, polymeric, ceramic and composite ) can strongly influence physical properties such as strength, toughness, ductility, hardness, corrosion resistance, high / low temperature behavior, wear resistance, and so on. most of the traditional materials ( such as metals and ceramics ) are microstructured. the manufacture of a perfect crystal of a material is physically impossible. for example, any crystalline material will contain defects such as precipitates, grain boundaries ( hall β petch relationship ), vacancies, interstitial atoms or substitutional atoms. the microstructure of materials reveals these larger defects and advances in simulation have allowed an increased understanding of how defects can be used to enhance material properties. = = = = macrostructure = = = = macrostructure is the appearance of a material in the scale millimeters to meters, it is the structure of the material as seen with the naked eye. = = = properties = = = materials exhibit myriad properties, including the following. mechanical properties, see strength of materials chemical properties, see chemistry electrical properties, see electricity thermal properties, see thermodynamics optical properties, see optics and photonics magnetic properties, see magnetism the properties of a material determine its usability and hence its engineering application. = = = processing = = = synthesis and processing involves the creation of a material with the desired micro - nanostructure. a material cannot be used in industry if no economically viable production method for it has been developed. therefore, developing processing methods for materials that are reasonably effective and cost - efficient is vital to the field of materials science. different materials require different processing or synthesis methods. for example, the processing of metals has historically defined eras such as the bronze age and iron age and is studied under the branch of materials science named physical metallurgy.
Question: Which would most likely be the characteristic of an object with a small volume and large mass?
A) high density
B) low pressure
C) low viscosity
D) high temperature
|
A) high density
|
Context:
wave, carrying an information signal, occupies a range of frequencies. the information in a radio signal is usually concentrated in narrow frequency bands called sidebands ( sb ) just above and below the carrier frequency. the width in hertz of the frequency range that the radio signal occupies, the highest frequency minus the lowest frequency, is called its bandwidth ( bw ). for any given signal - to - noise ratio, a given bandwidth can carry the same amount of information regardless of where in the radio frequency spectrum it is located ; bandwidth is a measure of information - carrying capacity. the bandwidth required by a radio transmission depends on the data rate of the information being sent, and the spectral efficiency of the modulation method used ; how much data it can transmit in each unit of bandwidth. different types of information signals carried by radio have different data rates. for example, a television signal has a greater data rate than an audio signal. the radio spectrum, the total range of radio frequencies that can be used for communication in a given area, is a limited resource. each radio transmission occupies a portion of the total bandwidth available. radio bandwidth is regarded as an economic good which has a monetary cost and is in increasing demand. in some parts of the radio spectrum, the right to use a frequency band or even a single radio channel is bought and sold for millions of dollars. so there is an incentive to employ technology to minimize the bandwidth used by radio services. a slow transition from analog to digital radio transmission technologies began in the late 1990s. part of the reason for this is that digital modulation can often transmit more information ( a greater data rate ) in a given bandwidth than analog modulation, by using data compression algorithms, which reduce redundancy in the data to be sent, and more efficient modulation. other reasons for the transition is that digital modulation has greater noise immunity than analog, digital signal processing chips have more power and flexibility than analog circuits, and a wide variety of types of information can be transmitted using the same digital modulation. because it is a fixed resource which is in demand by an increasing number of users, the radio spectrum has become increasingly congested in recent decades, and the need to use it more effectively is driving many additional radio innovations such as trunked radio systems, spread spectrum ( ultra - wideband ) transmission, frequency reuse, dynamic spectrum management, frequency pooling, and cognitive radio. = = = itu frequency bands = = = the itu arbitrarily divides the radio spectrum into 12 bands, each beginning at a wavelength which is a power
the higher microwave band 3 β 6 ghz, and millimeter wave band, around 28 and 39 ghz. since these frequencies have a shorter range than previous cellphone bands, the cells will be smaller than the cells in previous cellular networks which could be many miles across. millimeter - wave cells will only be a few blocks long, and instead of a cell base station and antenna tower, they will have many small antennas attached to utility poles and buildings. satellite phone ( satphone ) β a portable wireless telephone similar to a cell phone, connected to the telephone network through a radio link to an orbiting communications satellite instead of through cell towers. they are more expensive than cell phones ; but their advantage is that, unlike a cell phone which is limited to areas covered by cell towers, satphones can be used over most or all of the geographical area of the earth. in order for the phone to communicate with a satellite using a small omnidirectional antenna, first - generation systems use satellites in low earth orbit, about 400 β 700 miles ( 640 β 1, 100 km ) above the surface. with an orbital period of about 100 minutes, a satellite can only be in view of a phone for about 4 β 15 minutes, so the call is " handed off " to another satellite when one passes beyond the local horizon. therefore, large numbers of satellites, about 40 to 70, are required to ensure that at least one satellite is in view continuously from each point on earth. other satphone systems use satellites in geostationary orbit in which only a few satellites are needed, but these cannot be used at high latitudes because of terrestrial interference. cordless phone β a landline telephone in which the handset is portable and communicates with the rest of the phone by a short - range full duplex radio link, instead of being attached by a cord. both the handset and the base station have low - power radio transceivers that handle the short - range bidirectional radio link. as of 2022, cordless phones in most nations use the dect transmission standard. land mobile radio system β short - range mobile or portable half - duplex radio transceivers operating in the vhf or uhf band that can be used without a license. they are often installed in vehicles, with the mobile units communicating with a dispatcher at a fixed base station. special systems with reserved frequencies are used by first responder services ; police, fire, ambulance, and emergency services, and other government services. other systems are made for
the extremely small probability of tunneling through an almost classical potential barrier may become not small under the action of the specially adapted non - stationary signal which selects the certain particle energy e _ r. for particle energies close to this value, the tunneling rate is not small during a finite interval of time and has a very sharp peak at the energy e _ r. after entering inside the barrier, the particle emits electromagnetic quanta and exits the barrier with a lower energy. the signal amplitude can be much less compared to the field of the static barrier. this phenomenon can be called the euclidean resonance since the under - barrier motion occurs in imaginary time. the resonance may stimulate chemical and biochemical reactions in a selective way by adapting the signal to a certain particular chemical bond. the resonance may be used in search of the soft alpha - decay for which a conventional observation is impossible due to an extremely small decay rate.
stations located in places like light poles or building roofs. in the past, 4g networking had to rely on large cell towers in order to transmit signals over large distances. with the introduction of 5g networking, it is imperative that small cell stations are used because the mm wave spectrum, which is the specific type of band used in 5g services, strictly travels over short distances. if the distances between cell stations were longer, signals may suffer from interference from inclimate weather, or other objects such as houses, buildings, trees, and much more. in 5g networking, there are 3 main kinds of 5g : low - band, mid - band, and high - band. low - band frequencies operate below 2 ghz, mid - band frequencies operate between 2 β 10 ghz, and high - band frequencies operate between 20 and 100 ghz. verizon have seen outrageous numbers on their high - band 5g service, which they deem " ultraband ", which hit speeds of over 3 gbit / s. the main advantage of 5g networks is that the data transmission rate is much higher than the previous cellular network, up to 10 gbit / s, which is faster than the current wired internet and 100 times faster than the previous 4g lte cellular network. another advantage is lower network latency ( faster response time ), less than 1 millisecond, and 4g is 30 - 70 milliseconds. the peak rate needs to reach the gbit / s standard to meet the high data volume of high - definition video, virtual reality and so on. the air interface delay level needs to be around 1ms, which meets real - time applications such as autonomous driving and telemedicine. large network capacity, providing the connection capacity of 100 billion devices to meet iot communication. the spectrum efficiency is 10 times higher than lte. with continuous wide area coverage and high mobility, the user experience rate reaches 100 mbit / s. the flow density and the number of connections are greatly increased. since 5g is a relatively new type of service, only phones which are newly released or are upcoming can support 5g service. some of these phones include the iphone 12 / 13 ; select samsung devices such as the s21 series, note series, flip / fold series, a series ; google pixel 4a / 5 ; and a few more devices from other manufacturers. the first ever 5g smartphone, the samsung galaxy s20, was released by samsung in march 2020. following the release of samsung ' s s
rolling out 5g services across the us. 5g started being deployed at the beginning of 2020 and has been growing ever since. according to the gsm association, by 2025, approximately 1. 7 billion subscribers will have a subscription with 5g service. 5g wireless signals are transmitted through large numbers of small cell stations located in places like light poles or building roofs. in the past, 4g networking had to rely on large cell towers in order to transmit signals over large distances. with the introduction of 5g networking, it is imperative that small cell stations are used because the mm wave spectrum, which is the specific type of band used in 5g services, strictly travels over short distances. if the distances between cell stations were longer, signals may suffer from interference from inclimate weather, or other objects such as houses, buildings, trees, and much more. in 5g networking, there are 3 main kinds of 5g : low - band, mid - band, and high - band. low - band frequencies operate below 2 ghz, mid - band frequencies operate between 2 β 10 ghz, and high - band frequencies operate between 20 and 100 ghz. verizon have seen outrageous numbers on their high - band 5g service, which they deem " ultraband ", which hit speeds of over 3 gbit / s. the main advantage of 5g networks is that the data transmission rate is much higher than the previous cellular network, up to 10 gbit / s, which is faster than the current wired internet and 100 times faster than the previous 4g lte cellular network. another advantage is lower network latency ( faster response time ), less than 1 millisecond, and 4g is 30 - 70 milliseconds. the peak rate needs to reach the gbit / s standard to meet the high data volume of high - definition video, virtual reality and so on. the air interface delay level needs to be around 1ms, which meets real - time applications such as autonomous driving and telemedicine. large network capacity, providing the connection capacity of 100 billion devices to meet iot communication. the spectrum efficiency is 10 times higher than lte. with continuous wide area coverage and high mobility, the user experience rate reaches 100 mbit / s. the flow density and the number of connections are greatly increased. since 5g is a relatively new type of service, only phones which are newly released or are upcoming can support 5g service. some of these phones include the iphone 12 / 13 ; select
cobalt nanowires with a diameter in the range between 50 to 100nm can be prepared as single - crystal wires with the easy axis ( the c - axis ) perpendicular to the wire axis. the competition between the crystal anisotropy and demagnetization energy frustrates the magnetization direction. a periodic modulation of the angle between m and the wire axis yields a lower energy.
##directional range ( vor ) β a worldwide aircraft radio navigation system consisting of fixed ground radio beacons transmitting between 108. 00 and 117. 95 mhz in the very high frequency ( vhf ) band. an automated navigational instrument on the aircraft displays a bearing to a nearby vor transmitter. a vor beacon transmits two signals simultaneously on different frequencies. a directional antenna transmits a beam of radio waves that rotates like a lighthouse at a fixed rate, 30 times per second. when the directional beam is facing north, an omnidirectional antenna transmits a pulse. by measuring the difference in phase of these two signals, an aircraft can determine its bearing ( or " radial " ) from the station accurately. by taking a bearing on two vor beacons an aircraft can determine its position ( called a " fix " ) to an accuracy of about 90 metres ( 300 ft ). most vor beacons also have a distance measuring capability, called distance measuring equipment ( dme ) ; these are called vor / dme ' s. the aircraft transmits a radio signal to the vor / dme beacon and a transponder transmits a return signal. from the propagation delay between the transmitted and received signal the aircraft can calculate its distance from the beacon. this allows an aircraft to determine its location " fix " from only one vor beacon. since line - of - sight vhf frequencies are used vor beacons have a range of about 200 miles for aircraft at cruising altitude. tacan is a similar military radio beacon system which transmits in 962 β 1213 mhz, and a combined vor and tacan beacon is called a vortac. the number of vor beacons is declining as aviation switches to the rnav system that relies on global positioning system satellite navigation. instrument landing system ( ils ) - a short range radio navigation aid at airports which guides aircraft landing in low visibility conditions. it consists of multiple antennas at the end of each runway that radiate two beams of radio waves along the approach to the runway : the localizer ( 108 to 111. 95 mhz frequency ), which provides horizontal guidance, a heading line to keep the aircraft centered on the runway, and the glideslope ( 329. 15 to 335 mhz ) for vertical guidance, to keep the aircraft descending at the proper rate for a smooth touchdown at the correct point on the runway. each aircraft has a receiver instrument and antenna which receives the beams, with an indicator to tell the pilot whether he is
, lightning strikes, tornadoes, building fires, wildfires, and mass shootings disabling most of the system if not the entirety of it. geographic redundancy locations can be more than 621 miles ( 999 km ) continental, more than 62 miles apart and less than 93 miles ( 150 km ) apart, less than 62 miles apart, but not on the same campus, or different buildings that are more than 300 feet ( 91 m ) apart on the same campus. the following methods can reduce the risks of damage by a fire conflagration : large buildings at least 80 feet ( 24 m ) to 110 feet ( 34 m ) apart, but sometimes a minimum of 210 feet ( 64 m ) apart. : 9 high - rise buildings at least 82 feet ( 25 m ) apart : 12 open spaces clear of flammable vegetation within 200 feet ( 61 m ) on each side of objects different wings on the same building, in rooms that are separated by more than 300 feet ( 91 m ) different floors on the same wing of a building in rooms that are horizontally offset by a minimum of 70 feet ( 21 m ) with fire walls between the rooms that are on different floors two rooms separated by another room, leaving at least a 70 - foot gap between the two rooms there should be a minimum of two separated fire walls and on opposite sides of a corridor geographic redundancy is used by amazon web services ( aws ), google cloud platform ( gcp ), microsoft azure, netflix, dropbox, salesforce, linkedin, paypal, twitter, facebook, apple icloud, cisco meraki, and many others to provide geographic redundancy, high availability, fault tolerance and to ensure availability and reliability for their cloud services. as another example, to minimize risk of damage from severe windstorms or water damage, buildings can be located at least 2 miles ( 3. 2 km ) away from the shore, with an elevation of at least 5 feet ( 1. 5 m ) above sea level. for additional protection, they can be located at least 100 feet ( 30 m ) away from flood plain areas. = = functions of redundancy = = the two functions of redundancy are passive redundancy and active redundancy. both functions prevent performance decline from exceeding specification limits without human intervention using extra capacity. passive redundancy uses excess capacity to reduce the impact of component failures. one common form of passive redundancy is the extra strength of cabling and struts used in bridges.
- power radio transceivers that handle the short - range bidirectional radio link. as of 2022, cordless phones in most nations use the dect transmission standard. land mobile radio system β short - range mobile or portable half - duplex radio transceivers operating in the vhf or uhf band that can be used without a license. they are often installed in vehicles, with the mobile units communicating with a dispatcher at a fixed base station. special systems with reserved frequencies are used by first responder services ; police, fire, ambulance, and emergency services, and other government services. other systems are made for use by commercial firms such as taxi and delivery services. vhf systems use channels in the range 30 β 50 mhz and 150 β 172 mhz. uhf systems use the 450 β 470 mhz band and in some areas the 470 β 512 mhz range. in general, vhf systems have a longer range than uhf but require longer antennas. am or fm modulation is mainly used, but digital systems such as dmr are being introduced. the radiated power is typically limited to 4 watts. these systems have a fairly limited range, usually 3 to 20 miles ( 4. 8 to 32 km ) depending on terrain. repeaters installed on tall buildings, hills, or mountain peaks are often used to increase the range when it is desired to cover a larger area than line - of - sight. examples of land mobile systems are cb, frs, gmrs, and murs. modern digital systems, called trunked radio systems, have a digital channel management system using a control channel that automatically assigns frequency channels to user groups. walkie - talkie β a battery - powered portable handheld half - duplex two - way radio, used in land mobile radio systems. airband β half - duplex radio system used by aircraft pilots to talk to other aircraft and ground - based air traffic controllers. this vital system is the main communication channel for air traffic control. for most communication in overland flights in air corridors a vhf - am system using channels between 108 and 137 mhz in the vhf band is used. this system has a typical transmission range of 200 miles ( 320 km ) for aircraft flying at cruising altitude. for flights in more remote areas, such as transoceanic airline flights, aircraft use the hf band or channels on the inmarsat or iridium satphone satellites. military aircraft also use a dedicated uhf - am band from 225. 0 to 399. 95 mhz. marine radio β medium - range transceivers on
band that can be used without a license. they are often installed in vehicles, with the mobile units communicating with a dispatcher at a fixed base station. special systems with reserved frequencies are used by first responder services ; police, fire, ambulance, and emergency services, and other government services. other systems are made for use by commercial firms such as taxi and delivery services. vhf systems use channels in the range 30 β 50 mhz and 150 β 172 mhz. uhf systems use the 450 β 470 mhz band and in some areas the 470 β 512 mhz range. in general, vhf systems have a longer range than uhf but require longer antennas. am or fm modulation is mainly used, but digital systems such as dmr are being introduced. the radiated power is typically limited to 4 watts. these systems have a fairly limited range, usually 3 to 20 miles ( 4. 8 to 32 km ) depending on terrain. repeaters installed on tall buildings, hills, or mountain peaks are often used to increase the range when it is desired to cover a larger area than line - of - sight. examples of land mobile systems are cb, frs, gmrs, and murs. modern digital systems, called trunked radio systems, have a digital channel management system using a control channel that automatically assigns frequency channels to user groups. walkie - talkie β a battery - powered portable handheld half - duplex two - way radio, used in land mobile radio systems. airband β half - duplex radio system used by aircraft pilots to talk to other aircraft and ground - based air traffic controllers. this vital system is the main communication channel for air traffic control. for most communication in overland flights in air corridors a vhf - am system using channels between 108 and 137 mhz in the vhf band is used. this system has a typical transmission range of 200 miles ( 320 km ) for aircraft flying at cruising altitude. for flights in more remote areas, such as transoceanic airline flights, aircraft use the hf band or channels on the inmarsat or iridium satphone satellites. military aircraft also use a dedicated uhf - am band from 225. 0 to 399. 95 mhz. marine radio β medium - range transceivers on ships, used for ship - to - ship, ship - to - air, and ship - to - shore communication with harbormasters they use fm channels between 156 and 174 mhz in the vhf band with up to 25 watts power, giving them a range of about 60 miles ( 97 km ). some channels are half -
Question: Which of these describes a small range in the electromagnetic spectrum?
A) gravity
B) radiation
C) electricity
D) light
|
D) light
|
Context:
astronomical observations have shown that the expansion of the universe is at present accelerating, consistently with a constant negative pressure or tension. this is a major puzzle because we do not understand why this tension is so small compared to the planck density ; why, being so small, it is not exactly zero ; and why it has precisely the required value to make the expansion start accelerating just at the epoch when we are observing the universe. the recently proposed conjecture by afshordi that black holes create a gravitational aether owing to quantum gravity effects, which may be identified with this invisible tension, can solve this coincidence problem. the fact that the expansion of the universe is starting to accelerate at the epoch when we observe it is a necessity that is implied by our origin in a planet orbiting a star that formed when the age of the universe was of the same order as the lifetime of the star. this argument is unrelated to any anthropic reasoning.
in 1998 may, the soft gamma repeater sgr1900 + 14 emerged from several years of quiescence and emitted a series of intense bursts, one with a time history unlike any previously observed from this source. triangulation using ulysses, batse, and konus data give a 1. 6 square arcminute error box near the galactic supernova remnant g42. 8 + 0. 6. this error box contains a quiescent soft x - ray source which is probably a neutron star associated with the soft repeater.
the luminosity variation of a stellar source due to the gravitational microlensing effect can be considered also if the light rays are defocused ( instead of focused ) toward the observer. in this case, we should detect a gap instead of a peak in the light curve of the source. actually, we describe how the phenomenon depends on the relative position of source and lens with respect to the observer : if the lens is between, we have focusing, if the lens is behind, we have defocusing. it is shown that the number of events with predicted gaps is equal to the number of events with peaks in the light curves.
in a diagram of metallicity ( \ ~ z ) vs. luminosity ( m $ _ b $ ), the different types of nearby ( z $ < 0. 05 $ ) starburst galaxies seem to follow the same linear relationship as the normal spiral and irregular galaxies. however, for comparable luminosities the more massive starburst nucleus galaxies ( sbngs ) show a slight metallic defficiency as compared to the giant spiral galaxies. furthermore, the sbngs do not seem to follow the same relationship between \ ~ z and hubble type ( t ) than the normal galaxies. the early - type sbngs are metal poor as compared to normal galaxies. it may suggests that the chemical evolution of a majority of the nearby starbursts galaxies is not completely over and that the present burst represent an important phase of this process.
the belief that three dimensional space is infinite and flat in the absence of matter is a canon of physics that has been in place since the time of newton. the assumption that space is flat at infinity has guided several modern physical theories. but what do we actually know to support this belief? a simple argument, called the " telescope principle ", asserts that all that we can know about space is bounded by observations. physical theories are best when they can be verified by observations, and that should also apply to the geometry of space. the telescope principle is simple to state, but it leads to very interesting insights into relativity and yang - mills theory via projective equivalences of their respective spaces.
dynamical evolution of spiral galaxies is strongly dependent on non - axisymmetric patterns that develop from gravitational instabilities, either spontaneously or externally triggered. some evolutionary sequences are described through which a galaxy could possibly concentrate mass and build bulges, how external gas accretion from cosmic filaments could be funneled to the galaxy disks, and intermittently driven to the galaxy center, to form nuclear starbursts and fuel an active nucleus. the frequency of both bars and lopsidedness can be used to constrain the gas accretion rate.
the latest news from $ ^ 3 $ he universe are presented together with the extended map of the universe.
while the modern stellar imf shows a rapid decline with increasing mass, theoretical investigations suggest that very massive stars ( > 100 solar masses ) may have been abundant in the early universe. other calculations also indicate that, lacking metals, these same stars reach their late evolutionary stages without appreciable mass loss. after central helium burning, they encounter the electron - positron pair instability, collapse, and burn oxygen and silicon explosively. if sufficient energy is released by the burning, these stars explode as brilliant supernovae with energies up to 100 times that of an ordinary core collapse supernova. they also eject up to 50 solar masses of radioactive ni56. stars less massive than 140 solar masses or more massive than 260 solar masses should collapse into black holes instead of exploding, thus bounding the pair - creation supernovae with regions of stellar mass that are nucleosynthetically sterile. pair - instability supernovae might be detectable in the near infrared out to redshifts of 20 or more and their ashes should leave a distinctive nucleosynthetic pattern.
the theory outright... lakatos sought to reconcile the rationalism of popperian falsificationism with what seemed to be its own refutation by history ". many philosophers have tried to solve the problem of demarcation in the following terms : a statement constitutes knowledge if sufficiently many people believe it sufficiently strongly. but the history of thought shows us that many people were totally committed to absurd beliefs. if the strengths of beliefs were a hallmark of knowledge, we should have to rank some tales about demons, angels, devils, and of heaven and hell as knowledge. scientists, on the other hand, are very sceptical even of their best theories. newton ' s is the most powerful theory science has yet produced, but newton himself never believed that bodies attract each other at a distance. so no degree of commitment to beliefs makes them knowledge. indeed, the hallmark of scientific behaviour is a certain scepticism even towards one ' s most cherished theories. blind commitment to a theory is not an intellectual virtue : it is an intellectual crime. thus a statement may be pseudoscientific even if it is eminently ' plausible ' and everybody believes in it, and it may be scientifically valuable even if it is unbelievable and nobody believes in it. a theory may even be of supreme scientific value even if no one understands it, let alone believes in it. the boundary between science and pseudoscience is disputed and difficult to determine analytically, even after more than a century of study by philosophers of science and scientists, and despite some basic agreements on the fundamentals of the scientific method. the concept of pseudoscience rests on an understanding that the scientific method has been misrepresented or misapplied with respect to a given theory, but many philosophers of science maintain that different kinds of methods are held as appropriate across different fields and different eras of human history. according to lakatos, the typical descriptive unit of great scientific achievements is not an isolated hypothesis but " a powerful problem - solving machinery, which, with the help of sophisticated mathematical techniques, digests anomalies and even turns them into positive evidence ". to popper, pseudoscience uses induction to generate theories, and only performs experiments to seek to verify them. to popper, falsifiability is what determines the scientific status of a theory. taking a historical approach, kuhn observed that scientists did not follow popper ' s rule, and might ignore falsifying data, unless overwhelming. to kuhn, puzzle - solving within
it seems natural to ask why the universe exists at all. modern physics suggests that the universe can exist all by itself as a self - contained system, without anything external to create or sustain it. but there might not be an absolute answer to why it exists. i argue that any attempt to account for the existence of something rather than nothing must ultimately bottom out in a set of brute facts ; the universe simply is, without ultimate cause or explanation.
Question: An idea that the universe originated with an explosion is supported by repeated observations of light from distant galaxies traveling away from Earth. An idea that has been tested repeatedly and is supported by repeated results is most likely a
A) hypothesis.
B) theory.
C) law.
D) fact.
|
B) theory.
|
Context:
can the apparent complexity we observe in the real world be generated from simple initial conditions via simple, deterministic rules?
it is hard for us humans to recognize things in nature until we have invented them ourselves. for image - forming optics, nature has made virtually every kind of lens humans have devised. but what about lensless " imaging "? recently, we showed that a bare array of sensors on a curved substrate could achieve resolution not limited by diffraction - without any lens at all provided that the objects imaged conform to our a priori assumptions. is it possible that somewhere in nature we will find this kind of vision system? we think so and provide examples that seem to make no sense whatever unless they are using something like our lensless imaging work.
it seems natural to ask why the universe exists at all. modern physics suggests that the universe can exist all by itself as a self - contained system, without anything external to create or sustain it. but there might not be an absolute answer to why it exists. i argue that any attempt to account for the existence of something rather than nothing must ultimately bottom out in a set of brute facts ; the universe simply is, without ultimate cause or explanation.
living things, computers, societies, and even books are part of a grand evolutionary struggle to survive. that struggle shapes nature, nations, religions, art, science, and you. what you think, feel, and do is determined by it. darwinian evolution does not apply solely to the genes that are stored in dna. using the insights of alan turing and richard dawkins, we will see that it also applies to the memes we store in our brains and the information we store in our computers. the next time you run for president, fight a war, or just deal with the ordinary problems humans are heir to, perhaps this book will be of use. if you want to understand why and when you will die, or if you want to achieve greatness this book may help. if you are concerned about where the computer revolution is headed, this book may provide some answers.
time - dependent distribution of the global extinction of megafauna is compared with the growth of human population. there is no correlation between the two processes. furthermore, the size of human population and its growth rate were far too small to have any significant impact on the environment and on the life of megafauna.
quantum mechanics is nonlocal. classical mechanics is local. consequently classical mechanics can not explain all quantum phenomena. conversely, it is cumbersome to use quantum mechanics to describe classical phenomena. not only are the computations more complex, but - and this is the main point - it is conceptually more difficult : one has to argue that nonlocality, entanglement and the principle of superposition can be set aside when crossing the " quantum $ \ rightarrow $ classical " border. clearly, nonlocality, entanglement and the principle of superposition should become irrelevant in the classical limit. but why should one argue? shouldn ' t it just come out of the equations? does it come out of the equations? this contribution is about the last question. and the answer is : " it depends on which equation ".
dynamical evolution of spiral galaxies is strongly dependent on non - axisymmetric patterns that develop from gravitational instabilities, either spontaneously or externally triggered. some evolutionary sequences are described through which a galaxy could possibly concentrate mass and build bulges, how external gas accretion from cosmic filaments could be funneled to the galaxy disks, and intermittently driven to the galaxy center, to form nuclear starbursts and fuel an active nucleus. the frequency of both bars and lopsidedness can be used to constrain the gas accretion rate.
reference to recent papers and experimental feasibility are added. the paper will not be published in a hard - copy journal.
what violations of bell inequalities teach us is that the world is quantum mechanical, i. e., nonclassical. assertions that they imply the world is nonlocal arise from ignoring differences between quantum and classical physics.
the gravitational waves are non - physical sinuosities generated, in the last analysis, by undulating reference frames.
Question: Natural selection cannot occur without
A) competition for unlimited resources.
B) gradual warming of Earth.
C) genetic variation in species.
D) asexual reproduction in species.
|
C) genetic variation in species.
|
Context:
scientists look through telescopes, study images on electronic screens, record meter readings, and so on. generally, on a basic level, they can agree on what they see, e. g., the thermometer shows 37. 9 degrees c. but, if these scientists have different ideas about the theories that have been developed to explain these basic observations, they may disagree about what they are observing. for example, before albert einstein ' s general theory of relativity, observers would have likely interpreted an image of the einstein cross as five different objects in space. in light of that theory, however, astronomers will tell you that there are actually only two objects, one in the center and four different images of a second object around the sides. alternatively, if other scientists suspect that something is wrong with the telescope and only one object is actually being observed, they are operating under yet another theory. observations that cannot be separated from theoretical interpretation are said to be theory - laden. all observation involves both perception and cognition. that is, one does not make an observation passively, but rather is actively engaged in distinguishing the phenomenon being observed from surrounding sensory data. therefore, observations are affected by one ' s underlying understanding of the way in which the world functions, and that understanding may influence what is perceived, noticed, or deemed worthy of consideration. in this sense, it can be argued that all observation is theory - laden. = = = the purpose of science = = = should science aim to determine ultimate truth, or are there questions that science cannot answer? scientific realists claim that science aims at truth and that one ought to regard scientific theories as true, approximately true, or likely true. conversely, scientific anti - realists argue that science does not aim ( or at least does not succeed ) at truth, especially truth about unobservables like electrons or other universes. instrumentalists argue that scientific theories should only be evaluated on whether they are useful. in their view, whether theories are true or not is beside the point, because the purpose of science is to make predictions and enable effective technology. realists often point to the success of recent scientific theories as evidence for the truth ( or near truth ) of current theories. antirealists point to either the many false theories in the history of science, epistemic morals, the success of false modeling assumptions, or widely termed postmodern criticisms of objectivity as evidence against scientific realism. antirealists attempt to explain the success of scientific theories without reference to truth. some antirealists claim that scientific
education, science, in fact the whole society, extensively use images. between us and the world are the visual displays. screens, small and large, individual or not, are everywhere. images are increasingly the 2d substrate of our virtual interaction with reality. however images will never support a complete representation of the reality. three - dimensional representations will not change that. images are primarily a spatial representation of our world dedicated to our sight. key aspects such as energy and the associated forces are not spatially materialized. in classical physics, interaction description is based on newton equations with trajectory and force as the dual central concepts. images can in real time show all aspects of trajectories but not the associated dynamical aspects described by forces and energies. contrary to the real world, the world of images opposes no constrain, nor resistance to our actions. only the physical quantities, that do not contain mass in their dimension can be satisfactory represented by images. often symbols such as arrows are introduced to visualize the force vectors.
##physical processes which take place in human beings as they make sense of information received through the visual system. the subject of the image. when developing an imaging system, designers must consider the observables associated with the subjects which will be imaged. these observables generally take the form of emitted or reflected energy, such as electromagnetic energy or mechanical energy. the capture device. once the observables associated with the subject are characterized, designers can then identify and integrate the technologies needed to capture those observables. for example, in the case of consumer digital cameras, those technologies include optics for collecting energy in the visible portion of the electromagnetic spectrum, and electronic detectors for converting the electromagnetic energy into an electronic signal. the processor. for all digital imaging systems, the electronic signals produced by the capture device must be manipulated by an algorithm which formats the signals so they can be displayed as an image. in practice, there are often multiple processors involved in the creation of a digital image. the display. the display takes the electronic signals which have been manipulated by the processor and renders them on some visual medium. examples include paper ( for printed, or " hard copy " images ), television, computer monitor, or projector. note that some imaging scientists will include additional " links " in their description of the imaging chain. for example, some will include the " source " of the energy which " illuminates " or interacts with the subject of the image. others will include storage and / or transmission systems. = = subfields = = subfields within imaging science include : image processing, computer vision, 3d computer graphics, animations, atmospheric optics, astronomical imaging, biological imaging, digital image restoration, digital imaging, color science, digital photography, holography, magnetic resonance imaging, medical imaging, microdensitometry, optics, photography, remote sensing, radar imaging, radiometry, silver halide, ultrasound imaging, photoacoustic imaging, thermal imaging, visual perception, and various printing technologies. = = methodologies = = acoustic imaging coherent imaging uses an active coherent illumination source, such as in radar, synthetic aperture radar ( sar ), medical ultrasound and optical coherence tomography ; non - coherent imaging systems include fluorescent microscopes, optical microscopes, and telescopes. chemical imaging, the simultaneous measurement of spectra and pictures digital imaging, creating digital images, generally by scanning or through digital photography disk image, a file which contains the exact content of a data storage medium document imaging, replicating documents commonly
the nervous system. these kinds of tests can be divided into recordings of : ( 1 ) spontaneous or continuously running electrical activity, or ( 2 ) stimulus evoked responses. subspecialties include electroencephalography, electromyography, evoked potential, nerve conduction study and polysomnography. sometimes these tests are performed by techs without a medical degree, but the interpretation of these tests is done by a medical professional. diagnostic radiology is concerned with imaging of the body, e. g. by x - rays, x - ray computed tomography, ultrasonography, and nuclear magnetic resonance tomography. interventional radiologists can access areas in the body under imaging for an intervention or diagnostic sampling. nuclear medicine is concerned with studying human organ systems by administering radiolabelled substances ( radiopharmaceuticals ) to the body, which can then be imaged outside the body by a gamma camera or a pet scanner. each radiopharmaceutical consists of two parts : a tracer that is specific for the function under study ( e. g., neurotransmitter pathway, metabolic pathway, blood flow, or other ), and a radionuclide ( usually either a gamma - emitter or a positron emitter ). there is a degree of overlap between nuclear medicine and radiology, as evidenced by the emergence of combined devices such as the pet / ct scanner. pathology as a medical specialty is the branch of medicine that deals with the study of diseases and the morphologic, physiologic changes produced by them. as a diagnostic specialty, pathology can be considered the basis of modern scientific medical knowledge and plays a large role in evidence - based medicine. many modern molecular tests such as flow cytometry, polymerase chain reaction ( pcr ), immunohistochemistry, cytogenetics, gene rearrangements studies and fluorescent in situ hybridization ( fish ) fall within the territory of pathology. = = = = other major specialties = = = = the following are some major medical specialties that do not directly fit into any of the above - mentioned groups : anesthesiology ( also known as anaesthetics ) : concerned with the perioperative management of the surgical patient. the anesthesiologist ' s role during surgery is to prevent derangement in the vital organs ' ( i. e. brain, heart, kidneys ) functions and postoperative pain. outside of
of measuring methods. x - rays and gamma rays are used in industrial radiography to make images of the inside of solid products, as a means of nondestructive testing and inspection. the piece to be radiographed is placed between the source and a photographic film in a cassette. after a certain exposure time, the film is developed and it shows any internal defects of the material. gauges - gauges use the exponential absorption law of gamma rays level indicators : source and detector are placed at opposite sides of a container, indicating the presence or absence of material in the horizontal radiation path. beta or gamma sources are used, depending on the thickness and the density of the material to be measured. the method is used for containers of liquids or of grainy substances thickness gauges : if the material is of constant density, the signal measured by the radiation detector depends on the thickness of the material. this is useful for continuous production, like of paper, rubber, etc. electrostatic control - to avoid the build - up of static electricity in production of paper, plastics, synthetic textiles, etc., a ribbon - shaped source of the alpha emitter 241am can be placed close to the material at the end of the production line. the source ionizes the air to remove electric charges on the material. radioactive tracers - since radioactive isotopes behave, chemically, mostly like the inactive element, the behavior of a certain chemical substance can be followed by tracing the radioactivity. examples : adding a gamma tracer to a gas or liquid in a closed system makes it possible to find a hole in a tube. adding a tracer to the surface of the component of a motor makes it possible to measure wear by measuring the activity of the lubricating oil. oil and gas exploration - nuclear well logging is used to help predict the commercial viability of new or existing wells. the technology involves the use of a neutron or gamma - ray source and a radiation detector which are lowered into boreholes to determine the properties of the surrounding rock such as porosity and lithography. [ 1 ] road construction - nuclear moisture / density gauges are used to determine the density of soils, asphalt, and concrete. typically a cesium - 137 source is used. = = = commercial applications = = = radioluminescence tritium illumination : tritium is used with phosphor in rifle sights to increase nighttime firing accuracy. some runway markers and building exit signs use the same technology, to remain illuminated during blackouts. betavoltaics
the tests, assays, and procedures needed for providing the specific services. subspecialties include transfusion medicine, cellular pathology, clinical chemistry, hematology, clinical microbiology and clinical immunology. clinical neurophysiology is concerned with testing the physiology or function of the central and peripheral aspects of the nervous system. these kinds of tests can be divided into recordings of : ( 1 ) spontaneous or continuously running electrical activity, or ( 2 ) stimulus evoked responses. subspecialties include electroencephalography, electromyography, evoked potential, nerve conduction study and polysomnography. sometimes these tests are performed by techs without a medical degree, but the interpretation of these tests is done by a medical professional. diagnostic radiology is concerned with imaging of the body, e. g. by x - rays, x - ray computed tomography, ultrasonography, and nuclear magnetic resonance tomography. interventional radiologists can access areas in the body under imaging for an intervention or diagnostic sampling. nuclear medicine is concerned with studying human organ systems by administering radiolabelled substances ( radiopharmaceuticals ) to the body, which can then be imaged outside the body by a gamma camera or a pet scanner. each radiopharmaceutical consists of two parts : a tracer that is specific for the function under study ( e. g., neurotransmitter pathway, metabolic pathway, blood flow, or other ), and a radionuclide ( usually either a gamma - emitter or a positron emitter ). there is a degree of overlap between nuclear medicine and radiology, as evidenced by the emergence of combined devices such as the pet / ct scanner. pathology as a medical specialty is the branch of medicine that deals with the study of diseases and the morphologic, physiologic changes produced by them. as a diagnostic specialty, pathology can be considered the basis of modern scientific medical knowledge and plays a large role in evidence - based medicine. many modern molecular tests such as flow cytometry, polymerase chain reaction ( pcr ), immunohistochemistry, cytogenetics, gene rearrangements studies and fluorescent in situ hybridization ( fish ) fall within the territory of pathology. = = = = other major specialties = = = = the following are some major medical specialties that do not directly fit into any of the above - mentioned groups : anesthesiology ( also
astronomy uses methods from astronomy to determine past celestial constellations for forensic purposes. forensic botany is the study of plant life in order to gain information regarding possible crimes. forensic chemistry is the study of detection and identification of illicit drugs, accelerants used in arson cases, explosive and gunshot residue. forensic dactyloscopy is the study of fingerprints. forensic document examination or questioned document examination answers questions about a disputed document using a variety of scientific processes and methods. many examinations involve a comparison of the questioned document, or components of the document, with a set of known standards. the most common type of examination involves handwriting, whereby the examiner tries to address concerns about potential authorship. forensic dna analysis takes advantage of the uniqueness of an individual ' s dna to answer forensic questions such as paternity / maternity testing and placing a suspect at a crime scene, e. g. in a rape investigation. forensic engineering is the scientific examination and analysis of structures and products relating to their failure or cause of damage. forensic entomology deals with the examination of insects in, on and around human remains to assist in determination of time or location of death. it is also possible to determine if the body was moved after death using entomology. forensic geology deals with trace evidence in the form of soils, minerals and petroleum. forensic geomorphology is the study of the ground surface to look for potential location ( s ) of buried object ( s ). forensic geophysics is the application of geophysical techniques such as radar for detecting objects hidden underground or underwater. forensic intelligence process starts with the collection of data and ends with the integration of results within into the analysis of crimes under investigation. forensic interviews are conducted using the science of professionally using expertise to conduct a variety of investigative interviews with victims, witnesses, suspects or other sources to determine the facts regarding suspicions, allegations or specific incidents in either public or private sector settings. forensic histopathology is the application of histological techniques and examination to forensic pathology practice. forensic limnology is the analysis of evidence collected from crime scenes in or around fresh - water sources. examination of biological organisms, in particular diatoms, can be useful in connecting suspects with victims. forensic linguistics deals with issues in the legal system that requires linguistic expertise. forensic meteorology is a site - specific analysis of past weather conditions for a point of loss. forensic metrology is the application of metrology to assess the reliability of scientific evidence obtained through measurements forensic microbiology is the study of the necrobiome. forensic nursing
a comparison of the sensitivities of methods which allow us to determine the coordinates of a moving hot body is made.
, followed by a medical interview and a physical examination. basic diagnostic medical devices ( e. g., stethoscope, tongue depressor ) are typically used. after examining for signs and interviewing for symptoms, the doctor may order medical tests ( e. g., blood tests ), take a biopsy, or prescribe pharmaceutical drugs or other therapies. differential diagnosis methods help to rule out conditions based on the information provided. during the encounter, properly informing the patient of all relevant facts is an important part of the relationship and the development of trust. the medical encounter is then documented in the medical record, which is a legal document in many jurisdictions. follow - ups may be shorter but follow the same general procedure, and specialists follow a similar process. the diagnosis and treatment may take only a few minutes or a few weeks, depending on the complexity of the issue. the components of the medical interview and encounter are : chief complaint ( cc ) : the reason for the current medical visit. these are the symptoms. they are in the patient ' s own words and are recorded along with the duration of each one. also called chief concern or presenting complaint. current activity : occupation, hobbies, what the patient actually does. family history ( fh ) : listing of diseases in the family that may impact the patient. a family tree is sometimes used. history of present illness ( hpi ) : the chronological order of events of symptoms and further clarification of each symptom. distinguishable from history of previous illness, often called past medical history ( pmh ). medical history comprises hpi and pmh. medications ( rx ) : what drugs the patient takes including prescribed, over - the - counter, and home remedies, as well as alternative and herbal medicines or remedies. allergies are also recorded. past medical history ( pmh / pmhx ) : concurrent medical problems, past hospitalizations and operations, injuries, past infectious diseases or vaccinations, history of known allergies. review of systems ( ros ) or systems inquiry : a set of additional questions to ask, which may be missed on hpi : a general enquiry ( have you noticed any weight loss, change in sleep quality, fevers, lumps and bumps? etc. ), followed by questions on the body ' s main organ systems ( heart, lungs, digestive tract, urinary tract, etc. ). social history ( sh ) : birthplace, residences, marital history
. forensic histopathology is the application of histological techniques and examination to forensic pathology practice. forensic limnology is the analysis of evidence collected from crime scenes in or around fresh - water sources. examination of biological organisms, in particular diatoms, can be useful in connecting suspects with victims. forensic linguistics deals with issues in the legal system that requires linguistic expertise. forensic meteorology is a site - specific analysis of past weather conditions for a point of loss. forensic metrology is the application of metrology to assess the reliability of scientific evidence obtained through measurements forensic microbiology is the study of the necrobiome. forensic nursing is the application of nursing sciences to abusive crimes, like child abuse, or sexual abuse. categorization of wounds and traumas, collection of bodily fluids and emotional support are some of the duties of forensic nurses. forensic odontology is the study of the uniqueness of dentition, better known as the study of teeth. forensic optometry is the study of glasses and other eyewear relating to crime scenes and criminal investigations. forensic pathology is a field in which the principles of medicine and pathology are applied to determine a cause of death or injury in the context of a legal inquiry. forensic podiatry is an application of the study of feet footprint or footwear and their traces to analyze scene of crime and to establish personal identity in forensic examinations. forensic psychiatry is a specialized branch of psychiatry as applied to and based on scientific criminology. forensic psychology is the study of the mind of an individual, using forensic methods. usually it determines the circumstances behind a criminal ' s behavior. forensic seismology is the study of techniques to distinguish the seismic signals generated by underground nuclear explosions from those generated by earthquakes. forensic serology is the study of the body fluids. forensic social work is the specialist study of social work theories and their applications to a clinical, criminal justice or psychiatric setting. practitioners of forensic social work connected with the criminal justice system are often termed social supervisors, whilst the remaining use the interchangeable titles forensic social worker, approved mental health professional or forensic practitioner and they conduct specialist assessments of risk, care planning and act as an officer of the court. forensic toxicology is the study of the effect of drugs and poisons on / in the human body. forensic video analysis is the scientific examination, comparison and evaluation of video in legal matters. mobile device forensics is the scientific examination and evaluation of evidence found in mobile phones, e. g. call history and deleted sms, and includes sim card forensics
Question: In which of these investigations would pictures from a camera be most useful?
A) studying bird calls
B) studying exothermic reactions
C) studying the growth of plants
D) studying the diffusion of molecules
|
C) studying the growth of plants
|
Context:
time - dependent distribution of the global extinction of megafauna is compared with the growth of human population. there is no correlation between the two processes. furthermore, the size of human population and its growth rate were far too small to have any significant impact on the environment and on the life of megafauna.
they may have been present earlier, their diversification accelerated when they started using oxygen in their metabolism. later, around 1. 7 billion years ago, multicellular organisms began to appear, with differentiated cells performing specialised functions. algae - like multicellular land plants are dated back to about 1 billion years ago, although evidence suggests that microorganisms formed the earliest terrestrial ecosystems, at least 2. 7 billion years ago. microorganisms are thought to have paved the way for the inception of land plants in the ordovician period. land plants were so successful that they are thought to have contributed to the late devonian extinction event. ediacara biota appear during the ediacaran period, while vertebrates, along with most other modern phyla originated about 525 million years ago during the cambrian explosion. during the permian period, synapsids, including the ancestors of mammals, dominated the land, but most of this group became extinct in the permian β triassic extinction event 252 million years ago. during the recovery from this catastrophe, archosaurs became the most abundant land vertebrates ; one archosaur group, the dinosaurs, dominated the jurassic and cretaceous periods. after the cretaceous β paleogene extinction event 66 million years ago killed off the non - avian dinosaurs, mammals increased rapidly in size and diversity. such mass extinctions may have accelerated evolution by providing opportunities for new groups of organisms to diversify. = = diversity = = = = = bacteria and archaea = = = bacteria are a type of cell that constitute a large domain of prokaryotic microorganisms. typically a few micrometers in length, bacteria have a number of shapes, ranging from spheres to rods and spirals. bacteria were among the first life forms to appear on earth, and are present in most of its habitats. bacteria inhabit soil, water, acidic hot springs, radioactive waste, and the deep biosphere of the earth ' s crust. bacteria also live in symbiotic and parasitic relationships with plants and animals. most bacteria have not been characterised, and only about 27 percent of the bacterial phyla have species that can be grown in the laboratory. archaea constitute the other domain of prokaryotic cells and were initially classified as bacteria, receiving the name archaebacteria ( in the archaebacteria kingdom ), a term that has fallen out of use. archaeal cells have unique properties separating them from the other two domains, bacteria and eukaryota. archaea
it was the best of times ; it was the worst of times is the way dickens begins the tale of two cities. the line is appropriate to our time in particle physics. it is the best of times because we are in the midst of a revolution in understanding, the third to occur during my career. it is the worst of times because accelerator facilities are shutting down before new ones are opening, restricting the opportunity for experiments, and because of great uncertainty about future funding. my task today is to give you a view of the most important opportunities for our field under a scenario that is constrained by a tight budget. it is a time when we cannot afford the merely good, but must give first priority to the really important.
one of the greatest discoveries of modern times is that of the expanding universe, almost invariably attributed to hubble ( 1929 ). what is not widely known is that the original treatise by lemaitre ( 1927 ) contained a rich fusion of both theory and of observation. stiglers law of eponymy is yet again affirmed : no scientific discovery is named after its original discoverer ( merton, 1957 ). an appeal is made for a lemaitre telescope, to honour the discoverer of the expanding universe.
have evolved from the earliest emergence of life to present day. earth formed about 4. 5 billion years ago and all life on earth, both living and extinct, descended from a last universal common ancestor that lived about 3. 5 billion years ago. geologists have developed a geologic time scale that divides the history of the earth into major divisions, starting with four eons ( hadean, archean, proterozoic, and phanerozoic ), the first three of which are collectively known as the precambrian, which lasted approximately 4 billion years. each eon can be divided into eras, with the phanerozoic eon that began 539 million years ago being subdivided into paleozoic, mesozoic, and cenozoic eras. these three eras together comprise eleven periods ( cambrian, ordovician, silurian, devonian, carboniferous, permian, triassic, jurassic, cretaceous, tertiary, and quaternary ). the similarities among all known present - day species indicate that they have diverged through the process of evolution from their common ancestor. biologists regard the ubiquity of the genetic code as evidence of universal common descent for all bacteria, archaea, and eukaryotes. microbial mats of coexisting bacteria and archaea were the dominant form of life in the early archean eon and many of the major steps in early evolution are thought to have taken place in this environment. the earliest evidence of eukaryotes dates from 1. 85 billion years ago, and while they may have been present earlier, their diversification accelerated when they started using oxygen in their metabolism. later, around 1. 7 billion years ago, multicellular organisms began to appear, with differentiated cells performing specialised functions. algae - like multicellular land plants are dated back to about 1 billion years ago, although evidence suggests that microorganisms formed the earliest terrestrial ecosystems, at least 2. 7 billion years ago. microorganisms are thought to have paved the way for the inception of land plants in the ordovician period. land plants were so successful that they are thought to have contributed to the late devonian extinction event. ediacara biota appear during the ediacaran period, while vertebrates, along with most other modern phyla originated about 525 million years ago during the cambrian explosion. during the permian period, synapsids, including the ancestors of mammals, dominated the land, but most of this group became
. microbial mats of coexisting bacteria and archaea were the dominant form of life in the early archean eon and many of the major steps in early evolution are thought to have taken place in this environment. the earliest evidence of eukaryotes dates from 1. 85 billion years ago, and while they may have been present earlier, their diversification accelerated when they started using oxygen in their metabolism. later, around 1. 7 billion years ago, multicellular organisms began to appear, with differentiated cells performing specialised functions. algae - like multicellular land plants are dated back to about 1 billion years ago, although evidence suggests that microorganisms formed the earliest terrestrial ecosystems, at least 2. 7 billion years ago. microorganisms are thought to have paved the way for the inception of land plants in the ordovician period. land plants were so successful that they are thought to have contributed to the late devonian extinction event. ediacara biota appear during the ediacaran period, while vertebrates, along with most other modern phyla originated about 525 million years ago during the cambrian explosion. during the permian period, synapsids, including the ancestors of mammals, dominated the land, but most of this group became extinct in the permian β triassic extinction event 252 million years ago. during the recovery from this catastrophe, archosaurs became the most abundant land vertebrates ; one archosaur group, the dinosaurs, dominated the jurassic and cretaceous periods. after the cretaceous β paleogene extinction event 66 million years ago killed off the non - avian dinosaurs, mammals increased rapidly in size and diversity. such mass extinctions may have accelerated evolution by providing opportunities for new groups of organisms to diversify. = = diversity = = = = = bacteria and archaea = = = bacteria are a type of cell that constitute a large domain of prokaryotic microorganisms. typically a few micrometers in length, bacteria have a number of shapes, ranging from spheres to rods and spirals. bacteria were among the first life forms to appear on earth, and are present in most of its habitats. bacteria inhabit soil, water, acidic hot springs, radioactive waste, and the deep biosphere of the earth ' s crust. bacteria also live in symbiotic and parasitic relationships with plants and animals. most bacteria have not been characterised, and only about 27 percent of the bacterial phyla have species that can be grown in the laboratory. archaea constitute the other domain of
the celebrated franck - hertz experiment is reinterpreted by analogy with the glimmentladung experiment, formerly performed by heinrich hertz.
##rozoic eon that began 539 million years ago being subdivided into paleozoic, mesozoic, and cenozoic eras. these three eras together comprise eleven periods ( cambrian, ordovician, silurian, devonian, carboniferous, permian, triassic, jurassic, cretaceous, tertiary, and quaternary ). the similarities among all known present - day species indicate that they have diverged through the process of evolution from their common ancestor. biologists regard the ubiquity of the genetic code as evidence of universal common descent for all bacteria, archaea, and eukaryotes. microbial mats of coexisting bacteria and archaea were the dominant form of life in the early archean eon and many of the major steps in early evolution are thought to have taken place in this environment. the earliest evidence of eukaryotes dates from 1. 85 billion years ago, and while they may have been present earlier, their diversification accelerated when they started using oxygen in their metabolism. later, around 1. 7 billion years ago, multicellular organisms began to appear, with differentiated cells performing specialised functions. algae - like multicellular land plants are dated back to about 1 billion years ago, although evidence suggests that microorganisms formed the earliest terrestrial ecosystems, at least 2. 7 billion years ago. microorganisms are thought to have paved the way for the inception of land plants in the ordovician period. land plants were so successful that they are thought to have contributed to the late devonian extinction event. ediacara biota appear during the ediacaran period, while vertebrates, along with most other modern phyla originated about 525 million years ago during the cambrian explosion. during the permian period, synapsids, including the ancestors of mammals, dominated the land, but most of this group became extinct in the permian β triassic extinction event 252 million years ago. during the recovery from this catastrophe, archosaurs became the most abundant land vertebrates ; one archosaur group, the dinosaurs, dominated the jurassic and cretaceous periods. after the cretaceous β paleogene extinction event 66 million years ago killed off the non - avian dinosaurs, mammals increased rapidly in size and diversity. such mass extinctions may have accelerated evolution by providing opportunities for new groups of organisms to diversify. = = diversity = = = = = bacteria and archaea = = = bacteria are a type of cell that constitute a large domain of prokar
into major divisions, starting with four eons ( hadean, archean, proterozoic, and phanerozoic ), the first three of which are collectively known as the precambrian, which lasted approximately 4 billion years. each eon can be divided into eras, with the phanerozoic eon that began 539 million years ago being subdivided into paleozoic, mesozoic, and cenozoic eras. these three eras together comprise eleven periods ( cambrian, ordovician, silurian, devonian, carboniferous, permian, triassic, jurassic, cretaceous, tertiary, and quaternary ). the similarities among all known present - day species indicate that they have diverged through the process of evolution from their common ancestor. biologists regard the ubiquity of the genetic code as evidence of universal common descent for all bacteria, archaea, and eukaryotes. microbial mats of coexisting bacteria and archaea were the dominant form of life in the early archean eon and many of the major steps in early evolution are thought to have taken place in this environment. the earliest evidence of eukaryotes dates from 1. 85 billion years ago, and while they may have been present earlier, their diversification accelerated when they started using oxygen in their metabolism. later, around 1. 7 billion years ago, multicellular organisms began to appear, with differentiated cells performing specialised functions. algae - like multicellular land plants are dated back to about 1 billion years ago, although evidence suggests that microorganisms formed the earliest terrestrial ecosystems, at least 2. 7 billion years ago. microorganisms are thought to have paved the way for the inception of land plants in the ordovician period. land plants were so successful that they are thought to have contributed to the late devonian extinction event. ediacara biota appear during the ediacaran period, while vertebrates, along with most other modern phyla originated about 525 million years ago during the cambrian explosion. during the permian period, synapsids, including the ancestors of mammals, dominated the land, but most of this group became extinct in the permian β triassic extinction event 252 million years ago. during the recovery from this catastrophe, archosaurs became the most abundant land vertebrates ; one archosaur group, the dinosaurs, dominated the jurassic and cretaceous periods. after the cretaceous β paleogene extinction event 66 million years ago killed off
, tertiary, and quaternary ). the similarities among all known present - day species indicate that they have diverged through the process of evolution from their common ancestor. biologists regard the ubiquity of the genetic code as evidence of universal common descent for all bacteria, archaea, and eukaryotes. microbial mats of coexisting bacteria and archaea were the dominant form of life in the early archean eon and many of the major steps in early evolution are thought to have taken place in this environment. the earliest evidence of eukaryotes dates from 1. 85 billion years ago, and while they may have been present earlier, their diversification accelerated when they started using oxygen in their metabolism. later, around 1. 7 billion years ago, multicellular organisms began to appear, with differentiated cells performing specialised functions. algae - like multicellular land plants are dated back to about 1 billion years ago, although evidence suggests that microorganisms formed the earliest terrestrial ecosystems, at least 2. 7 billion years ago. microorganisms are thought to have paved the way for the inception of land plants in the ordovician period. land plants were so successful that they are thought to have contributed to the late devonian extinction event. ediacara biota appear during the ediacaran period, while vertebrates, along with most other modern phyla originated about 525 million years ago during the cambrian explosion. during the permian period, synapsids, including the ancestors of mammals, dominated the land, but most of this group became extinct in the permian β triassic extinction event 252 million years ago. during the recovery from this catastrophe, archosaurs became the most abundant land vertebrates ; one archosaur group, the dinosaurs, dominated the jurassic and cretaceous periods. after the cretaceous β paleogene extinction event 66 million years ago killed off the non - avian dinosaurs, mammals increased rapidly in size and diversity. such mass extinctions may have accelerated evolution by providing opportunities for new groups of organisms to diversify. = = diversity = = = = = bacteria and archaea = = = bacteria are a type of cell that constitute a large domain of prokaryotic microorganisms. typically a few micrometers in length, bacteria have a number of shapes, ranging from spheres to rods and spirals. bacteria were among the first life forms to appear on earth, and are present in most of its habitats. bacteria inhabit soil, water, acidic hot springs, radioactive
Question: Of the examples listed, which event is thought by scientists to have led to the greatest extinction?
A) a volcano in Hawaii erupting
B) the construction of the Hoover Dam
C) whalers of the 17th, 18th, and 19th centuries
D) the Mississippi River flooding its banks each summer
|
C) whalers of the 17th, 18th, and 19th centuries
|
Context:
the creation of your own reality and your own world. the metaphor i used was humans being like magic markers. for so long, they painted black and white pictures in their life because that ' s all they thought they could do. but they can paint with a different color and make a very vibrant and beautiful picture if they take control. on the single " new skin ", he further elaborated : in " new skin ", i attribute a scab to the present state of society. the way the scab looks in its worst state is gross and chaotic and horrible, that ' s now, but when it breaks away, there ' s a brand new piece of skin that ' s stronger than before. it ' s like creation out of chaos. the song " favorite things ", according to boyd, related to the topic of religion : " my favorite things " is my personal beliefs about religion and how it oppresses the things i enjoy the most. unfortunately, the simplest things, such as thinking for myself, creating my own reality and being whatever the hell i want to be each day of my life, are a sin. to be a good christian basically means to give up the reigns of your life and let some unseen force do it for you. " favorite things " also includes a sample of the 1959 track " flamenco fantasy ", by easy listening group the 101 strings orchestra. the song has a similar title to " my favorite things ", from the mary poppins musical and film, with both songs repeatedly mentioning their titles in the lyrics. however, it does not musically reference " my favorite things ". the single " a certain shade of green " has been described as being a song about procrastination. the line " are you gonna stand around till 2012 a. d.? " is a reference to an interpretation of the mayan calendar which dictated that the world would end on december 21, 2012. boyd did not believe this to be true, but it was on his mind as his mother was researching it for a book called maya memory : the glory that was palenque. while recording " nebula ", boyd said in 1997, " we found out what it ' s like to actually plug a phaser pedal into the wall while it ' s on. it sounds like a laser gun, and that ' s the first sound you hear in ' nebula '. " he added that for the song, " we used these walkie - talkies for children that have this slinky - like coil between them. when
of cells = = = autologous : the donor and the recipient of the cells are the same individual. cells are harvested, cultured or stored, and then reintroduced to the host. as a result of the host ' s own cells being reintroduced, an antigenic response is not elicited. the body ' s immune system recognizes these re - implanted cells as its own, and does not target them for attack. autologous cell dependence on host cell health and donor site morbidity may be deterrents to their use. adipose - derived and bone marrow - derived mesenchymal stem cells are commonly autologous in nature, and can be used in a myriad of ways, from helping repair skeletal tissue to replenishing beta cells in diabetic patients. allogenic : cells are obtained from the body of a donor of the same species as the recipient. while there are some ethical constraints to the use of human cells for in vitro studies ( i. e. human brain tissue chimera development ), the employment of dermal fibroblasts from human foreskin demonstrates an immunologically safe and thus a viable choice for allogenic tissue engineering of the skin. xenogenic : these cells are derived isolated cells from alternate species from the recipient. a notable example of xenogeneic tissue utilization is cardiovascular implant construction via animal cells. chimeric human - animal farming raises ethical concerns around the potential for improved consciousness from implanting human organs in animals. syngeneic or isogenic : these cells describe those borne from identical genetic code. this imparts an immunologic benefit similar to autologous cell lines ( see above ). autologous cells can be considered syngenic, but the classification also extends to non - autologously derived cells such as those from an identical twin, from genetically identical ( cloned ) research models, or induced stem cells ( isc ) as related to the donor. = = = stem cells = = = stem cells are undifferentiated cells with the ability to divide in culture and give rise to different forms of specialized cells. stem cells are divided into " adult " and " embryonic " stem cells according to their source. while there is still a large ethical debate related to the use of embryonic stem cells, it is thought that another alternative source β induced pluripotent stem cells β may be useful for the repair of diseased or damaged tissues, or may be used to grow new organs. totipotent cells
, buses, trucks, etc. it includes branch study of mechanical, electronic, software and safety elements. some of the engineering attributes and disciplines that are of importance to the automotive engineer include : safety engineering : safety engineering is the assessment of various crash scenarios and their impact on the vehicle occupants. these are tested against very stringent governmental regulations. some of these requirements include : seat belt and air bag functionality testing, front and side - impact testing, and tests of rollover resistance. assessments are done with various methods and tools, including computer crash simulation ( typically finite element analysis ), crash - test dummy, and partial system sled and full vehicle crashes. fuel economy / emissions : fuel economy is the measured fuel efficiency of the vehicle in miles per gallon or kilometers per liter. emissions - testing covers the measurement of vehicle emissions, including hydrocarbons, nitrogen oxides ( nox ), carbon monoxide ( co ), carbon dioxide ( co2 ), and evaporative emissions. nvh engineering ( noise, vibration, and harshness ) : nvh involves customer feedback ( both tactile [ felt ] and audible [ heard ] ) concerning a vehicle. while sound can be interpreted as a rattle, squeal, or hot, a tactile response can be seat vibration or a buzz in the steering wheel. this feedback is generated by components either rubbing, vibrating, or rotating. nvh response can be classified in various ways : powertrain nvh, road noise, wind noise, component noise, and squeak and rattle. note, there are both good and bad nvh qualities. the nvh engineer works to either eliminate bad nvh or change the " bad nvh " to good ( i. e., exhaust tones ). vehicle electronics : automotive electronics is an increasingly important aspect of automotive engineering. modern vehicles employ dozens of electronic systems. these systems are responsible for operational controls such as the throttle, brake and steering controls ; as well as many comfort - and - convenience systems such as the hvac, infotainment, and lighting systems. it would not be possible for automobiles to meet modern safety and fuel - economy requirements without electronic controls. performance : performance is a measurable and testable value of a vehicle ' s ability to perform in various conditions. performance can be considered in a wide variety of tasks, but it generally considers how quickly a car can accelerate ( e. g. standing start 1 / 4 mile elapsed time, 0 β 60 mph, etc. )
? if the latter, an important question is how the internal experiences of others can be measured. self - reports of feelings and beliefs may not be reliable because, even in cases in which there is no apparent incentive for subjects to intentionally deceive in their answers, self - deception or selective memory may affect their responses. then even in the case of accurate self - reports, how can responses be compared across individuals? even if two individuals respond with the same answer on a likert scale, they may be experiencing very different things. other issues in philosophy of psychology are philosophical questions about the nature of mind, brain, and cognition, and are perhaps more commonly thought of as part of cognitive science, or philosophy of mind. for example, are humans rational creatures? is there any sense in which they have free will, and how does that relate to the experience of making choices? philosophy of psychology also closely monitors contemporary work conducted in cognitive neuroscience, psycholinguistics, and artificial intelligence, questioning what they can and cannot explain in psychology. philosophy of psychology is a relatively young field, because psychology only became a discipline of its own in the late 1800s. in particular, neurophilosophy has just recently become its own field with the works of paul churchland and patricia churchland. philosophy of mind, by contrast, has been a well - established discipline since before psychology was a field of study at all. it is concerned with questions about the very nature of mind, the qualities of experience, and particular issues like the debate between dualism and monism. = = = philosophy of social science = = = the philosophy of social science is the study of the logic and method of the social sciences, such as sociology and cultural anthropology. philosophers of social science are concerned with the differences and similarities between the social and the natural sciences, causal relationships between social phenomena, the possible existence of social laws, and the ontological significance of structure and agency. the french philosopher, auguste comte ( 1798 β 1857 ), established the epistemological perspective of positivism in the course in positivist philosophy, a series of texts published between 1830 and 1842. the first three volumes of the course dealt chiefly with the natural sciences already in existence ( geoscience, astronomy, physics, chemistry, biology ), whereas the latter two emphasised the inevitable coming of social science : " sociologie ". for comte, the natural sciences had to necessarily arrive first, before humanity could adequately channel its efforts into the most challenging and complex " queen science " of human society
ambient air ( see lockheed f - 117 nighthawk, rectangular nozzles on the lockheed martin f - 22 raptor, and serrated nozzle flaps on the lockheed martin f - 35 lightning ). often, cool air is deliberately injected into the exhaust flow to boost this process ( see ryan aqm - 91 firefly and northrop b - 2 spirit ). the stefan β boltzmann law shows how this results in less energy ( thermal radiation in infrared spectrum ) being released and thus reduces the heat signature. in some aircraft, the jet exhaust is vented above the wing surface to shield it from observers below, as in the lockheed f - 117 nighthawk, and the unstealthy fairchild republic a - 10 thunderbolt ii. to achieve infrared stealth, the exhaust gas is cooled to the temperatures where the brightest wavelengths it radiates are absorbed by atmospheric carbon dioxide and water vapor, greatly reducing the infrared visibility of the exhaust plume. another way to reduce the exhaust temperature is to circulate coolant fluids such as fuel inside the exhaust pipe, where the fuel tanks serve as heat sinks cooled by the flow of air along the wings. ground combat includes the use of both active and passive infrared sensors. thus, the united states marine corps ( usmc ) ground combat uniform requirements document specifies infrared reflective quality standards. = = reducing radio frequency ( rf ) emissions = = in addition to reducing infrared and acoustic emissions, a stealth vehicle must avoid radiating any other detectable energy, such as from onboard radars, communications systems, or rf leakage from electronics enclosures. the f - 117 uses passive infrared and low light level television sensor systems to aim its weapons and the f - 22 raptor has an advanced lpi radar which can illuminate enemy aircraft without triggering a radar warning receiver response. = = measuring = = the size of a target ' s image on radar is measured by the rcs, often represented by the symbol Ο and expressed in square meters. this does not equal geometric area. a perfectly conducting sphere of projected cross sectional area 1 m2 ( i. e. a diameter of 1. 13 m ) will have an rcs of 1 m2. note that for radar wavelengths much less than the diameter of the sphere, rcs is independent of frequency. conversely, a square flat plate of area 1 m2 will have an rcs of Ο = 4Ο a2 / Ξ»2 ( where a = area, Ξ» = wavelength ), or 13, 982 m2 at 10 ghz if the radar is perpendicular to the flat
and reduces its radar profile. the flying wing design most closely resembles a so - called infinite flat plate ( as vertical control surfaces dramatically increase rcs ), the perfect stealth shape, as it would have no angles to reflect back radar waves. in addition to altering the tail, stealth design must bury the engines within the wing or fuselage, or in some cases where stealth is applied to an extant aircraft, install baffles in the air intakes, so that the compressor blades are not visible to radar. a stealthy shape must be devoid of complex bumps or protrusions of any kind, meaning that weapons, fuel tanks, and other stores must not be carried externally. any stealthy vehicle becomes un - stealthy when a door or hatch opens. parallel alignment of edges or even surfaces is also often used in stealth designs. the technique involves using a small number of edge orientations in the shape of the structure. for example, on the f - 22a raptor, the leading edges of the wing and the tail planes are set at the same angle. other smaller structures, such as the air intake bypass doors and the air refueling aperture, also use the same angles. the effect of this is to return a narrow radar signal in a very specific direction away from the radar emitter rather than returning a diffuse signal detectable at many angles. the effect is sometimes called " glitter " after the very brief signal seen when the reflected beam passes across a detector. it can be difficult for the radar operator to distinguish between a glitter event and a digital glitch in the processing system. stealth airframes sometimes display distinctive serrations on some exposed edges, such as the engine ports. the yf - 23 has such serrations on the exhaust ports. this is another example in the parallel alignment of features, this time on the external airframe. the shaping requirements detracted greatly from the f - 117 ' s aerodynamic properties. it is inherently unstable, and cannot be flown without a fly - by - wire control system. similarly, coating the cockpit canopy with a thin film transparent conductor ( vapor - deposited gold or indium tin oxide ) helps to reduce the aircraft ' s radar profile, because radar waves would normally enter the cockpit, reflect off objects ( the inside of a cockpit has a complex shape, with a pilot helmet alone forming a sizeable return ), and possibly return to the radar, but the conductive coating creates a controlled shape that deflects the incoming radar waves away from the radar. the coating is thin enough that it has
or fuselage, or in some cases where stealth is applied to an extant aircraft, install baffles in the air intakes, so that the compressor blades are not visible to radar. a stealthy shape must be devoid of complex bumps or protrusions of any kind, meaning that weapons, fuel tanks, and other stores must not be carried externally. any stealthy vehicle becomes un - stealthy when a door or hatch opens. parallel alignment of edges or even surfaces is also often used in stealth designs. the technique involves using a small number of edge orientations in the shape of the structure. for example, on the f - 22a raptor, the leading edges of the wing and the tail planes are set at the same angle. other smaller structures, such as the air intake bypass doors and the air refueling aperture, also use the same angles. the effect of this is to return a narrow radar signal in a very specific direction away from the radar emitter rather than returning a diffuse signal detectable at many angles. the effect is sometimes called " glitter " after the very brief signal seen when the reflected beam passes across a detector. it can be difficult for the radar operator to distinguish between a glitter event and a digital glitch in the processing system. stealth airframes sometimes display distinctive serrations on some exposed edges, such as the engine ports. the yf - 23 has such serrations on the exhaust ports. this is another example in the parallel alignment of features, this time on the external airframe. the shaping requirements detracted greatly from the f - 117 ' s aerodynamic properties. it is inherently unstable, and cannot be flown without a fly - by - wire control system. similarly, coating the cockpit canopy with a thin film transparent conductor ( vapor - deposited gold or indium tin oxide ) helps to reduce the aircraft ' s radar profile, because radar waves would normally enter the cockpit, reflect off objects ( the inside of a cockpit has a complex shape, with a pilot helmet alone forming a sizeable return ), and possibly return to the radar, but the conductive coating creates a controlled shape that deflects the incoming radar waves away from the radar. the coating is thin enough that it has no adverse effect on pilot vision. = = = = ships = = = = ships have also adopted similar methods. though the earlier american arleigh burke - class destroyers incorporated some signature - reduction features. the norwegian skjold - class corvettes was the first coastal defence and the french la fayette - class frigates the
the injuries of the inundations they have been designed to prevent, as the escape of floods from the raised river must occur sooner or later. inadequate planning controls which have permitted development on floodplains have been blamed for the flooding of domestic properties. channelization was done under the auspices or overall direction of engineers employed by the local authority or the national government. one of the most heavily channelized areas in the united states is west tennessee, where every major stream with one exception ( the hatchie river ) has been partially or completely channelized. channelization of a stream may be undertaken for several reasons. one is to make a stream more suitable for navigation or for navigation by larger vessels with deep draughts. another is to restrict water to a certain area of a stream ' s natural bottom lands so that the bulk of such lands can be made available for agriculture. a third reason is flood control, with the idea of giving a stream a sufficiently large and deep channel so that flooding beyond those limits will be minimal or nonexistent, at least on a routine basis. one major reason is to reduce natural erosion ; as a natural waterway curves back and forth, it usually deposits sand and gravel on the inside of the corners where the water flows slowly, and cuts sand, gravel, subsoil, and precious topsoil from the outside corners where it flows rapidly due to a change in direction. unlike sand and gravel, the topsoil that is eroded does not get deposited on the inside of the next corner of the river. it simply washes away. = = loss of wetlands = = channelization has several predictable and negative effects. one of them is loss of wetlands. wetlands are an excellent habitat for multiple forms of wildlife, and additionally serve as a " filter " for much of the world ' s surface fresh water. another is the fact that channelized streams are almost invariably straightened. for example, the channelization of florida ' s kissimmee river has been cited as a cause contributing to the loss of wetlands. this straightening causes the streams to flow more rapidly, which can, in some instances, vastly increase soil erosion. it can also increase flooding downstream from the channelized area, as larger volumes of water traveling more rapidly than normal can reach choke points over a shorter period of time than they otherwise would, with a net effect of flood control in one area coming at the expense of aggravated flooding in another. in addition, studies have shown that stream channelization results in declines of river fish populations. : 3 - 1ff a
to sustain innovation and safeguard national security, the u. s. must strengthen domestic pathways to computing phds by engaging talented undergraduates early - before they are committed to industry - with research experiences, mentorship, and financial support for graduate studies.
and by processing power using accelerators. food irradiation is only a niche application compared to medical supplies, plastic materials, raw materials, gemstones, cables and wires, etc. = = accidents = = nuclear accidents, because of the powerful forces involved, are often very dangerous. historically, the first incidents involved fatal radiation exposure. marie curie died from aplastic anemia which resulted from her high levels of exposure. two scientists, an american and canadian respectively, harry daghlian and louis slotin, died after mishandling the same plutonium mass. unlike conventional weapons, the intense light, heat, and explosive force is not the only deadly component to a nuclear weapon. approximately half of the deaths from hiroshima and nagasaki died two to five years afterward from radiation exposure. civilian nuclear and radiological accidents primarily involve nuclear power plants. most common are nuclear leaks that expose workers to hazardous material. a nuclear meltdown refers to the more serious hazard of releasing nuclear material into the surrounding environment. the most significant meltdowns occurred at three mile island in pennsylvania and chernobyl in the soviet ukraine. the earthquake and tsunami on march 11, 2011 caused serious damage to three nuclear reactors and a spent fuel storage pond at the fukushima daiichi nuclear power plant in japan. military reactors that experienced similar accidents were windscale in the united kingdom and sl - 1 in the united states. military accidents usually involve the loss or unexpected detonation of nuclear weapons. the castle bravo test in 1954 produced a larger yield than expected, which contaminated nearby islands, a japanese fishing boat ( with one fatality ), and raised concerns about contaminated fish in japan. in the 1950s through 1970s, several nuclear bombs were lost from submarines and aircraft, some of which have never been recovered. the last twenty years have seen a marked decline in such accidents. = = examples of environmental benefits = = proponents of nuclear energy note that annually, nuclear - generated electricity reduces 470 million metric tons of carbon dioxide emissions that would otherwise come from fossil fuels. additionally, the amount of comparatively low waste that nuclear energy does create is safely disposed of by the large scale nuclear energy production facilities or it is repurposed / recycled for other energy uses. proponents of nuclear energy also bring to attention the opportunity cost of utilizing other forms of electricity. for example, the environmental protection agency estimates that coal kills 30, 000 people a year, as a result of its environmental impact, while 60 people died in the chernobyl disaster. a real world example of impact provided by proponents of nuclear energy is
Question: The human stomach is protected from self-digestion by a layer of
A) bile.
B) mucus.
C) muscle.
D) white blood cells.
|
B) mucus.
|
Context:
directly on the skin is currently available as a sole study source. the significance of epidermal electronics involves their mechanical properties, which resemble those of skin. the skin can be modeled as bilayer, composed of an epidermis having young ' s modulus ( e ) of 2 - 80 kpa and thickness of 0. 3 β 3 mm and a dermis having e of 140 - 600 kpa and thickness of 0. 05 - 1. 5 mm. together this bilayer responds plastically to tensile strains β₯ 30 %, below which the skin ' s surface stretches and wrinkles without deforming. properties of epidermal electronics mirror those of skin to allow them to perform in this same way. like skin, epidermal electronics are ultrathin ( h < 100 ΞΌm ), low - modulus ( e β70 kpa ), and lightweight ( < 10 mg / cm2 ), enabling them to conform to the skin without applying strain. conformal contact and proper adhesion enable the device to bend and stretch without delaminating, deforming or failing, thereby eliminating the challenges with conventional, bulky wearables, including measurement artifacts, hysteresis, and motion - induced irritation to the skin. with this inherent ability to take the shape of skin, epidermal electronics can accurately acquire data without altering the natural motion or behavior of skin. the thin, soft, flexible design of epidermal electronics resembles that of temporary tattoos laminated on the skin. essentially, these devices are " mechanically invisible " to the wearer. epidermal electronics devices may adhere to the skin via van der waals forces or elastomeric substrates. with only van der waals forces, an epidermal device has the same thermal mass per unit area ( 150 mj / cm2k ) as skin, when the skin ' s thickness is < 500 nm. along with van der waals forces, the low values of e and thickness are effective in maximizing adhesion because they prevent deformation - induced detachment due to tension or compression. introducing an elastomeric substrate can improve adhesion but will raise the thermal mass per unit area slightly. several materials have been studied to produce these skin - like properties, including photolithography patterned serpentine gold nanofilm and patterned doping of silicon nanomembranes. = = = foot - worn = = = smart shoes are an example of wearable technology that incorporate smart features into shoes. smart shoes often work with smartphone applications to support
dimension 3, typically r3. a surface that is contained in a projective space is called a projective surface ( see Β§ projective surface ). a surface that is not supposed to be included in another space is called an abstract surface. = = examples = = the graph of a continuous function of two variables, defined over a connected open subset of r2 is a topological surface. if the function is differentiable, the graph is a differentiable surface. a plane is both an algebraic surface and a differentiable surface. it is also a ruled surface and a surface of revolution. a circular cylinder ( that is, the locus of a line crossing a circle and parallel to a given direction ) is an algebraic surface and a differentiable surface. a circular cone ( locus of a line crossing a circle, and passing through a fixed point, the apex, which is outside the plane of the circle ) is an algebraic surface which is not a differentiable surface. if one removes the apex, the remainder of the cone is the union of two differentiable surfaces. the surface of a polyhedron is a topological surface, which is neither a differentiable surface nor an algebraic surface. a hyperbolic paraboloid ( the graph of the function z = xy ) is a differentiable surface and an algebraic surface. it is also a ruled surface, and, for this reason, is often used in architecture. a two - sheet hyperboloid is an algebraic surface and the union of two non - intersecting differentiable surfaces. = = parametric surface = = a parametric surface is the image of an open subset of the euclidean plane ( typically r 2 { \ displaystyle \ mathbb { r } ^ { 2 } } ) by a continuous function, in a topological space, generally a euclidean space of dimension at least three. usually the function is supposed to be continuously differentiable, and this will be always the case in this article. specifically, a parametric surface in r 3 { \ displaystyle \ mathbb { r } ^ { 3 } } is given by three functions of two variables u and v, called parameters x = f 1 ( u, v ), y = f 2 ( u, v ), z = f 3 ( u, v ). { \ displaystyle { \ begin { aligned } x & = f _ { 1 } ( u, v ), \ \ [ 4pt ] y & = f _ { 2 } ( u, v ), \ \ [ 4pt ] z & = f _ { 3 }
the skin without applying strain. conformal contact and proper adhesion enable the device to bend and stretch without delaminating, deforming or failing, thereby eliminating the challenges with conventional, bulky wearables, including measurement artifacts, hysteresis, and motion - induced irritation to the skin. with this inherent ability to take the shape of skin, epidermal electronics can accurately acquire data without altering the natural motion or behavior of skin. the thin, soft, flexible design of epidermal electronics resembles that of temporary tattoos laminated on the skin. essentially, these devices are " mechanically invisible " to the wearer. epidermal electronics devices may adhere to the skin via van der waals forces or elastomeric substrates. with only van der waals forces, an epidermal device has the same thermal mass per unit area ( 150 mj / cm2k ) as skin, when the skin ' s thickness is < 500 nm. along with van der waals forces, the low values of e and thickness are effective in maximizing adhesion because they prevent deformation - induced detachment due to tension or compression. introducing an elastomeric substrate can improve adhesion but will raise the thermal mass per unit area slightly. several materials have been studied to produce these skin - like properties, including photolithography patterned serpentine gold nanofilm and patterned doping of silicon nanomembranes. = = = foot - worn = = = smart shoes are an example of wearable technology that incorporate smart features into shoes. smart shoes often work with smartphone applications to support tasks cannot be done with standard footwear. the uses include vibrating of the smart phone to tell users when and where to turn to reach their destination via google maps or self - lacing. self - lacing sneaker technology, similar to the nike mag in back to the future part ii, is another use of the smart shoe. in 2019 german footwear company puma was recognized as one of the " 100 best inventions of 2019 " by time for its fi laceless shoe that uses micro - motors to adjust the fit from an iphone. nike also introduced a smart shoe in 2019 known as adapt bb. the shoe featured buttons on the side to loosen or tighten the fit with a custom motor and gear, which could also be controlled by a smartphone. = = modern technologies = = on april 16, 2013, google invited " glass explorers " who had pre - ordered its wearable glasses at the 2012 google i / o conference to pick up their devices.
manifold of dimension two ( see Β§ topological surface ). a differentiable surface is a surfaces that is a differentiable manifold ( see Β§ differentiable surface ). every differentiable surface is a topological surface, but the converse is false. a " surface " is often implicitly supposed to be contained in a euclidean space of dimension 3, typically r3. a surface that is contained in a projective space is called a projective surface ( see Β§ projective surface ). a surface that is not supposed to be included in another space is called an abstract surface. = = examples = = the graph of a continuous function of two variables, defined over a connected open subset of r2 is a topological surface. if the function is differentiable, the graph is a differentiable surface. a plane is both an algebraic surface and a differentiable surface. it is also a ruled surface and a surface of revolution. a circular cylinder ( that is, the locus of a line crossing a circle and parallel to a given direction ) is an algebraic surface and a differentiable surface. a circular cone ( locus of a line crossing a circle, and passing through a fixed point, the apex, which is outside the plane of the circle ) is an algebraic surface which is not a differentiable surface. if one removes the apex, the remainder of the cone is the union of two differentiable surfaces. the surface of a polyhedron is a topological surface, which is neither a differentiable surface nor an algebraic surface. a hyperbolic paraboloid ( the graph of the function z = xy ) is a differentiable surface and an algebraic surface. it is also a ruled surface, and, for this reason, is often used in architecture. a two - sheet hyperboloid is an algebraic surface and the union of two non - intersecting differentiable surfaces. = = parametric surface = = a parametric surface is the image of an open subset of the euclidean plane ( typically r 2 { \ displaystyle \ mathbb { r } ^ { 2 } } ) by a continuous function, in a topological space, generally a euclidean space of dimension at least three. usually the function is supposed to be continuously differentiable, and this will be always the case in this article. specifically, a parametric surface in r 3 { \ displaystyle \ mathbb { r } ^ { 3 } } is given by three functions of two variables u and v, called parameters x = f 1 ( u, v ), y = f 2 ( u, v ), z = f 3
, etc. electrostatic control - to avoid the build - up of static electricity in production of paper, plastics, synthetic textiles, etc., a ribbon - shaped source of the alpha emitter 241am can be placed close to the material at the end of the production line. the source ionizes the air to remove electric charges on the material. radioactive tracers - since radioactive isotopes behave, chemically, mostly like the inactive element, the behavior of a certain chemical substance can be followed by tracing the radioactivity. examples : adding a gamma tracer to a gas or liquid in a closed system makes it possible to find a hole in a tube. adding a tracer to the surface of the component of a motor makes it possible to measure wear by measuring the activity of the lubricating oil. oil and gas exploration - nuclear well logging is used to help predict the commercial viability of new or existing wells. the technology involves the use of a neutron or gamma - ray source and a radiation detector which are lowered into boreholes to determine the properties of the surrounding rock such as porosity and lithography. [ 1 ] road construction - nuclear moisture / density gauges are used to determine the density of soils, asphalt, and concrete. typically a cesium - 137 source is used. = = = commercial applications = = = radioluminescence tritium illumination : tritium is used with phosphor in rifle sights to increase nighttime firing accuracy. some runway markers and building exit signs use the same technology, to remain illuminated during blackouts. betavoltaics. smoke detector : an ionization smoke detector includes a tiny mass of radioactive americium - 241, which is a source of alpha radiation. two ionisation chambers are placed next to each other. both contain a small source of 241am that gives rise to a small constant current. one is closed and serves for comparison, the other is open to ambient air ; it has a gridded electrode. when smoke enters the open chamber, the current is disrupted as the smoke particles attach to the charged ions and restore them to a neutral electrical state. this reduces the current in the open chamber. when the current drops below a certain threshold, the alarm is triggered. = = = food processing and agriculture = = = in biology and agriculture, radiation is used to induce mutations to produce new or improved species, such as in atomic gardening. another use in insect control is the sterile insect technique, where male insects are sterilized by radiation and released, so they have
##idermal electronics mirror those of skin to allow them to perform in this same way. like skin, epidermal electronics are ultrathin ( h < 100 ΞΌm ), low - modulus ( e β70 kpa ), and lightweight ( < 10 mg / cm2 ), enabling them to conform to the skin without applying strain. conformal contact and proper adhesion enable the device to bend and stretch without delaminating, deforming or failing, thereby eliminating the challenges with conventional, bulky wearables, including measurement artifacts, hysteresis, and motion - induced irritation to the skin. with this inherent ability to take the shape of skin, epidermal electronics can accurately acquire data without altering the natural motion or behavior of skin. the thin, soft, flexible design of epidermal electronics resembles that of temporary tattoos laminated on the skin. essentially, these devices are " mechanically invisible " to the wearer. epidermal electronics devices may adhere to the skin via van der waals forces or elastomeric substrates. with only van der waals forces, an epidermal device has the same thermal mass per unit area ( 150 mj / cm2k ) as skin, when the skin ' s thickness is < 500 nm. along with van der waals forces, the low values of e and thickness are effective in maximizing adhesion because they prevent deformation - induced detachment due to tension or compression. introducing an elastomeric substrate can improve adhesion but will raise the thermal mass per unit area slightly. several materials have been studied to produce these skin - like properties, including photolithography patterned serpentine gold nanofilm and patterned doping of silicon nanomembranes. = = = foot - worn = = = smart shoes are an example of wearable technology that incorporate smart features into shoes. smart shoes often work with smartphone applications to support tasks cannot be done with standard footwear. the uses include vibrating of the smart phone to tell users when and where to turn to reach their destination via google maps or self - lacing. self - lacing sneaker technology, similar to the nike mag in back to the future part ii, is another use of the smart shoe. in 2019 german footwear company puma was recognized as one of the " 100 best inventions of 2019 " by time for its fi laceless shoe that uses micro - motors to adjust the fit from an iphone. nike also introduced a smart shoe in 2019 known as adapt bb. the shoe featured buttons on the
discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under ordinary conditions, most of the materials brought down from the high lands by torrential water courses are carried forward by the main river to the sea, or partially strewn over flat alluvial plains during floods ; the size of the materials forming the bed of the river or borne along by the stream is gradually reduced on proceeding seawards, so that in the po river in italy, for instance, pebbles and gravel are found for about 140 miles below turin, sand along the next 100 miles, and silt and mud in the last 110 miles ( 176 km ). = = channelization = = the removal of obstructions, natural or artificial ( e. g., trunks of trees, boulders and accumulations of gravel ) from a river bed furnishes a simple and efficient means of increasing the discharging capacity of its channel. such removals will consequently lower the height of floods upstream. every impediment to the flow, in proportion to its extent, raises the level of the river above it so as to produce the additional artificial fall necessary to convey the flow through the restricted channel, thereby reducing the total available fall. reducing the length of the channel by substituting straight cuts for a winding course is the only way in which the effective fall can be increased. this involves some loss of capacity in the channel as a whole, and in the case of a large river with a considerable flow it is difficult to maintain a straight cut owing to the tendency of the current to erode the banks and form again a sinuous channel. even if the cut is preserved by protecting the banks,
layer of the skin. these wearables are mounted directly onto the skin to continuously monitor physiological and metabolic processes, both dermal and subdermal. wireless capability is typically achieved through battery, bluetooth or nfc, making these devices convenient and portable as a type of wearable technology. currently, epidermal electronics are being developed in the fields of fitness and medical monitoring. current usage of epidermal technology is limited by existing fabrication processes. its current application relies on various sophisticated fabrication techniques such as by lithography or by directly printing on a carrier substrate before attaching directly to the body. research into printing epidermal electronics directly on the skin is currently available as a sole study source. the significance of epidermal electronics involves their mechanical properties, which resemble those of skin. the skin can be modeled as bilayer, composed of an epidermis having young ' s modulus ( e ) of 2 - 80 kpa and thickness of 0. 3 β 3 mm and a dermis having e of 140 - 600 kpa and thickness of 0. 05 - 1. 5 mm. together this bilayer responds plastically to tensile strains β₯ 30 %, below which the skin ' s surface stretches and wrinkles without deforming. properties of epidermal electronics mirror those of skin to allow them to perform in this same way. like skin, epidermal electronics are ultrathin ( h < 100 ΞΌm ), low - modulus ( e β70 kpa ), and lightweight ( < 10 mg / cm2 ), enabling them to conform to the skin without applying strain. conformal contact and proper adhesion enable the device to bend and stretch without delaminating, deforming or failing, thereby eliminating the challenges with conventional, bulky wearables, including measurement artifacts, hysteresis, and motion - induced irritation to the skin. with this inherent ability to take the shape of skin, epidermal electronics can accurately acquire data without altering the natural motion or behavior of skin. the thin, soft, flexible design of epidermal electronics resembles that of temporary tattoos laminated on the skin. essentially, these devices are " mechanically invisible " to the wearer. epidermal electronics devices may adhere to the skin via van der waals forces or elastomeric substrates. with only van der waals forces, an epidermal device has the same thermal mass per unit area ( 150 mj / cm2k ) as skin, when the skin ' s thickness is < 500
floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under ordinary conditions, most of the materials brought down from the high lands by torrential water courses are carried forward by the main river to the sea, or partially strewn over flat alluvial plains during floods ; the size of the materials forming the bed of the river or borne along by the stream is gradually reduced on proceeding seawards, so that in the po river in italy, for instance, pebbles and gravel are found for about 140 miles below turin, sand along the next 100 miles, and silt and mud in the last 110 miles ( 176 km ). = = channelization = = the removal of obstructions, natural or artificial ( e. g., trunks of trees, boulders and accumulations of gravel ) from a river bed furnishes a simple and efficient means of increasing the discharging capacity of its channel. such removals will consequently lower the height of floods upstream. every impediment to the flow, in proportion to its extent, raises the level of the river above it so as to produce the additional artificial fall necessary to convey the flow through the restricted channel, thereby reducing the total available fall. reducing the length of the channel by substituting straight cuts for a winding course is the only way in which the effective fall can be increased.
designates the relationship between two or more variables. conceptual definition : description of a concept by relating it to other concepts. operational definition : details in regards to defining the variables and how they will be measured / assessed in the study. gathering of data : consists of identifying a population and selecting samples, gathering information from or about these samples by using specific research instruments. the instruments used for data collection must be valid and reliable. analysis of data : involves breaking down the individual pieces of data to draw conclusions about it. data interpretation : this can be represented through tables, figures, and pictures, and then described in words. test, revising of hypothesis conclusion, reiteration if necessary a common misconception is that a hypothesis will be proven ( see, rather, null hypothesis ). generally, a hypothesis is used to make predictions that can be tested by observing the outcome of an experiment. if the outcome is inconsistent with the hypothesis, then the hypothesis is rejected ( see falsifiability ). however, if the outcome is consistent with the hypothesis, the experiment is said to support the hypothesis. this careful language is used because researchers recognize that alternative hypotheses may also be consistent with the observations. in this sense, a hypothesis can never be proven, but rather only supported by surviving rounds of scientific testing and, eventually, becoming widely thought of as true. a useful hypothesis allows prediction and within the accuracy of observation of the time, the prediction will be verified. as the accuracy of observation improves with time, the hypothesis may no longer provide an accurate prediction. in this case, a new hypothesis will arise to challenge the old, and to the extent that the new hypothesis makes more accurate predictions than the old, the new will supplant it. researchers can also use a null hypothesis, which states no relationship or difference between the independent or dependent variables. = = = research in the humanities = = = research in the humanities involves different methods such as for example hermeneutics and semiotics. humanities scholars usually do not search for the ultimate correct answer to a question, but instead, explore the issues and details that surround it. context is always important, and context can be social, historical, political, cultural, or ethnic. an example of research in the humanities is historical research, which is embodied in historical method. historians use primary sources and other evidence to systematically investigate a topic, and then to write histories in the form of accounts of the past. other studies aim to merely examine the occurrence of behaviours in societies and communities
Question: On a field trip, a student holds a frog and reports that it has slippery skin. This is an example of
A) a measurement
B) a prediction
C) an explanation
D) an observation
|
D) an observation
|
Context:
be the more significant to modern soil theory than fallou ' s. previously, soil had been considered a product of chemical transformations of rocks, a dead substrate from which plants derive nutritious elements. soil and bedrock were in fact equated. dokuchaev considers the soil as a natural body having its own genesis and its own history of development, a body with complex and multiform processes taking place within it. the soil is considered as different from bedrock. the latter becomes soil under the influence of a series of soil - formation factors ( climate, vegetation, country, relief and age ). according to him, soil should be called the " daily " or outward horizons of rocks regardless of the type ; they are changed naturally by the common effect of water, air and various kinds of living and dead organisms. a 1914 encyclopedic definition : " the different forms of earth on the surface of the rocks, formed by the breaking down or weathering of rocks ". serves to illustrate the historic view of soil which persisted from the 19th century. dokuchaev ' s late 19th century soil concept developed in the 20th century to one of soil as earthy material that has been altered by living processes. a corollary concept is that soil without a living component is simply a part of earth ' s outer layer. further refinement of the soil concept is occurring in view of an appreciation of energy transport and transformation within soil. the term is popularly applied to the material on the surface of the earth ' s moon and mars, a usage acceptable within a portion of the scientific community. accurate to this modern understanding of soil is nikiforoff ' s 1959 definition of soil as the " excited skin of the sub aerial part of the earth ' s crust ". = = areas of practice = = academically, soil scientists tend to be drawn to one of five areas of specialization : microbiology, pedology, edaphology, physics, or chemistry. yet the work specifics are very much dictated by the challenges facing our civilization ' s desire to sustain the land that supports it, and the distinctions between the sub - disciplines of soil science often blur in the process. soil science professionals commonly stay current in soil chemistry, soil physics, soil microbiology, pedology, and applied soil science in related disciplines. one exciting effort drawing in soil scientists in the u. s. as of 2004 is the soil quality initiative. central to the soil quality initiative is developing indices of soil health and then monitoring them in a way
on a large scale provided protection from insect pests or tolerance to herbicides. fungal and virus resistant crops have also been developed or are in development. this makes the insect and weed management of crops easier and can indirectly increase crop yield. gm crops that directly improve yield by accelerating growth or making the plant more hardy ( by improving salt, cold or drought tolerance ) are also under development. in 2016 salmon have been genetically modified with growth hormones to reach normal adult size much faster. gmos have been developed that modify the quality of produce by increasing the nutritional value or providing more industrially useful qualities or quantities. the amflora potato produces a more industrially useful blend of starches. soybeans and canola have been genetically modified to produce more healthy oils. the first commercialised gm food was a tomato that had delayed ripening, increasing its shelf life. plants and animals have been engineered to produce materials they do not normally make. pharming uses crops and animals as bioreactors to produce vaccines, drug intermediates, or the drugs themselves ; the useful product is purified from the harvest and then used in the standard pharmaceutical production process. cows and goats have been engineered to express drugs and other proteins in their milk, and in 2009 the fda approved a drug produced in goat milk. = = = other applications = = = genetic engineering has potential applications in conservation and natural area management. gene transfer through viral vectors has been proposed as a means of controlling invasive species as well as vaccinating threatened fauna from disease. transgenic trees have been suggested as a way to confer resistance to pathogens in wild populations. with the increasing risks of maladaptation in organisms as a result of climate change and other perturbations, facilitated adaptation through gene tweaking could be one solution to reducing extinction risks. applications of genetic engineering in conservation are thus far mostly theoretical and have yet to be put into practice. genetic engineering is also being used to create microbial art. some bacteria have been genetically engineered to create black and white photographs. novelty items such as lavender - colored carnations, blue roses, and glowing fish, have also been produced through genetic engineering. = = regulation = = the regulation of genetic engineering concerns the approaches taken by governments to assess and manage the risks associated with the development and release of gmos. the development of a regulatory framework began in 1975, at asilomar, california. the asilomar meeting recommended a set of voluntary guidelines regarding the use of recombinant technology. as the technology improved
to be separated conceptually from geology and crop production and treated as a whole. as a founding father of soil science, fallou has primacy in time. fallou was working on the origins of soil before dokuchaev was born ; however dokuchaev ' s work was more extensive and is considered to be the more significant to modern soil theory than fallou ' s. previously, soil had been considered a product of chemical transformations of rocks, a dead substrate from which plants derive nutritious elements. soil and bedrock were in fact equated. dokuchaev considers the soil as a natural body having its own genesis and its own history of development, a body with complex and multiform processes taking place within it. the soil is considered as different from bedrock. the latter becomes soil under the influence of a series of soil - formation factors ( climate, vegetation, country, relief and age ). according to him, soil should be called the " daily " or outward horizons of rocks regardless of the type ; they are changed naturally by the common effect of water, air and various kinds of living and dead organisms. a 1914 encyclopedic definition : " the different forms of earth on the surface of the rocks, formed by the breaking down or weathering of rocks ". serves to illustrate the historic view of soil which persisted from the 19th century. dokuchaev ' s late 19th century soil concept developed in the 20th century to one of soil as earthy material that has been altered by living processes. a corollary concept is that soil without a living component is simply a part of earth ' s outer layer. further refinement of the soil concept is occurring in view of an appreciation of energy transport and transformation within soil. the term is popularly applied to the material on the surface of the earth ' s moon and mars, a usage acceptable within a portion of the scientific community. accurate to this modern understanding of soil is nikiforoff ' s 1959 definition of soil as the " excited skin of the sub aerial part of the earth ' s crust ". = = areas of practice = = academically, soil scientists tend to be drawn to one of five areas of specialization : microbiology, pedology, edaphology, physics, or chemistry. yet the work specifics are very much dictated by the challenges facing our civilization ' s desire to sustain the land that supports it, and the distinctions between the sub - disciplines of soil science often blur in the process. soil science professionals commonly stay current
generation times. corn has been used to study mechanisms of photosynthesis and phloem loading of sugar in c4 plants. the single celled green alga chlamydomonas reinhardtii, while not an embryophyte itself, contains a green - pigmented chloroplast related to that of land plants, making it useful for study. a red alga cyanidioschyzon merolae has also been used to study some basic chloroplast functions. spinach, peas, soybeans and a moss physcomitrella patens are commonly used to study plant cell biology. agrobacterium tumefaciens, a soil rhizosphere bacterium, can attach to plant cells and infect them with a callus - inducing ti plasmid by horizontal gene transfer, causing a callus infection called crown gall disease. schell and van montagu ( 1977 ) hypothesised that the ti plasmid could be a natural vector for introducing the nif gene responsible for nitrogen fixation in the root nodules of legumes and other plant species. today, genetic modification of the ti plasmid is one of the main techniques for introduction of transgenes to plants and the creation of genetically modified crops. = = = epigenetics = = = epigenetics is the study of heritable changes in gene function that cannot be explained by changes in the underlying dna sequence but cause the organism ' s genes to behave ( or " express themselves " ) differently. one example of epigenetic change is the marking of the genes by dna methylation which determines whether they will be expressed or not. gene expression can also be controlled by repressor proteins that attach to silencer regions of the dna and prevent that region of the dna code from being expressed. epigenetic marks may be added or removed from the dna during programmed stages of development of the plant, and are responsible, for example, for the differences between anthers, petals and normal leaves, despite the fact that they all have the same underlying genetic code. epigenetic changes may be temporary or may remain through successive cell divisions for the remainder of the cell ' s life. some epigenetic changes have been shown to be heritable, while others are reset in the germ cells. epigenetic changes in eukaryotic biology serve to regulate the process of cellular differentiation. during morphogenesis, totipotent stem cells become the various
the best - suited crops ( e. g., those with the highest yields ) to produce enough food to support a growing population. as crops and fields became increasingly large and difficult to maintain, it was discovered that specific organisms and their by - products could effectively fertilize, restore nitrogen, and control pests. throughout the history of agriculture, farmers have inadvertently altered the genetics of their crops through introducing them to new environments and breeding them with other plants β one of the first forms of biotechnology. these processes also were included in early fermentation of beer. these processes were introduced in early mesopotamia, egypt, china and india, and still use the same basic biological methods. in brewing, malted grains ( containing enzymes ) convert starch from grains into sugar and then adding specific yeasts to produce beer. in this process, carbohydrates in the grains broke down into alcohols, such as ethanol. later, other cultures produced the process of lactic acid fermentation, which produced other preserved foods, such as soy sauce. fermentation was also used in this time period to produce leavened bread. although the process of fermentation was not fully understood until louis pasteur ' s work in 1857, it is still the first use of biotechnology to convert a food source into another form. before the time of charles darwin ' s work and life, animal and plant scientists had already used selective breeding. darwin added to that body of work with his scientific observations about the ability of science to change species. these accounts contributed to darwin ' s theory of natural selection. for thousands of years, humans have used selective breeding to improve the production of crops and livestock to use them for food. in selective breeding, organisms with desirable characteristics are mated to produce offspring with the same characteristics. for example, this technique was used with corn to produce the largest and sweetest crops. in the early twentieth century scientists gained a greater understanding of microbiology and explored ways of manufacturing specific products. in 1917, chaim weizmann first used a pure microbiological culture in an industrial process, that of manufacturing corn starch using clostridium acetobutylicum, to produce acetone, which the united kingdom desperately needed to manufacture explosives during world war i. biotechnology has also led to the development of antibiotics. in 1928, alexander fleming discovered the mold penicillium. his work led to the purification of the antibiotic formed by the mold by howard florey, ernst boris chain and norman heatley β to form
the broad definition of " utilizing a biotechnological system to make products ". indeed, the cultivation of plants may be viewed as the earliest biotechnological enterprise. agriculture has been theorized to have become the dominant way of producing food since the neolithic revolution. through early biotechnology, the earliest farmers selected and bred the best - suited crops ( e. g., those with the highest yields ) to produce enough food to support a growing population. as crops and fields became increasingly large and difficult to maintain, it was discovered that specific organisms and their by - products could effectively fertilize, restore nitrogen, and control pests. throughout the history of agriculture, farmers have inadvertently altered the genetics of their crops through introducing them to new environments and breeding them with other plants β one of the first forms of biotechnology. these processes also were included in early fermentation of beer. these processes were introduced in early mesopotamia, egypt, china and india, and still use the same basic biological methods. in brewing, malted grains ( containing enzymes ) convert starch from grains into sugar and then adding specific yeasts to produce beer. in this process, carbohydrates in the grains broke down into alcohols, such as ethanol. later, other cultures produced the process of lactic acid fermentation, which produced other preserved foods, such as soy sauce. fermentation was also used in this time period to produce leavened bread. although the process of fermentation was not fully understood until louis pasteur ' s work in 1857, it is still the first use of biotechnology to convert a food source into another form. before the time of charles darwin ' s work and life, animal and plant scientists had already used selective breeding. darwin added to that body of work with his scientific observations about the ability of science to change species. these accounts contributed to darwin ' s theory of natural selection. for thousands of years, humans have used selective breeding to improve the production of crops and livestock to use them for food. in selective breeding, organisms with desirable characteristics are mated to produce offspring with the same characteristics. for example, this technique was used with corn to produce the largest and sweetest crops. in the early twentieth century scientists gained a greater understanding of microbiology and explored ways of manufacturing specific products. in 1917, chaim weizmann first used a pure microbiological culture in an industrial process, that of manufacturing corn starch using clostridium acetobutylicum, to produce acetone, which the united
. throughout the history of agriculture, farmers have inadvertently altered the genetics of their crops through introducing them to new environments and breeding them with other plants β one of the first forms of biotechnology. these processes also were included in early fermentation of beer. these processes were introduced in early mesopotamia, egypt, china and india, and still use the same basic biological methods. in brewing, malted grains ( containing enzymes ) convert starch from grains into sugar and then adding specific yeasts to produce beer. in this process, carbohydrates in the grains broke down into alcohols, such as ethanol. later, other cultures produced the process of lactic acid fermentation, which produced other preserved foods, such as soy sauce. fermentation was also used in this time period to produce leavened bread. although the process of fermentation was not fully understood until louis pasteur ' s work in 1857, it is still the first use of biotechnology to convert a food source into another form. before the time of charles darwin ' s work and life, animal and plant scientists had already used selective breeding. darwin added to that body of work with his scientific observations about the ability of science to change species. these accounts contributed to darwin ' s theory of natural selection. for thousands of years, humans have used selective breeding to improve the production of crops and livestock to use them for food. in selective breeding, organisms with desirable characteristics are mated to produce offspring with the same characteristics. for example, this technique was used with corn to produce the largest and sweetest crops. in the early twentieth century scientists gained a greater understanding of microbiology and explored ways of manufacturing specific products. in 1917, chaim weizmann first used a pure microbiological culture in an industrial process, that of manufacturing corn starch using clostridium acetobutylicum, to produce acetone, which the united kingdom desperately needed to manufacture explosives during world war i. biotechnology has also led to the development of antibiotics. in 1928, alexander fleming discovered the mold penicillium. his work led to the purification of the antibiotic formed by the mold by howard florey, ernst boris chain and norman heatley β to form what we today know as penicillin. in 1940, penicillin became available for medicinal use to treat bacterial infections in humans. the field of modern biotechnology is generally thought of as having been born in 1971 when paul berg ' s ( stanford ) experiments in gene splicing had early success. herbert w. boyer
kilometers ( 4, 200, 000 to 395, 400, 000 acres ). 10 % of the world ' s crop lands were planted with gm crops in 2010. as of 2011, 11 different transgenic crops were grown commercially on 395 million acres ( 160 million hectares ) in 29 countries such as the us, brazil, argentina, india, canada, china, paraguay, pakistan, south africa, uruguay, bolivia, australia, philippines, myanmar, burkina faso, mexico and spain. genetically modified foods are foods produced from organisms that have had specific changes introduced into their dna with the methods of genetic engineering. these techniques have allowed for the introduction of new crop traits as well as a far greater control over a food ' s genetic structure than previously afforded by methods such as selective breeding and mutation breeding. commercial sale of genetically modified foods began in 1994, when calgene first marketed its flavr savr delayed ripening tomato. to date most genetic modification of foods have primarily focused on cash crops in high demand by farmers such as soybean, corn, canola, and cotton seed oil. these have been engineered for resistance to pathogens and herbicides and better nutrient profiles. gm livestock have also been experimentally developed ; in november 2013 none were available on the market, but in 2015 the fda approved the first gm salmon for commercial production and consumption. there is a scientific consensus that currently available food derived from gm crops poses no greater risk to human health than conventional food, but that each gm food needs to be tested on a case - by - case basis before introduction. nonetheless, members of the public are much less likely than scientists to perceive gm foods as safe. the legal and regulatory status of gm foods varies by country, with some nations banning or restricting them, and others permitting them with widely differing degrees of regulation. gm crops also provide a number of ecological benefits, if not used in excess. insect - resistant crops have proven to lower pesticide usage, therefore reducing the environmental impact of pesticides as a whole. however, opponents have objected to gm crops per se on several grounds, including environmental concerns, whether food produced from gm crops is safe, whether gm crops are needed to address the world ' s food needs, and economic concerns raised by the fact these organisms are subject to intellectual property law. biotechnology has several applications in the realm of food security. crops like golden rice are engineered to have higher nutritional content, and there is potential for food products with longer shelf lives. though not a form of agricultural biotechnology, vaccines can help prevent diseases found in
eat them. plants and other photosynthetic organisms are at the base of most food chains because they use the energy from the sun and nutrients from the soil and atmosphere, converting them into a form that can be used by animals. this is what ecologists call the first trophic level. the modern forms of the major staple foods, such as hemp, teff, maize, rice, wheat and other cereal grasses, pulses, bananas and plantains, as well as hemp, flax and cotton grown for their fibres, are the outcome of prehistoric selection over thousands of years from among wild ancestral plants with the most desirable characteristics. botanists study how plants produce food and how to increase yields, for example through plant breeding, making their work important to humanity ' s ability to feed the world and provide food security for future generations. botanists also study weeds, which are a considerable problem in agriculture, and the biology and control of plant pathogens in agriculture and natural ecosystems. ethnobotany is the study of the relationships between plants and people. when applied to the investigation of historical plant β people relationships ethnobotany may be referred to as archaeobotany or palaeoethnobotany. some of the earliest plant - people relationships arose between the indigenous people of canada in identifying edible plants from inedible plants. this relationship the indigenous people had with plants was recorded by ethnobotanists. = = plant biochemistry = = plant biochemistry is the study of the chemical processes used by plants. some of these processes are used in their primary metabolism like the photosynthetic calvin cycle and crassulacean acid metabolism. others make specialised materials like the cellulose and lignin used to build their bodies, and secondary products like resins and aroma compounds. plants and various other groups of photosynthetic eukaryotes collectively known as " algae " have unique organelles known as chloroplasts. chloroplasts are thought to be descended from cyanobacteria that formed endosymbiotic relationships with ancient plant and algal ancestors. chloroplasts and cyanobacteria contain the blue - green pigment chlorophyll a. chlorophyll a ( as well as its plant and green algal - specific cousin chlorophyll b ) absorbs light in the blue - violet and orange / red parts of the spectrum while reflecting and transmitting the green light that we see as the characteristic colour
in agrobiotechnology. examples in food crops include resistance to certain pests, diseases, stressful environmental conditions, resistance to chemical treatments ( e. g. resistance to a herbicide ), reduction of spoilage, or improving the nutrient profile of the crop. examples in non - food crops include production of pharmaceutical agents, biofuels, and other industrially useful goods, as well as for bioremediation. farmers have widely adopted gm technology. between 1996 and 2011, the total surface area of land cultivated with gm crops had increased by a factor of 94, from 17, 000 to 1, 600, 000 square kilometers ( 4, 200, 000 to 395, 400, 000 acres ). 10 % of the world ' s crop lands were planted with gm crops in 2010. as of 2011, 11 different transgenic crops were grown commercially on 395 million acres ( 160 million hectares ) in 29 countries such as the us, brazil, argentina, india, canada, china, paraguay, pakistan, south africa, uruguay, bolivia, australia, philippines, myanmar, burkina faso, mexico and spain. genetically modified foods are foods produced from organisms that have had specific changes introduced into their dna with the methods of genetic engineering. these techniques have allowed for the introduction of new crop traits as well as a far greater control over a food ' s genetic structure than previously afforded by methods such as selective breeding and mutation breeding. commercial sale of genetically modified foods began in 1994, when calgene first marketed its flavr savr delayed ripening tomato. to date most genetic modification of foods have primarily focused on cash crops in high demand by farmers such as soybean, corn, canola, and cotton seed oil. these have been engineered for resistance to pathogens and herbicides and better nutrient profiles. gm livestock have also been experimentally developed ; in november 2013 none were available on the market, but in 2015 the fda approved the first gm salmon for commercial production and consumption. there is a scientific consensus that currently available food derived from gm crops poses no greater risk to human health than conventional food, but that each gm food needs to be tested on a case - by - case basis before introduction. nonetheless, members of the public are much less likely than scientists to perceive gm foods as safe. the legal and regulatory status of gm foods varies by country, with some nations banning or restricting them, and others permitting them with widely differing degrees of regulation. gm crops also provide a number of ecological benefits, if not used in excess. insect - resistant
Question: What most likely happens to soil when crops are not rotated from year to year?
A) The soil nutrients are depleted.
B) The soil fertility increases.
C) The soil becomes more porous.
D) The soil erodes less rapidly.
|
A) The soil nutrients are depleted.
|
Context:
oxygen, carbon dioxide, and water to pass through while restricting the movement of larger molecules and charged particles such as ions. cell membranes also contain membrane proteins, including integral membrane proteins that go across the membrane serving as membrane transporters, and peripheral proteins that loosely attach to the outer side of the cell membrane, acting as enzymes shaping the cell. cell membranes are involved in various cellular processes such as cell adhesion, storing electrical energy, and cell signalling and serve as the attachment surface for several extracellular structures such as a cell wall, glycocalyx, and cytoskeleton. within the cytoplasm of a cell, there are many biomolecules such as proteins and nucleic acids. in addition to biomolecules, eukaryotic cells have specialized structures called organelles that have their own lipid bilayers or are spatially units. these organelles include the cell nucleus, which contains most of the cell ' s dna, or mitochondria, which generate adenosine triphosphate ( atp ) to power cellular processes. other organelles such as endoplasmic reticulum and golgi apparatus play a role in the synthesis and packaging of proteins, respectively. biomolecules such as proteins can be engulfed by lysosomes, another specialized organelle. plant cells have additional organelles that distinguish them from animal cells such as a cell wall that provides support for the plant cell, chloroplasts that harvest sunlight energy to produce sugar, and vacuoles that provide storage and structural support as well as being involved in reproduction and breakdown of plant seeds. eukaryotic cells also have cytoskeleton that is made up of microtubules, intermediate filaments, and microfilaments, all of which provide support for the cell and are involved in the movement of the cell and its organelles. in terms of their structural composition, the microtubules are made up of tubulin ( e. g., Ξ± - tubulin and Ξ² - tubulin ) whereas intermediate filaments are made up of fibrous proteins. microfilaments are made up of actin molecules that interact with other strands of proteins. = = = metabolism = = = all cells require energy to sustain cellular processes. metabolism is the set of chemical reactions in an organism. the three main purposes of metabolism are : the conversion of food to energy to run cellular processes ; the conversion of food / fuel to monomer building blocks ; and
shaping the cell. cell membranes are involved in various cellular processes such as cell adhesion, storing electrical energy, and cell signalling and serve as the attachment surface for several extracellular structures such as a cell wall, glycocalyx, and cytoskeleton. within the cytoplasm of a cell, there are many biomolecules such as proteins and nucleic acids. in addition to biomolecules, eukaryotic cells have specialized structures called organelles that have their own lipid bilayers or are spatially units. these organelles include the cell nucleus, which contains most of the cell ' s dna, or mitochondria, which generate adenosine triphosphate ( atp ) to power cellular processes. other organelles such as endoplasmic reticulum and golgi apparatus play a role in the synthesis and packaging of proteins, respectively. biomolecules such as proteins can be engulfed by lysosomes, another specialized organelle. plant cells have additional organelles that distinguish them from animal cells such as a cell wall that provides support for the plant cell, chloroplasts that harvest sunlight energy to produce sugar, and vacuoles that provide storage and structural support as well as being involved in reproduction and breakdown of plant seeds. eukaryotic cells also have cytoskeleton that is made up of microtubules, intermediate filaments, and microfilaments, all of which provide support for the cell and are involved in the movement of the cell and its organelles. in terms of their structural composition, the microtubules are made up of tubulin ( e. g., Ξ± - tubulin and Ξ² - tubulin ) whereas intermediate filaments are made up of fibrous proteins. microfilaments are made up of actin molecules that interact with other strands of proteins. = = = metabolism = = = all cells require energy to sustain cellular processes. metabolism is the set of chemical reactions in an organism. the three main purposes of metabolism are : the conversion of food to energy to run cellular processes ; the conversion of food / fuel to monomer building blocks ; and the elimination of metabolic wastes. these enzyme - catalyzed reactions allow organisms to grow and reproduce, maintain their structures, and respond to their environments. metabolic reactions may be categorized as catabolic β the breaking down of compounds ( for example, the breaking down of glucose to pyruvate by cellular respiration
within a cell membrane that separates its cytoplasm from the extracellular space. a cell membrane consists of a lipid bilayer, including cholesterols that sit between phospholipids to maintain their fluidity at various temperatures. cell membranes are semipermeable, allowing small molecules such as oxygen, carbon dioxide, and water to pass through while restricting the movement of larger molecules and charged particles such as ions. cell membranes also contain membrane proteins, including integral membrane proteins that go across the membrane serving as membrane transporters, and peripheral proteins that loosely attach to the outer side of the cell membrane, acting as enzymes shaping the cell. cell membranes are involved in various cellular processes such as cell adhesion, storing electrical energy, and cell signalling and serve as the attachment surface for several extracellular structures such as a cell wall, glycocalyx, and cytoskeleton. within the cytoplasm of a cell, there are many biomolecules such as proteins and nucleic acids. in addition to biomolecules, eukaryotic cells have specialized structures called organelles that have their own lipid bilayers or are spatially units. these organelles include the cell nucleus, which contains most of the cell ' s dna, or mitochondria, which generate adenosine triphosphate ( atp ) to power cellular processes. other organelles such as endoplasmic reticulum and golgi apparatus play a role in the synthesis and packaging of proteins, respectively. biomolecules such as proteins can be engulfed by lysosomes, another specialized organelle. plant cells have additional organelles that distinguish them from animal cells such as a cell wall that provides support for the plant cell, chloroplasts that harvest sunlight energy to produce sugar, and vacuoles that provide storage and structural support as well as being involved in reproduction and breakdown of plant seeds. eukaryotic cells also have cytoskeleton that is made up of microtubules, intermediate filaments, and microfilaments, all of which provide support for the cell and are involved in the movement of the cell and its organelles. in terms of their structural composition, the microtubules are made up of tubulin ( e. g., Ξ± - tubulin and Ξ² - tubulin ) whereas intermediate filaments are made up of fibrous proteins. microfilaments are made up of actin molecules that interact with
the surface of the membrane, retentate is removed from the same side further downstream, whereas the permeate flow is tracked on the other side. in dead - end filtration, the direction of the fluid flow is normal to the membrane surface. both flow geometries offer some advantages and disadvantages. generally, dead - end filtration is used for feasibility studies on a laboratory scale. the dead - end membranes are relatively easy to fabricate which reduces the cost of the separation process. the dead - end membrane separation process is easy to implement and the process is usually cheaper than cross - flow membrane filtration. the dead - end filtration process is usually a batch - type process, where the filtering solution is loaded ( or slowly fed ) into the membrane device, which then allows passage of some particles subject to the driving force. the main disadvantage of dead - end filtration is the extensive membrane fouling and concentration polarization. the fouling is usually induced faster at higher driving forces. membrane fouling and particle retention in a feed solution also builds up a concentration gradients and particle backflow ( concentration polarization ). the tangential flow devices are more cost and labor - intensive, but they are less susceptible to fouling due to the sweeping effects and high shear rates of the passing flow. the most commonly used synthetic membrane devices ( modules ) are flat sheets / plates, spiral wounds, and hollow fibers. flat membranes used in filtration and separation processes can be enhanced with surface patterning, where microscopic structures are introduced to improve performance. these patterns increase surface area, optimize water flow, and reduce fouling, leading to higher permeability and longer membrane lifespan. research has shown that such modifications can significantly enhance efficiency in water purification, energy applications, and industrial separations. flat plates are usually constructed as circular thin flat membrane surfaces to be used in dead - end geometry modules. spiral wounds are constructed from similar flat membranes but in the form of a " pocket " containing two membrane sheets separated by a highly porous support plate. several such pockets are then wound around a tube to create a tangential flow geometry and to reduce membrane fouling. hollow fiber modules consist of an assembly of self - supporting fibers with dense skin separation layers, and a more open matrix helping to withstand pressure gradients and maintain structural integrity. the hollow fiber modules can contain up to 10, 000 fibers ranging from 200 to 2500 ΞΌm in diameter ; the main advantage of hollow fiber modules is the very large surface area within
use less energy than conventional thermal separation processes such as distillation, sublimation or crystallization. the separation process is purely physical and both fractions ( permeate and retentate ) can be obtained as useful products. cold separation using membrane technology is widely used in the food technology, biotechnology and pharmaceutical industries. furthermore, using membranes enables separations to take place that would be impossible using thermal separation methods. for example, it is impossible to separate the constituents of azeotropic liquids or solutes which form isomorphic crystals by distillation or recrystallization but such separations can be achieved using membrane technology. depending on the type of membrane, the selective separation of certain individual substances or substance mixtures is possible. important technical applications include the production of drinking water by reverse osmosis. in waste water treatment, membrane technology is becoming increasingly important. ultra / microfiltration can be very effective in removing colloids and macromolecules from wastewater. this is needed if wastewater is discharged into sensitive waters especially those designated for contact water sports and recreation. about half of the market is in medical applications such as artificial kidneys to remove toxic substances by hemodialysis and as artificial lung for bubble - free supply of oxygen in the blood. the importance of membrane technology is growing in the field of environmental protection ( nano - mem - pro ippc database ). even in modern energy recovery techniques, membranes are increasingly used, for example in fuel cells and in osmotic power plants. = = mass transfer = = two basic models can be distinguished for mass transfer through the membrane : the solution - diffusion model and the hydrodynamic model. in real membranes, these two transport mechanisms certainly occur side by side, especially during ultra - filtration. = = = solution - diffusion model = = = in the solution - diffusion model, transport occurs only by diffusion. the component that needs to be transported must first be dissolved in the membrane. the general approach of the solution - diffusion model is to assume that the chemical potential of the feed and permeate fluids are in equilibrium with the adjacent membrane surfaces such that appropriate expressions for the chemical potential in the fluid and membrane phases can be equated at the solution - membrane interface. this principle is more important for dense membranes without natural pores such as those used for reverse osmosis and in fuel cells. during the filtration process a boundary layer forms on the membrane. this concentration gradient is created by molecules which cannot pass through the membrane. the
generally, dead - end filtration is used for feasibility studies on a laboratory scale. the dead - end membranes are relatively easy to fabricate which reduces the cost of the separation process. the dead - end membrane separation process is easy to implement and the process is usually cheaper than cross - flow membrane filtration. the dead - end filtration process is usually a batch - type process, where the filtering solution is loaded ( or slowly fed ) into the membrane device, which then allows passage of some particles subject to the driving force. the main disadvantage of dead - end filtration is the extensive membrane fouling and concentration polarization. the fouling is usually induced faster at higher driving forces. membrane fouling and particle retention in a feed solution also builds up a concentration gradients and particle backflow ( concentration polarization ). the tangential flow devices are more cost and labor - intensive, but they are less susceptible to fouling due to the sweeping effects and high shear rates of the passing flow. the most commonly used synthetic membrane devices ( modules ) are flat sheets / plates, spiral wounds, and hollow fibers. flat membranes used in filtration and separation processes can be enhanced with surface patterning, where microscopic structures are introduced to improve performance. these patterns increase surface area, optimize water flow, and reduce fouling, leading to higher permeability and longer membrane lifespan. research has shown that such modifications can significantly enhance efficiency in water purification, energy applications, and industrial separations. flat plates are usually constructed as circular thin flat membrane surfaces to be used in dead - end geometry modules. spiral wounds are constructed from similar flat membranes but in the form of a " pocket " containing two membrane sheets separated by a highly porous support plate. several such pockets are then wound around a tube to create a tangential flow geometry and to reduce membrane fouling. hollow fiber modules consist of an assembly of self - supporting fibers with dense skin separation layers, and a more open matrix helping to withstand pressure gradients and maintain structural integrity. the hollow fiber modules can contain up to 10, 000 fibers ranging from 200 to 2500 ΞΌm in diameter ; the main advantage of hollow fiber modules is the very large surface area within an enclosed volume, increasing the efficiency of the separation process. the disc tube module uses a cross - flow geometry and consists of a pressure tube and hydraulic discs, which are held by a central tension rod, and membrane cushions that lie between two discs. = = membrane performance and governing equations = = the selection of synthetic membranes
##able. additionally, they must be biocompatible, meaning that they do not cause any adverse effects to cells. silicone, for example, is a synthetic, non - biodegradable material commonly used as a drug delivery material, while gelatin is a biodegradable, natural material commonly used in cell - culture scaffolds the material needed for each application is different, and dependent on the desired mechanical properties of the material. tissue engineering of long bone defects for example, will require a rigid scaffold with a compressive strength similar to that of cortical bone ( 100 - 150 mpa ), which is much higher compared to a scaffold for skin regeneration. there are a few versatile synthetic materials used for many different scaffold applications. one of these commonly used materials is polylactic acid ( pla ), a synthetic polymer. pla β polylactic acid. this is a polyester which degrades within the human body to form lactic acid, a naturally occurring chemical which is easily removed from the body. similar materials are polyglycolic acid ( pga ) and polycaprolactone ( pcl ) : their degradation mechanism is similar to that of pla, but pcl degrades slower and pga degrades faster. pla is commonly combined with pga to create poly - lactic - co - glycolic acid ( plga ). this is especially useful because the degradation of plga can be tailored by altering the weight percentages of pla and pga : more pla β slower degradation, more pga β faster degradation. this tunability, along with its biocompatibility, makes it an extremely useful material for scaffold creation. scaffolds may also be constructed from natural materials : in particular different derivatives of the extracellular matrix have been studied to evaluate their ability to support cell growth. protein based materials β such as collagen, or fibrin, and polysaccharidic materials - like chitosan or glycosaminoglycans ( gags ), have all proved suitable in terms of cell compatibility. among gags, hyaluronic acid, possibly in combination with cross linking agents ( e. g. glutaraldehyde, water - soluble carbodiimide, etc. ), is one of the possible choices as scaffold material. due to the covalent attachment of thiol groups to these polymers, they can crosslink via disulfide bond
are single - celled organisms such as bacteria, whereas eukaryotes can be single - celled or multicellular. in multicellular organisms, every cell in the organism ' s body is derived ultimately from a single cell in a fertilized egg. = = = cell structure = = = every cell is enclosed within a cell membrane that separates its cytoplasm from the extracellular space. a cell membrane consists of a lipid bilayer, including cholesterols that sit between phospholipids to maintain their fluidity at various temperatures. cell membranes are semipermeable, allowing small molecules such as oxygen, carbon dioxide, and water to pass through while restricting the movement of larger molecules and charged particles such as ions. cell membranes also contain membrane proteins, including integral membrane proteins that go across the membrane serving as membrane transporters, and peripheral proteins that loosely attach to the outer side of the cell membrane, acting as enzymes shaping the cell. cell membranes are involved in various cellular processes such as cell adhesion, storing electrical energy, and cell signalling and serve as the attachment surface for several extracellular structures such as a cell wall, glycocalyx, and cytoskeleton. within the cytoplasm of a cell, there are many biomolecules such as proteins and nucleic acids. in addition to biomolecules, eukaryotic cells have specialized structures called organelles that have their own lipid bilayers or are spatially units. these organelles include the cell nucleus, which contains most of the cell ' s dna, or mitochondria, which generate adenosine triphosphate ( atp ) to power cellular processes. other organelles such as endoplasmic reticulum and golgi apparatus play a role in the synthesis and packaging of proteins, respectively. biomolecules such as proteins can be engulfed by lysosomes, another specialized organelle. plant cells have additional organelles that distinguish them from animal cells such as a cell wall that provides support for the plant cell, chloroplasts that harvest sunlight energy to produce sugar, and vacuoles that provide storage and structural support as well as being involved in reproduction and breakdown of plant seeds. eukaryotic cells also have cytoskeleton that is made up of microtubules, intermediate filaments, and microfilaments, all of which provide support for the cell and are involved in the movement of the cell and its
there are a variety of bioreactors designed for 3d cell cultures. there are small plastic cylindrical chambers, as well as glass chambers, with regulated internal humidity and moisture specifically engineered for the purpose of growing cells in three dimensions. the bioreactor uses bioactive synthetic materials such as polyethylene terephthalate membranes to surround the spheroid cells in an environment that maintains high levels of nutrients. they are easy to open and close, so that cell spheroids can be removed for testing, yet the chamber is able to maintain 100 % humidity throughout. this humidity is important to achieve maximum cell growth and function. the bioreactor chamber is part of a larger device that rotates to ensure equal cell growth in each direction across three dimensions. quinxell technologies now under quintech life sciences from singapore has developed a bioreactor known as the tisxell biaxial bioreactor which is specially designed for the purpose of tissue engineering. it is the first bioreactor in the world to have a spherical glass chamber with biaxial rotation ; specifically to mimic the rotation of the fetus in the womb ; which provides a conducive environment for the growth of tissues. multiple forms of mechanical stimulation have also been combined into a single bioreactor. using gene expression analysis, one academic study found that applying a combination of cyclic strain and ultrasound stimulation to pre - osteoblast cells in a bioreactor accelerated matrix maturation and differentiation. the technology of this combined stimulation bioreactor could be used to grow bone cells more quickly and effectively in future clinical stem cell therapies. mc2 biotek has also developed a bioreactor known as prototissue that uses gas exchange to maintain high oxygen levels within the cell chamber ; improving upon previous bioreactors, since the higher oxygen levels help the cell grow and undergo normal cell respiration. active areas of research on bioreactors includes increasing production scale and refining the physiological environment, both of which could improve the efficiency and efficacy of bioreactors in research or clinical use. bioreactors are currently used to study, among other things, cell and tissue level therapies, cell and tissue response to specific physiological environment changes, and development of disease and injury. = = = long fiber generation = = = in 2013, a group from the university of tokyo developed cell laden fibers up to a meter in length and on the order of 100 ΞΌm in size. these fibers were created using a microfluidic device that forms a
. most cells are very small, with diameters ranging from 1 to 100 micrometers and are therefore only visible under a light or electron microscope. there are generally two types of cells : eukaryotic cells, which contain a nucleus, and prokaryotic cells, which do not. prokaryotes are single - celled organisms such as bacteria, whereas eukaryotes can be single - celled or multicellular. in multicellular organisms, every cell in the organism ' s body is derived ultimately from a single cell in a fertilized egg. = = = cell structure = = = every cell is enclosed within a cell membrane that separates its cytoplasm from the extracellular space. a cell membrane consists of a lipid bilayer, including cholesterols that sit between phospholipids to maintain their fluidity at various temperatures. cell membranes are semipermeable, allowing small molecules such as oxygen, carbon dioxide, and water to pass through while restricting the movement of larger molecules and charged particles such as ions. cell membranes also contain membrane proteins, including integral membrane proteins that go across the membrane serving as membrane transporters, and peripheral proteins that loosely attach to the outer side of the cell membrane, acting as enzymes shaping the cell. cell membranes are involved in various cellular processes such as cell adhesion, storing electrical energy, and cell signalling and serve as the attachment surface for several extracellular structures such as a cell wall, glycocalyx, and cytoskeleton. within the cytoplasm of a cell, there are many biomolecules such as proteins and nucleic acids. in addition to biomolecules, eukaryotic cells have specialized structures called organelles that have their own lipid bilayers or are spatially units. these organelles include the cell nucleus, which contains most of the cell ' s dna, or mitochondria, which generate adenosine triphosphate ( atp ) to power cellular processes. other organelles such as endoplasmic reticulum and golgi apparatus play a role in the synthesis and packaging of proteins, respectively. biomolecules such as proteins can be engulfed by lysosomes, another specialized organelle. plant cells have additional organelles that distinguish them from animal cells such as a cell wall that provides support for the plant cell, chloroplasts that harvest sunlight energy to produce sugar, and vacuoles that provide storage and structural support
Question: Materials must pass through a cell's semi-permeable membrane in order for energy to be utilized. This process is called
A) diffusion.
B) oxidation.
C) respiration.
D) reproduction.
|
A) diffusion.
|
Context:
the decomposition theorem is deduced from local purity.
oil umbrella ) ; for calculating the time of death ( allowing for weather and insect activity ) ; described how to wash and examine the dead body to ascertain the reason for death. at that time the book had described methods for distinguishing between suicide and faked suicide. he wrote the book on forensics stating that all wounds or dead bodies should be examined, not avoided. the book became the first form of literature to help determine the cause of death. in one of song ci ' s accounts ( washing away of wrongs ), the case of a person murdered with a sickle was solved by an investigator who instructed each suspect to bring his sickle to one location. ( he realized it was a sickle by testing various blades on an animal carcass and comparing the wounds. ) flies, attracted by the smell of blood, eventually gathered on a single sickle. in light of this, the owner of that sickle confessed to the murder. the book also described how to distinguish between a drowning ( water in the lungs ) and strangulation ( broken neck cartilage ), and described evidence from examining corpses to determine if a death was caused by murder, suicide or accident. methods from around the world involved saliva and examination of the mouth and tongue to determine innocence or guilt, as a precursor to the polygraph test. in ancient india, some suspects were made to fill their mouths with dried rice and spit it back out. similarly, in ancient china, those accused of a crime would have rice powder placed in their mouths. in ancient middle - eastern cultures, the accused were made to lick hot metal rods briefly. it is thought that these tests had some validity since a guilty person would produce less saliva and thus have a drier mouth ; the accused would be considered guilty if rice was sticking to their mouths in abundance or if their tongues were severely burned due to lack of shielding from saliva. = = education and training = = initial glance, forensic intelligence may appear as a nascent facet of forensic science facilitated by advancements in information technologies such as computers, databases, and data - flow management software. however, a more profound examination reveals that forensic intelligence represents a genuine and emerging inclination among forensic practitioners to actively participate in investigative and policing strategies. in doing so, it elucidates existing practices within scientific literature, advocating for a paradigm shift from the prevailing conception of forensic science as a conglomerate of disciplines merely aiding the criminal justice system. instead, it urges a perspective that views forensic science as a discipline studying the informative potential of
the surface of the membrane, retentate is removed from the same side further downstream, whereas the permeate flow is tracked on the other side. in dead - end filtration, the direction of the fluid flow is normal to the membrane surface. both flow geometries offer some advantages and disadvantages. generally, dead - end filtration is used for feasibility studies on a laboratory scale. the dead - end membranes are relatively easy to fabricate which reduces the cost of the separation process. the dead - end membrane separation process is easy to implement and the process is usually cheaper than cross - flow membrane filtration. the dead - end filtration process is usually a batch - type process, where the filtering solution is loaded ( or slowly fed ) into the membrane device, which then allows passage of some particles subject to the driving force. the main disadvantage of dead - end filtration is the extensive membrane fouling and concentration polarization. the fouling is usually induced faster at higher driving forces. membrane fouling and particle retention in a feed solution also builds up a concentration gradients and particle backflow ( concentration polarization ). the tangential flow devices are more cost and labor - intensive, but they are less susceptible to fouling due to the sweeping effects and high shear rates of the passing flow. the most commonly used synthetic membrane devices ( modules ) are flat sheets / plates, spiral wounds, and hollow fibers. flat membranes used in filtration and separation processes can be enhanced with surface patterning, where microscopic structures are introduced to improve performance. these patterns increase surface area, optimize water flow, and reduce fouling, leading to higher permeability and longer membrane lifespan. research has shown that such modifications can significantly enhance efficiency in water purification, energy applications, and industrial separations. flat plates are usually constructed as circular thin flat membrane surfaces to be used in dead - end geometry modules. spiral wounds are constructed from similar flat membranes but in the form of a " pocket " containing two membrane sheets separated by a highly porous support plate. several such pockets are then wound around a tube to create a tangential flow geometry and to reduce membrane fouling. hollow fiber modules consist of an assembly of self - supporting fibers with dense skin separation layers, and a more open matrix helping to withstand pressure gradients and maintain structural integrity. the hollow fiber modules can contain up to 10, 000 fibers ranging from 200 to 2500 ΞΌm in diameter ; the main advantage of hollow fiber modules is the very large surface area within
use less energy than conventional thermal separation processes such as distillation, sublimation or crystallization. the separation process is purely physical and both fractions ( permeate and retentate ) can be obtained as useful products. cold separation using membrane technology is widely used in the food technology, biotechnology and pharmaceutical industries. furthermore, using membranes enables separations to take place that would be impossible using thermal separation methods. for example, it is impossible to separate the constituents of azeotropic liquids or solutes which form isomorphic crystals by distillation or recrystallization but such separations can be achieved using membrane technology. depending on the type of membrane, the selective separation of certain individual substances or substance mixtures is possible. important technical applications include the production of drinking water by reverse osmosis. in waste water treatment, membrane technology is becoming increasingly important. ultra / microfiltration can be very effective in removing colloids and macromolecules from wastewater. this is needed if wastewater is discharged into sensitive waters especially those designated for contact water sports and recreation. about half of the market is in medical applications such as artificial kidneys to remove toxic substances by hemodialysis and as artificial lung for bubble - free supply of oxygen in the blood. the importance of membrane technology is growing in the field of environmental protection ( nano - mem - pro ippc database ). even in modern energy recovery techniques, membranes are increasingly used, for example in fuel cells and in osmotic power plants. = = mass transfer = = two basic models can be distinguished for mass transfer through the membrane : the solution - diffusion model and the hydrodynamic model. in real membranes, these two transport mechanisms certainly occur side by side, especially during ultra - filtration. = = = solution - diffusion model = = = in the solution - diffusion model, transport occurs only by diffusion. the component that needs to be transported must first be dissolved in the membrane. the general approach of the solution - diffusion model is to assume that the chemical potential of the feed and permeate fluids are in equilibrium with the adjacent membrane surfaces such that appropriate expressions for the chemical potential in the fluid and membrane phases can be equated at the solution - membrane interface. this principle is more important for dense membranes without natural pores such as those used for reverse osmosis and in fuel cells. during the filtration process a boundary layer forms on the membrane. this concentration gradient is created by molecules which cannot pass through the membrane. the
". = = extraction = = extractive metallurgy is the practice of removing valuable metals from an ore and refining the extracted raw metals into a purer form. in order to convert a metal oxide or sulphide to a purer metal, the ore must be reduced physically, chemically, or electrolytically. extractive metallurgists are interested in three primary streams : feed, concentrate ( metal oxide / sulphide ) and tailings ( waste ). after mining, large pieces of the ore feed are broken through crushing or grinding in order to obtain particles small enough, where each particle is either mostly valuable or mostly waste. concentrating the particles of value in a form supporting separation enables the desired metal to be removed from waste products. mining may not be necessary, if the ore body and physical environment are conducive to leaching. leaching dissolves minerals in an ore body and results in an enriched solution. the solution is collected and processed to extract valuable metals. ore bodies often contain more than one valuable metal. tailings of a previous process may be used as a feed in another process to extract a secondary product from the original ore. additionally, a concentrate may contain more than one valuable metal. that concentrate would then be processed to separate the valuable metals into individual constituents. = = metal and its alloys = = much effort has been placed on understanding iron β carbon alloy system, which includes steels and cast irons. plain carbon steels ( those that contain essentially only carbon as an alloying element ) are used in low - cost, high - strength applications, where neither weight nor corrosion are a major concern. cast irons, including ductile iron, are also part of the iron - carbon system. iron - manganese - chromium alloys ( hadfield - type steels ) are also used in non - magnetic applications such as directional drilling. other engineering metals include aluminium, chromium, copper, magnesium, nickel, titanium, zinc, and silicon. these metals are most often used as alloys with the noted exception of silicon, which is not a metal. other forms include : stainless steel, particularly austenitic stainless steels, galvanized steel, nickel alloys, titanium alloys, or occasionally copper alloys are used, where resistance to corrosion is important. aluminium alloys and magnesium alloys are commonly used, when a lightweight strong part is required such as in automotive and aerospace applications. copper - nickel alloys ( such as monel ) are used in highly corrosive environments and for non - magnetic applications
= = = = = = environmental remediation = = = environmental remediation is the process through which contaminants or pollutants in soil, water and other media are removed to improve environmental quality. the main focus is the reduction of hazardous substances within the environment. some of the areas involved in environmental remediation include ; soil contamination, hazardous waste, groundwater contamination, oil, gas and chemical spills. there are three most common types of environmental remediation. these include soil, water, and sediment remediation. soil remediation consists of removing contaminants in soil, as these pose great risks to humans and the ecosystem. some examples of this are heavy metals, pesticides, and radioactive materials. depending on the contaminant the remedial processes can be physical, chemical, thermal, or biological. water remediation is one of the most important considering water is an essential natural resource. depending on the source of water there will be different contaminants. surface water contamination mainly consists of agricultural, animal, and industrial waste, as well as acid mine drainage. there has been a rise in the need for water remediation due to the increased discharge of industrial waste, leading to a demand for sustainable water solutions. the market for water remediation is expected to consistently increase to $ 19. 6 billion by 2030. sediment remediation consists of removing contaminated sediments. is it almost similar to soil remediation except it is often more sophisticated as it involves additional contaminants. to reduce the contaminants it is likely to use physical, chemical, and biological processes that help with source control, but if these processes are executed correctly, there ' s a risk of contamination resurfacing. = = = solid waste management = = = solid waste management is the purification, consumption, reuse, disposal, and treatment of solid waste that is undertaken by the government or the ruling bodies of a city / town. it refers to the collection, treatment, and disposal of non - soluble, solid waste material. solid waste is associated with both industrial, institutional, commercial and residential activities. hazardous solid waste, when improperly disposed can encourage the infestation of insects and rodents, contributing to the spread of diseases. some of the most common types of solid waste management include ; landfills, vermicomposting, composting, recycling, and incineration. however, a major barrier for solid waste management practices is the high costs associated with recycling
generally, dead - end filtration is used for feasibility studies on a laboratory scale. the dead - end membranes are relatively easy to fabricate which reduces the cost of the separation process. the dead - end membrane separation process is easy to implement and the process is usually cheaper than cross - flow membrane filtration. the dead - end filtration process is usually a batch - type process, where the filtering solution is loaded ( or slowly fed ) into the membrane device, which then allows passage of some particles subject to the driving force. the main disadvantage of dead - end filtration is the extensive membrane fouling and concentration polarization. the fouling is usually induced faster at higher driving forces. membrane fouling and particle retention in a feed solution also builds up a concentration gradients and particle backflow ( concentration polarization ). the tangential flow devices are more cost and labor - intensive, but they are less susceptible to fouling due to the sweeping effects and high shear rates of the passing flow. the most commonly used synthetic membrane devices ( modules ) are flat sheets / plates, spiral wounds, and hollow fibers. flat membranes used in filtration and separation processes can be enhanced with surface patterning, where microscopic structures are introduced to improve performance. these patterns increase surface area, optimize water flow, and reduce fouling, leading to higher permeability and longer membrane lifespan. research has shown that such modifications can significantly enhance efficiency in water purification, energy applications, and industrial separations. flat plates are usually constructed as circular thin flat membrane surfaces to be used in dead - end geometry modules. spiral wounds are constructed from similar flat membranes but in the form of a " pocket " containing two membrane sheets separated by a highly porous support plate. several such pockets are then wound around a tube to create a tangential flow geometry and to reduce membrane fouling. hollow fiber modules consist of an assembly of self - supporting fibers with dense skin separation layers, and a more open matrix helping to withstand pressure gradients and maintain structural integrity. the hollow fiber modules can contain up to 10, 000 fibers ranging from 200 to 2500 ΞΌm in diameter ; the main advantage of hollow fiber modules is the very large surface area within an enclosed volume, increasing the efficiency of the separation process. the disc tube module uses a cross - flow geometry and consists of a pressure tube and hydraulic discs, which are held by a central tension rod, and membrane cushions that lie between two discs. = = membrane performance and governing equations = = the selection of synthetic membranes
the elimination of metabolic wastes. these enzyme - catalyzed reactions allow organisms to grow and reproduce, maintain their structures, and respond to their environments. metabolic reactions may be categorized as catabolic β the breaking down of compounds ( for example, the breaking down of glucose to pyruvate by cellular respiration ) ; or anabolic β the building up ( synthesis ) of compounds ( such as proteins, carbohydrates, lipids, and nucleic acids ). usually, catabolism releases energy, and anabolism consumes energy. the chemical reactions of metabolism are organized into metabolic pathways, in which one chemical is transformed through a series of steps into another chemical, each step being facilitated by a specific enzyme. enzymes are crucial to metabolism because they allow organisms to drive desirable reactions that require energy that will not occur by themselves, by coupling them to spontaneous reactions that release energy. enzymes act as catalysts β they allow a reaction to proceed more rapidly without being consumed by it β by reducing the amount of activation energy needed to convert reactants into products. enzymes also allow the regulation of the rate of a metabolic reaction, for example in response to changes in the cell ' s environment or to signals from other cells. = = = cellular respiration = = = cellular respiration is a set of metabolic reactions and processes that take place in cells to convert chemical energy from nutrients into adenosine triphosphate ( atp ), and then release waste products. the reactions involved in respiration are catabolic reactions, which break large molecules into smaller ones, releasing energy. respiration is one of the key ways a cell releases chemical energy to fuel cellular activity. the overall reaction occurs in a series of biochemical steps, some of which are redox reactions. although cellular respiration is technically a combustion reaction, it clearly does not resemble one when it occurs in a cell because of the slow, controlled release of energy from the series of reactions. sugar in the form of glucose is the main nutrient used by animal and plant cells in respiration. cellular respiration involving oxygen is called aerobic respiration, which has four stages : glycolysis, citric acid cycle ( or krebs cycle ), electron transport chain, and oxidative phosphorylation. glycolysis is a metabolic process that occurs in the cytoplasm whereby glucose is converted into two pyruvates, with two net molecules of atp being produced at the same time. each pyruvate is then
the walls of a victim ' s stomach. toxicology, a subfield of forensic chemistry, focuses on detecting and identifying drugs, poisons, and other toxic substances in biological samples. forensic toxicologists work on cases involving drug overdoses, poisoning, and substance abuse. their work is critical in determining whether harmful substances play a role in a person β s death or impairment. read more james marsh was the first to apply this new science to the art of forensics. he was called by the prosecution in a murder trial to give evidence as a chemist in 1832. the defendant, john bodle, was accused of poisoning his grandfather with arsenic - laced coffee. marsh performed the standard test by mixing a suspected sample with hydrogen sulfide and hydrochloric acid. while he was able to detect arsenic as yellow arsenic trisulfide, when it was shown to the jury it had deteriorated, allowing the suspect to be acquitted due to reasonable doubt. annoyed by that, marsh developed a much better test. he combined a sample containing arsenic with sulfuric acid and arsenic - free zinc, resulting in arsine gas. the gas was ignited, and it decomposed to pure metallic arsenic, which, when passed to a cold surface, would appear as a silvery - black deposit. so sensitive was the test, known formally as the marsh test, that it could detect as little as one - fiftieth of a milligram of arsenic. he first described this test in the edinburgh philosophical journal in 1836. = = = ballistics and firearms = = = ballistics is " the science of the motion of projectiles in flight ". in forensic science, analysts examine the patterns left on bullets and cartridge casings after being ejected from a weapon. when fired, a bullet is left with indentations and markings that are unique to the barrel and firing pin of the firearm that ejected the bullet. this examination can help scientists identify possible makes and models of weapons connected to a crime. henry goddard at scotland yard pioneered the use of bullet comparison in 1835. he noticed a flaw in the bullet that killed the victim and was able to trace this back to the mold that was used in the manufacturing process. = = = anthropometry = = = the french police officer alphonse bertillon was the first to apply the anthropological technique of anthropometry to law enforcement, thereby creating an identification system based on physical measurements. before that time, criminals could be identified only by name or photograph. dissatisfied with the ad hoc methods used to identify captured
the structural components of cells. as a by - product of photosynthesis, plants release oxygen into the atmosphere, a gas that is required by nearly all living things to carry out cellular respiration. in addition, they are influential in the global carbon and water cycles and plant roots bind and stabilise soils, preventing soil erosion. plants are crucial to the future of human society as they provide food, oxygen, biochemicals, and products for people, as well as creating and preserving soil. historically, all living things were classified as either animals or plants and botany covered the study of all organisms not considered animals. botanists examine both the internal functions and processes within plant organelles, cells, tissues, whole plants, plant populations and plant communities. at each of these levels, a botanist may be concerned with the classification ( taxonomy ), phylogeny and evolution, structure ( anatomy and morphology ), or function ( physiology ) of plant life. the strictest definition of " plant " includes only the " land plants " or embryophytes, which include seed plants ( gymnosperms, including the pines, and flowering plants ) and the free - sporing cryptogams including ferns, clubmosses, liverworts, hornworts and mosses. embryophytes are multicellular eukaryotes descended from an ancestor that obtained its energy from sunlight by photosynthesis. they have life cycles with alternating haploid and diploid phases. the sexual haploid phase of embryophytes, known as the gametophyte, nurtures the developing diploid embryo sporophyte within its tissues for at least part of its life, even in the seed plants, where the gametophyte itself is nurtured by its parent sporophyte. other groups of organisms that were previously studied by botanists include bacteria ( now studied in bacteriology ), fungi ( mycology ) β including lichen - forming fungi ( lichenology ), non - chlorophyte algae ( phycology ), and viruses ( virology ). however, attention is still given to these groups by botanists, and fungi ( including lichens ) and photosynthetic protists are usually covered in introductory botany courses. palaeobotanists study ancient plants in the fossil record to provide information about the evolutionary history of plants. cyanobacteria, the first oxygen - releasing photosynthetic organisms on earth, are thought to have given rise to the
Question: What is a decomposer?
A) a living thing that breaks down dead organisms
B) a living thing that makes its own food
C) an animal that eats live plants
D) an animal that hunts other animals
|
A) a living thing that breaks down dead organisms
|
Context:
and cell phones are a particular challenge because the stream of data can interfere with focusing and learning. although these technologies affect adults too, young people may be more influenced by it as their developing brains can easily become habituated to switching tasks and become unaccustomed to sustaining attention. too much information, coming too rapidly, can overwhelm thinking. technology is " rapidly and profoundly altering our brains. " high exposure levels stimulate brain cell alteration and release neurotransmitters, which causes the strengthening of some neural pathways and the weakening of others. this leads to heightened stress levels on the brain that, at first, boost energy levels, but, over time, actually augment memory, impair cognition, lead to depression, and alter the neural circuitry of the hippocampus, amygdala and prefrontal cortex. these are the brain regions that control mood and thought. if unchecked, the underlying structure of the brain could be altered. overstimulation due to technology may begin too young. when children are exposed before the age of seven, important developmental tasks may be delayed, and bad learning habits might develop, which " deprives children of the exploration and play that they need to develop. " media psychology is an emerging specialty field that embraces electronic devices and the sensory behaviors occurring from the use of educational technology in learning. = = = sociocultural criticism = = = according to lai, " the learning environment is a complex system where the interplay and interactions of many things impact the outcome of learning. " when technology is brought into an educational setting, the pedagogical setting changes in that technology - driven teaching can change the entire meaning of an activity without adequate research validation. if technology monopolizes an activity, students can begin to develop the sense that " life would scarcely be thinkable without technology. " leo marx considered the word " technology " itself as problematic, susceptible to reification and " phantom objectivity ", which conceals its fundamental nature as something that is only valuable insofar as it benefits the human condition. technology ultimately comes down to affecting the relations between people, but this notion is obfuscated when technology is treated as an abstract notion devoid of good and evil. langdon winner makes a similar point by arguing that the underdevelopment of the philosophy of technology leaves us with an overly simplistic reduction in our discourse to the supposedly dichotomous notions of the " making " versus the " uses " of new technologies and that a narrow focus on " use
delay of ripening, increase of juice yield, and improvement of re - hydration. irradiation is a more general term of deliberate exposure of materials to radiation to achieve a technical goal ( in this context ' ionizing radiation ' is implied ). as such it is also used on non - food items, such as medical hardware, plastics, tubes for gas - pipelines, hoses for floor - heating, shrink - foils for food packaging, automobile parts, wires and cables ( isolation ), tires, and even gemstones. compared to the amount of food irradiated, the volume of those every - day applications is huge but not noticed by the consumer. the genuine effect of processing food by ionizing radiation relates to damages to the dna, the basic genetic information for life. microorganisms can no longer proliferate and continue their malignant or pathogenic activities. spoilage causing micro - organisms cannot continue their activities. insects do not survive or become incapable of procreation. plants cannot continue the natural ripening or aging process. all these effects are beneficial to the consumer and the food industry, likewise. the amount of energy imparted for effective food irradiation is low compared to cooking the same ; even at a typical dose of 10 kgy most food, which is ( with regard to warming ) physically equivalent to water, would warm by only about 2. 5 Β°c ( 4. 5 Β°f ). the specialty of processing food by ionizing radiation is the fact, that the energy density per atomic transition is very high, it can cleave molecules and induce ionization ( hence the name ) which cannot be achieved by mere heating. this is the reason for new beneficial effects, however at the same time, for new concerns. the treatment of solid food by ionizing radiation can provide an effect similar to heat pasteurization of liquids, such as milk. however, the use of the term, cold pasteurization, to describe irradiated foods is controversial, because pasteurization and irradiation are fundamentally different processes, although the intended end results can in some cases be similar. detractors of food irradiation have concerns about the health hazards of induced radioactivity. a report for the industry advocacy group american council on science and health entitled " irradiated foods " states : " the types of radiation sources approved for the treatment of foods have specific energy levels well below that which would cause any element in food to become radioactive. food undergoing irradiation does not become any more
. an example of this would be bt corn. whether or not green biotechnology products such as this are ultimately more environmentally friendly is a topic of considerable debate. it is commonly considered as the next phase of green revolution, which can be seen as a platform to eradicate world hunger by using technologies which enable the production of more fertile and resistant, towards biotic and abiotic stress, plants and ensures application of environmentally friendly fertilizers and the use of biopesticides, it is mainly focused on the development of agriculture. on the other hand, some of the uses of green biotechnology involve microorganisms to clean and reduce waste. red biotechnology is the use of biotechnology in the medical and pharmaceutical industries, and health preservation. this branch involves the production of vaccines and antibiotics, regenerative therapies, creation of artificial organs and new diagnostics of diseases. as well as the development of hormones, stem cells, antibodies, sirna and diagnostic tests. white biotechnology, also known as industrial biotechnology, is biotechnology applied to industrial processes. an example is the designing of an organism to produce a useful chemical. another example is the using of enzymes as industrial catalysts to either produce valuable chemicals or destroy hazardous / polluting chemicals. white biotechnology tends to consume less in resources than traditional processes used to produce industrial goods. yellow biotechnology refers to the use of biotechnology in food production ( food industry ), for example in making wine ( winemaking ), cheese ( cheesemaking ), and beer ( brewing ) by fermentation. it has also been used to refer to biotechnology applied to insects. this includes biotechnology - based approaches for the control of harmful insects, the characterisation and utilisation of active ingredients or genes of insects for research, or application in agriculture and medicine and various other approaches. gray biotechnology is dedicated to environmental applications, and focused on the maintenance of biodiversity and the remotion of pollutants. brown biotechnology is related to the management of arid lands and deserts. one application is the creation of enhanced seeds that resist extreme environmental conditions of arid regions, which is related to the innovation, creation of agriculture techniques and management of resources. violet biotechnology is related to law, ethical and philosophical issues around biotechnology. microbial biotechnology has been proposed for the rapidly emerging area of biotechnology applications in space and microgravity ( space bioeconomy ) dark biotechnology is the color associated with bioterrorism or biological weapons and biowarfare which uses microorganisms, and toxins to cause diseases and death in humans, livestock and
the world is changing at an ever - increasing pace. and it has changed in a much more fundamental way than one would think, primarily because it has become more connected and interdependent than in our entire history. every new product, every new invention can be combined with those that existed before, thereby creating an explosion of complexity : structural complexity, dynamic complexity, functional complexity, and algorithmic complexity. how to respond to this challenge? and what are the costs?
industrial applications. this branch of biotechnology is the most used for the industries of refining and combustion principally on the production of bio - oils with photosynthetic micro - algae. green biotechnology is biotechnology applied to agricultural processes. an example would be the selection and domestication of plants via micropropagation. another example is the designing of transgenic plants to grow under specific environments in the presence ( or absence ) of chemicals. one hope is that green biotechnology might produce more environmentally friendly solutions than traditional industrial agriculture. an example of this is the engineering of a plant to express a pesticide, thereby ending the need of external application of pesticides. an example of this would be bt corn. whether or not green biotechnology products such as this are ultimately more environmentally friendly is a topic of considerable debate. it is commonly considered as the next phase of green revolution, which can be seen as a platform to eradicate world hunger by using technologies which enable the production of more fertile and resistant, towards biotic and abiotic stress, plants and ensures application of environmentally friendly fertilizers and the use of biopesticides, it is mainly focused on the development of agriculture. on the other hand, some of the uses of green biotechnology involve microorganisms to clean and reduce waste. red biotechnology is the use of biotechnology in the medical and pharmaceutical industries, and health preservation. this branch involves the production of vaccines and antibiotics, regenerative therapies, creation of artificial organs and new diagnostics of diseases. as well as the development of hormones, stem cells, antibodies, sirna and diagnostic tests. white biotechnology, also known as industrial biotechnology, is biotechnology applied to industrial processes. an example is the designing of an organism to produce a useful chemical. another example is the using of enzymes as industrial catalysts to either produce valuable chemicals or destroy hazardous / polluting chemicals. white biotechnology tends to consume less in resources than traditional processes used to produce industrial goods. yellow biotechnology refers to the use of biotechnology in food production ( food industry ), for example in making wine ( winemaking ), cheese ( cheesemaking ), and beer ( brewing ) by fermentation. it has also been used to refer to biotechnology applied to insects. this includes biotechnology - based approaches for the control of harmful insects, the characterisation and utilisation of active ingredients or genes of insects for research, or application in agriculture and medicine and various other approaches. gray biotechnology is dedicated to environmental applications, and focused on the maintenance of biodiversity and the remotion of poll
no offspring, to reduce the population. in industrial and food applications, radiation is used for sterilization of tools and equipment. an advantage is that the object may be sealed in plastic before sterilization. an emerging use in food production is the sterilization of food using food irradiation. food irradiation is the process of exposing food to ionizing radiation in order to destroy microorganisms, bacteria, viruses, or insects that might be present in the food. the radiation sources used include radioisotope gamma ray sources, x - ray generators and electron accelerators. further applications include sprout inhibition, delay of ripening, increase of juice yield, and improvement of re - hydration. irradiation is a more general term of deliberate exposure of materials to radiation to achieve a technical goal ( in this context ' ionizing radiation ' is implied ). as such it is also used on non - food items, such as medical hardware, plastics, tubes for gas - pipelines, hoses for floor - heating, shrink - foils for food packaging, automobile parts, wires and cables ( isolation ), tires, and even gemstones. compared to the amount of food irradiated, the volume of those every - day applications is huge but not noticed by the consumer. the genuine effect of processing food by ionizing radiation relates to damages to the dna, the basic genetic information for life. microorganisms can no longer proliferate and continue their malignant or pathogenic activities. spoilage causing micro - organisms cannot continue their activities. insects do not survive or become incapable of procreation. plants cannot continue the natural ripening or aging process. all these effects are beneficial to the consumer and the food industry, likewise. the amount of energy imparted for effective food irradiation is low compared to cooking the same ; even at a typical dose of 10 kgy most food, which is ( with regard to warming ) physically equivalent to water, would warm by only about 2. 5 Β°c ( 4. 5 Β°f ). the specialty of processing food by ionizing radiation is the fact, that the energy density per atomic transition is very high, it can cleave molecules and induce ionization ( hence the name ) which cannot be achieved by mere heating. this is the reason for new beneficial effects, however at the same time, for new concerns. the treatment of solid food by ionizing radiation can provide an effect similar to heat pasteurization of liquids, such as milk. however
best - known and controversial applications of genetic engineering is the creation and use of genetically modified crops or genetically modified livestock to produce genetically modified food. crops have been developed to increase production, increase tolerance to abiotic stresses, alter the composition of the food, or to produce novel products. the first crops to be released commercially on a large scale provided protection from insect pests or tolerance to herbicides. fungal and virus resistant crops have also been developed or are in development. this makes the insect and weed management of crops easier and can indirectly increase crop yield. gm crops that directly improve yield by accelerating growth or making the plant more hardy ( by improving salt, cold or drought tolerance ) are also under development. in 2016 salmon have been genetically modified with growth hormones to reach normal adult size much faster. gmos have been developed that modify the quality of produce by increasing the nutritional value or providing more industrially useful qualities or quantities. the amflora potato produces a more industrially useful blend of starches. soybeans and canola have been genetically modified to produce more healthy oils. the first commercialised gm food was a tomato that had delayed ripening, increasing its shelf life. plants and animals have been engineered to produce materials they do not normally make. pharming uses crops and animals as bioreactors to produce vaccines, drug intermediates, or the drugs themselves ; the useful product is purified from the harvest and then used in the standard pharmaceutical production process. cows and goats have been engineered to express drugs and other proteins in their milk, and in 2009 the fda approved a drug produced in goat milk. = = = other applications = = = genetic engineering has potential applications in conservation and natural area management. gene transfer through viral vectors has been proposed as a means of controlling invasive species as well as vaccinating threatened fauna from disease. transgenic trees have been suggested as a way to confer resistance to pathogens in wild populations. with the increasing risks of maladaptation in organisms as a result of climate change and other perturbations, facilitated adaptation through gene tweaking could be one solution to reducing extinction risks. applications of genetic engineering in conservation are thus far mostly theoretical and have yet to be put into practice. genetic engineering is also being used to create microbial art. some bacteria have been genetically engineered to create black and white photographs. novelty items such as lavender - colored carnations, blue roses, and glowing fish, have also been produced through genetic engineering. = = regulation = = the regulation of genetic engineering
and child health in boston, said of the digital generation, " their brains are rewarded not for staying on task, but for jumping to the next thing. the worry is we ' re raising a generation of kids in front of screens whose brains are going to be wired differently. " students have always faced distractions ; computers and cell phones are a particular challenge because the stream of data can interfere with focusing and learning. although these technologies affect adults too, young people may be more influenced by it as their developing brains can easily become habituated to switching tasks and become unaccustomed to sustaining attention. too much information, coming too rapidly, can overwhelm thinking. technology is " rapidly and profoundly altering our brains. " high exposure levels stimulate brain cell alteration and release neurotransmitters, which causes the strengthening of some neural pathways and the weakening of others. this leads to heightened stress levels on the brain that, at first, boost energy levels, but, over time, actually augment memory, impair cognition, lead to depression, and alter the neural circuitry of the hippocampus, amygdala and prefrontal cortex. these are the brain regions that control mood and thought. if unchecked, the underlying structure of the brain could be altered. overstimulation due to technology may begin too young. when children are exposed before the age of seven, important developmental tasks may be delayed, and bad learning habits might develop, which " deprives children of the exploration and play that they need to develop. " media psychology is an emerging specialty field that embraces electronic devices and the sensory behaviors occurring from the use of educational technology in learning. = = = sociocultural criticism = = = according to lai, " the learning environment is a complex system where the interplay and interactions of many things impact the outcome of learning. " when technology is brought into an educational setting, the pedagogical setting changes in that technology - driven teaching can change the entire meaning of an activity without adequate research validation. if technology monopolizes an activity, students can begin to develop the sense that " life would scarcely be thinkable without technology. " leo marx considered the word " technology " itself as problematic, susceptible to reification and " phantom objectivity ", which conceals its fundamental nature as something that is only valuable insofar as it benefits the human condition. technology ultimately comes down to affecting the relations between people, but this notion is obfuscated when technology is treated as an abstract notion devoid of
the third millennium bc in palmela, portugal, los millares, spain, and stonehenge, united kingdom. the precise beginnings, however, have not be clearly ascertained and new discoveries are both continuous and ongoing. in approximately 1900 bc, ancient iron smelting sites existed in tamil nadu. in the near east, about 3, 500 bc, it was discovered that by combining copper and tin, a superior metal could be made, an alloy called bronze. this represented a major technological shift known as the bronze age. the extraction of iron from its ore into a workable metal is much more difficult than for copper or tin. the process appears to have been invented by the hittites in about 1200 bc, beginning the iron age. the secret of extracting and working iron was a key factor in the success of the philistines. historical developments in ferrous metallurgy can be found in a wide variety of past cultures and civilizations. this includes the ancient and medieval kingdoms and empires of the middle east and near east, ancient iran, ancient egypt, ancient nubia, and anatolia in present - day turkey, ancient nok, carthage, the celts, greeks and romans of ancient europe, medieval europe, ancient and medieval china, ancient and medieval india, ancient and medieval japan, amongst others. a 16th century book by georg agricola, de re metallica, describes the highly developed and complex processes of mining metal ores, metal extraction, and metallurgy of the time. agricola has been described as the " father of metallurgy ". = = extraction = = extractive metallurgy is the practice of removing valuable metals from an ore and refining the extracted raw metals into a purer form. in order to convert a metal oxide or sulphide to a purer metal, the ore must be reduced physically, chemically, or electrolytically. extractive metallurgists are interested in three primary streams : feed, concentrate ( metal oxide / sulphide ) and tailings ( waste ). after mining, large pieces of the ore feed are broken through crushing or grinding in order to obtain particles small enough, where each particle is either mostly valuable or mostly waste. concentrating the particles of value in a form supporting separation enables the desired metal to be removed from waste products. mining may not be necessary, if the ore body and physical environment are conducive to leaching. leaching dissolves minerals in an ore body and results in an enriched solution. the solution
was used before copper smelting was known. copper smelting is believed to have originated when the technology of pottery kilns allowed sufficiently high temperatures. the concentration of various elements such as arsenic increase with depth in copper ore deposits and smelting of these ores yields arsenical bronze, which can be sufficiently work hardened to be suitable for making tools. bronze is an alloy of copper with tin ; the latter being found in relatively few deposits globally caused a long time to elapse before true tin bronze became widespread. ( see : tin sources and trade in ancient times ) bronze was a major advancement over stone as a material for making tools, both because of its mechanical properties like strength and ductility and because it could be cast in molds to make intricately shaped objects. bronze significantly advanced shipbuilding technology with better tools and bronze nails. bronze nails replaced the old method of attaching boards of the hull with cord woven through drilled holes. better ships enabled long - distance trade and the advance of civilization. this technological trend apparently began in the fertile crescent and spread outward over time. these developments were not, and still are not, universal. the three - age system does not accurately describe the technology history of groups outside of eurasia, and does not apply at all in the case of some isolated populations, such as the spinifex people, the sentinelese, and various amazonian tribes, which still make use of stone age technology, and have not developed agricultural or metal technology. these villages preserve traditional customs in the face of global modernity, exhibiting a remarkable resistance to the rapid advancement of technology. = = = = iron age = = = = before iron smelting was developed the only iron was obtained from meteorites and is usually identified by having nickel content. meteoric iron was rare and valuable, but was sometimes used to make tools and other implements, such as fish hooks. the iron age involved the adoption of iron smelting technology. it generally replaced bronze and made it possible to produce tools which were stronger, lighter and cheaper to make than bronze equivalents. the raw materials to make iron, such as ore and limestone, are far more abundant than copper and especially tin ores. consequently, iron was produced in many areas. it was not possible to mass manufacture steel or pure iron because of the high temperatures required. furnaces could reach melting temperature but the crucibles and molds needed for melting and casting had not been developed. steel could be produced by forging bloomery iron to reduce the carbon content in a
Question: Modern technology has had some positive and negative effects on society. Which would be a negative effect of advances in technology?
A) human jobs replaced by more efficient machines
B) decrease in production of solid waste
C) use of alternate sources of energy
D) reduction in habitat destruction
|
A) human jobs replaced by more efficient machines
|
Context:
various forms that are characteristic of its life cycle. there are four key processes that underlie development : determination, differentiation, morphogenesis, and growth. determination sets the developmental fate of a cell, which becomes more restrictive during development. differentiation is the process by which specialized cells arise from less specialized cells such as stem cells. stem cells are undifferentiated or partially differentiated cells that can differentiate into various types of cells and proliferate indefinitely to produce more of the same stem cell. cellular differentiation dramatically changes a cell ' s size, shape, membrane potential, metabolic activity, and responsiveness to signals, which are largely due to highly controlled modifications in gene expression and epigenetics. with a few exceptions, cellular differentiation almost never involves a change in the dna sequence itself. thus, different cells can have very different physical characteristics despite having the same genome. morphogenesis, or the development of body form, is the result of spatial differences in gene expression. a small fraction of the genes in an organism ' s genome called the developmental - genetic toolkit control the development of that organism. these toolkit genes are highly conserved among phyla, meaning that they are ancient and very similar in widely separated groups of animals. differences in deployment of toolkit genes affect the body plan and the number, identity, and pattern of body parts. among the most important toolkit genes are the hox genes. hox genes determine where repeating parts, such as the many vertebrae of snakes, will grow in a developing embryo or larva. = = evolution = = = = = evolutionary processes = = = evolution is a central organizing concept in biology. it is the change in heritable characteristics of populations over successive generations. in artificial selection, animals were selectively bred for specific traits. given that traits are inherited, populations contain a varied mix of traits, and reproduction is able to increase any population, darwin argued that in the natural world, it was nature that played the role of humans in selecting for specific traits. darwin inferred that individuals who possessed heritable traits better adapted to their environments are more likely to survive and produce more offspring than other individuals. he further inferred that this would lead to the accumulation of favorable traits over successive generations, thereby increasing the match between the organisms and their environment. = = = speciation = = = a species is a group of organisms that mate with one another and speciation is the process by which one lineage splits into two lineages as a result of having evolved independently from each other
to chromatin, which is a complex of dna and protein found in eukaryotic cells. = = = genes, development, and evolution = = = development is the process by which a multicellular organism ( plant or animal ) goes through a series of changes, starting from a single cell, and taking on various forms that are characteristic of its life cycle. there are four key processes that underlie development : determination, differentiation, morphogenesis, and growth. determination sets the developmental fate of a cell, which becomes more restrictive during development. differentiation is the process by which specialized cells arise from less specialized cells such as stem cells. stem cells are undifferentiated or partially differentiated cells that can differentiate into various types of cells and proliferate indefinitely to produce more of the same stem cell. cellular differentiation dramatically changes a cell ' s size, shape, membrane potential, metabolic activity, and responsiveness to signals, which are largely due to highly controlled modifications in gene expression and epigenetics. with a few exceptions, cellular differentiation almost never involves a change in the dna sequence itself. thus, different cells can have very different physical characteristics despite having the same genome. morphogenesis, or the development of body form, is the result of spatial differences in gene expression. a small fraction of the genes in an organism ' s genome called the developmental - genetic toolkit control the development of that organism. these toolkit genes are highly conserved among phyla, meaning that they are ancient and very similar in widely separated groups of animals. differences in deployment of toolkit genes affect the body plan and the number, identity, and pattern of body parts. among the most important toolkit genes are the hox genes. hox genes determine where repeating parts, such as the many vertebrae of snakes, will grow in a developing embryo or larva. = = evolution = = = = = evolutionary processes = = = evolution is a central organizing concept in biology. it is the change in heritable characteristics of populations over successive generations. in artificial selection, animals were selectively bred for specific traits. given that traits are inherited, populations contain a varied mix of traits, and reproduction is able to increase any population, darwin argued that in the natural world, it was nature that played the role of humans in selecting for specific traits. darwin inferred that individuals who possessed heritable traits better adapted to their environments are more likely to survive and produce more offspring than other individuals. he further inferred that this would lead to the
shaping the cell. cell membranes are involved in various cellular processes such as cell adhesion, storing electrical energy, and cell signalling and serve as the attachment surface for several extracellular structures such as a cell wall, glycocalyx, and cytoskeleton. within the cytoplasm of a cell, there are many biomolecules such as proteins and nucleic acids. in addition to biomolecules, eukaryotic cells have specialized structures called organelles that have their own lipid bilayers or are spatially units. these organelles include the cell nucleus, which contains most of the cell ' s dna, or mitochondria, which generate adenosine triphosphate ( atp ) to power cellular processes. other organelles such as endoplasmic reticulum and golgi apparatus play a role in the synthesis and packaging of proteins, respectively. biomolecules such as proteins can be engulfed by lysosomes, another specialized organelle. plant cells have additional organelles that distinguish them from animal cells such as a cell wall that provides support for the plant cell, chloroplasts that harvest sunlight energy to produce sugar, and vacuoles that provide storage and structural support as well as being involved in reproduction and breakdown of plant seeds. eukaryotic cells also have cytoskeleton that is made up of microtubules, intermediate filaments, and microfilaments, all of which provide support for the cell and are involved in the movement of the cell and its organelles. in terms of their structural composition, the microtubules are made up of tubulin ( e. g., Ξ± - tubulin and Ξ² - tubulin ) whereas intermediate filaments are made up of fibrous proteins. microfilaments are made up of actin molecules that interact with other strands of proteins. = = = metabolism = = = all cells require energy to sustain cellular processes. metabolism is the set of chemical reactions in an organism. the three main purposes of metabolism are : the conversion of food to energy to run cellular processes ; the conversion of food / fuel to monomer building blocks ; and the elimination of metabolic wastes. these enzyme - catalyzed reactions allow organisms to grow and reproduce, maintain their structures, and respond to their environments. metabolic reactions may be categorized as catabolic β the breaking down of compounds ( for example, the breaking down of glucose to pyruvate by cellular respiration
young plant cells, and electroporation, which involves using an electric shock to make the cell membrane permeable to plasmid dna. as only a single cell is transformed with genetic material, the organism must be regenerated from that single cell. in plants this is accomplished through the use of tissue culture. in animals it is necessary to ensure that the inserted dna is present in the embryonic stem cells. bacteria consist of a single cell and reproduce clonally so regeneration is not necessary. selectable markers are used to easily differentiate transformed from untransformed cells. these markers are usually present in the transgenic organism, although a number of strategies have been developed that can remove the selectable marker from the mature transgenic plant. further testing using pcr, southern hybridization, and dna sequencing is conducted to confirm that an organism contains the new gene. these tests can also confirm the chromosomal location and copy number of the inserted gene. the presence of the gene does not guarantee it will be expressed at appropriate levels in the target tissue so methods that look for and measure the gene products ( rna and protein ) are also used. these include northern hybridisation, quantitative rt - pcr, western blot, immunofluorescence, elisa and phenotypic analysis. the new genetic material can be inserted randomly within the host genome or targeted to a specific location. the technique of gene targeting uses homologous recombination to make desired changes to a specific endogenous gene. this tends to occur at a relatively low frequency in plants and animals and generally requires the use of selectable markers. the frequency of gene targeting can be greatly enhanced through genome editing. genome editing uses artificially engineered nucleases that create specific double - stranded breaks at desired locations in the genome, and use the cell ' s endogenous mechanisms to repair the induced break by the natural processes of homologous recombination and nonhomologous end - joining. there are four families of engineered nucleases : meganucleases, zinc finger nucleases, transcription activator - like effector nucleases ( talens ), and the cas9 - guiderna system ( adapted from crispr ). talen and crispr are the two most commonly used and each has its own advantages. talens have greater target specificity, while crispr is easier to design and more efficient. in addition to enhancing gene targeting, engineered nucleases can be used to introduce mutations
oxygen, carbon dioxide, and water to pass through while restricting the movement of larger molecules and charged particles such as ions. cell membranes also contain membrane proteins, including integral membrane proteins that go across the membrane serving as membrane transporters, and peripheral proteins that loosely attach to the outer side of the cell membrane, acting as enzymes shaping the cell. cell membranes are involved in various cellular processes such as cell adhesion, storing electrical energy, and cell signalling and serve as the attachment surface for several extracellular structures such as a cell wall, glycocalyx, and cytoskeleton. within the cytoplasm of a cell, there are many biomolecules such as proteins and nucleic acids. in addition to biomolecules, eukaryotic cells have specialized structures called organelles that have their own lipid bilayers or are spatially units. these organelles include the cell nucleus, which contains most of the cell ' s dna, or mitochondria, which generate adenosine triphosphate ( atp ) to power cellular processes. other organelles such as endoplasmic reticulum and golgi apparatus play a role in the synthesis and packaging of proteins, respectively. biomolecules such as proteins can be engulfed by lysosomes, another specialized organelle. plant cells have additional organelles that distinguish them from animal cells such as a cell wall that provides support for the plant cell, chloroplasts that harvest sunlight energy to produce sugar, and vacuoles that provide storage and structural support as well as being involved in reproduction and breakdown of plant seeds. eukaryotic cells also have cytoskeleton that is made up of microtubules, intermediate filaments, and microfilaments, all of which provide support for the cell and are involved in the movement of the cell and its organelles. in terms of their structural composition, the microtubules are made up of tubulin ( e. g., Ξ± - tubulin and Ξ² - tubulin ) whereas intermediate filaments are made up of fibrous proteins. microfilaments are made up of actin molecules that interact with other strands of proteins. = = = metabolism = = = all cells require energy to sustain cellular processes. metabolism is the set of chemical reactions in an organism. the three main purposes of metabolism are : the conversion of food to energy to run cellular processes ; the conversion of food / fuel to monomer building blocks ; and
combination of chemical and mechanical stimulation. cell proliferation and differentiation are largely influenced by mechanical and biochemical cues in the surrounding extracellular matrix environment. bioreactors are typically developed to replicate the specific physiological environment of the tissue being grown ( e. g., flex and fluid shearing for heart tissue growth ). this can allow specialized cell lines to thrive in cultures replicating their native environments, but it also makes bioreactors attractive tools for culturing stem cells. a successful stem - cell - based bioreactor is effective at expanding stem cells with uniform properties and / or promoting controlled, reproducible differentiation into selected mature cell types. there are a variety of bioreactors designed for 3d cell cultures. there are small plastic cylindrical chambers, as well as glass chambers, with regulated internal humidity and moisture specifically engineered for the purpose of growing cells in three dimensions. the bioreactor uses bioactive synthetic materials such as polyethylene terephthalate membranes to surround the spheroid cells in an environment that maintains high levels of nutrients. they are easy to open and close, so that cell spheroids can be removed for testing, yet the chamber is able to maintain 100 % humidity throughout. this humidity is important to achieve maximum cell growth and function. the bioreactor chamber is part of a larger device that rotates to ensure equal cell growth in each direction across three dimensions. quinxell technologies now under quintech life sciences from singapore has developed a bioreactor known as the tisxell biaxial bioreactor which is specially designed for the purpose of tissue engineering. it is the first bioreactor in the world to have a spherical glass chamber with biaxial rotation ; specifically to mimic the rotation of the fetus in the womb ; which provides a conducive environment for the growth of tissues. multiple forms of mechanical stimulation have also been combined into a single bioreactor. using gene expression analysis, one academic study found that applying a combination of cyclic strain and ultrasound stimulation to pre - osteoblast cells in a bioreactor accelerated matrix maturation and differentiation. the technology of this combined stimulation bioreactor could be used to grow bone cells more quickly and effectively in future clinical stem cell therapies. mc2 biotek has also developed a bioreactor known as prototissue that uses gas exchange to maintain high oxygen levels within the cell chamber ; improving upon previous bioreactors, since the higher oxygen levels help the cell grow and undergo normal cell respiration. active
the cell ' s life. some epigenetic changes have been shown to be heritable, while others are reset in the germ cells. epigenetic changes in eukaryotic biology serve to regulate the process of cellular differentiation. during morphogenesis, totipotent stem cells become the various pluripotent cell lines of the embryo, which in turn become fully differentiated cells. a single fertilised egg cell, the zygote, gives rise to the many different plant cell types including parenchyma, xylem vessel elements, phloem sieve tubes, guard cells of the epidermis, etc. as it continues to divide. the process results from the epigenetic activation of some genes and inhibition of others. unlike animals, many plant cells, particularly those of the parenchyma, do not terminally differentiate, remaining totipotent with the ability to give rise to a new individual plant. exceptions include highly lignified cells, the sclerenchyma and xylem which are dead at maturity, and the phloem sieve tubes which lack nuclei. while plants use many of the same epigenetic mechanisms as animals, such as chromatin remodelling, an alternative hypothesis is that plants set their gene expression patterns using positional information from the environment and surrounding cells to determine their developmental fate. epigenetic changes can lead to paramutations, which do not follow the mendelian heritage rules. these epigenetic marks are carried from one generation to the next, with one allele inducing a change on the other. = = plant evolution = = the chloroplasts of plants have a number of biochemical, structural and genetic similarities to cyanobacteria, ( commonly but incorrectly known as " blue - green algae " ) and are thought to be derived from an ancient endosymbiotic relationship between an ancestral eukaryotic cell and a cyanobacterial resident. the algae are a polyphyletic group and are placed in various divisions, some more closely related to plants than others. there are many differences between them in features such as cell wall composition, biochemistry, pigmentation, chloroplast structure and nutrient reserves. the algal division charophyta, sister to the green algal division chlorophyta, is considered to contain the ancestor of true plants. the charophyte class charophyceae and the land plant sub - kingdom embryophy
, there are many biomolecules such as proteins and nucleic acids. in addition to biomolecules, eukaryotic cells have specialized structures called organelles that have their own lipid bilayers or are spatially units. these organelles include the cell nucleus, which contains most of the cell ' s dna, or mitochondria, which generate adenosine triphosphate ( atp ) to power cellular processes. other organelles such as endoplasmic reticulum and golgi apparatus play a role in the synthesis and packaging of proteins, respectively. biomolecules such as proteins can be engulfed by lysosomes, another specialized organelle. plant cells have additional organelles that distinguish them from animal cells such as a cell wall that provides support for the plant cell, chloroplasts that harvest sunlight energy to produce sugar, and vacuoles that provide storage and structural support as well as being involved in reproduction and breakdown of plant seeds. eukaryotic cells also have cytoskeleton that is made up of microtubules, intermediate filaments, and microfilaments, all of which provide support for the cell and are involved in the movement of the cell and its organelles. in terms of their structural composition, the microtubules are made up of tubulin ( e. g., Ξ± - tubulin and Ξ² - tubulin ) whereas intermediate filaments are made up of fibrous proteins. microfilaments are made up of actin molecules that interact with other strands of proteins. = = = metabolism = = = all cells require energy to sustain cellular processes. metabolism is the set of chemical reactions in an organism. the three main purposes of metabolism are : the conversion of food to energy to run cellular processes ; the conversion of food / fuel to monomer building blocks ; and the elimination of metabolic wastes. these enzyme - catalyzed reactions allow organisms to grow and reproduce, maintain their structures, and respond to their environments. metabolic reactions may be categorized as catabolic β the breaking down of compounds ( for example, the breaking down of glucose to pyruvate by cellular respiration ) ; or anabolic β the building up ( synthesis ) of compounds ( such as proteins, carbohydrates, lipids, and nucleic acids ). usually, catabolism releases energy, and anabolism consumes energy. the chemical reactions of metabolism are organized into metabolic pathways, in which
s immune system recognizes these re - implanted cells as its own, and does not target them for attack. autologous cell dependence on host cell health and donor site morbidity may be deterrents to their use. adipose - derived and bone marrow - derived mesenchymal stem cells are commonly autologous in nature, and can be used in a myriad of ways, from helping repair skeletal tissue to replenishing beta cells in diabetic patients. allogenic : cells are obtained from the body of a donor of the same species as the recipient. while there are some ethical constraints to the use of human cells for in vitro studies ( i. e. human brain tissue chimera development ), the employment of dermal fibroblasts from human foreskin demonstrates an immunologically safe and thus a viable choice for allogenic tissue engineering of the skin. xenogenic : these cells are derived isolated cells from alternate species from the recipient. a notable example of xenogeneic tissue utilization is cardiovascular implant construction via animal cells. chimeric human - animal farming raises ethical concerns around the potential for improved consciousness from implanting human organs in animals. syngeneic or isogenic : these cells describe those borne from identical genetic code. this imparts an immunologic benefit similar to autologous cell lines ( see above ). autologous cells can be considered syngenic, but the classification also extends to non - autologously derived cells such as those from an identical twin, from genetically identical ( cloned ) research models, or induced stem cells ( isc ) as related to the donor. = = = stem cells = = = stem cells are undifferentiated cells with the ability to divide in culture and give rise to different forms of specialized cells. stem cells are divided into " adult " and " embryonic " stem cells according to their source. while there is still a large ethical debate related to the use of embryonic stem cells, it is thought that another alternative source β induced pluripotent stem cells β may be useful for the repair of diseased or damaged tissues, or may be used to grow new organs. totipotent cells are stem cells which can divide into further stem cells or differentiate into any cell type in the body, including extra - embryonic tissue. pluripotent cells are stem cells which can differentiate into any cell type in the body except extra - embryonic tissue. induced pluripotent stem cells ( ipscs )
. most cells are very small, with diameters ranging from 1 to 100 micrometers and are therefore only visible under a light or electron microscope. there are generally two types of cells : eukaryotic cells, which contain a nucleus, and prokaryotic cells, which do not. prokaryotes are single - celled organisms such as bacteria, whereas eukaryotes can be single - celled or multicellular. in multicellular organisms, every cell in the organism ' s body is derived ultimately from a single cell in a fertilized egg. = = = cell structure = = = every cell is enclosed within a cell membrane that separates its cytoplasm from the extracellular space. a cell membrane consists of a lipid bilayer, including cholesterols that sit between phospholipids to maintain their fluidity at various temperatures. cell membranes are semipermeable, allowing small molecules such as oxygen, carbon dioxide, and water to pass through while restricting the movement of larger molecules and charged particles such as ions. cell membranes also contain membrane proteins, including integral membrane proteins that go across the membrane serving as membrane transporters, and peripheral proteins that loosely attach to the outer side of the cell membrane, acting as enzymes shaping the cell. cell membranes are involved in various cellular processes such as cell adhesion, storing electrical energy, and cell signalling and serve as the attachment surface for several extracellular structures such as a cell wall, glycocalyx, and cytoskeleton. within the cytoplasm of a cell, there are many biomolecules such as proteins and nucleic acids. in addition to biomolecules, eukaryotic cells have specialized structures called organelles that have their own lipid bilayers or are spatially units. these organelles include the cell nucleus, which contains most of the cell ' s dna, or mitochondria, which generate adenosine triphosphate ( atp ) to power cellular processes. other organelles such as endoplasmic reticulum and golgi apparatus play a role in the synthesis and packaging of proteins, respectively. biomolecules such as proteins can be engulfed by lysosomes, another specialized organelle. plant cells have additional organelles that distinguish them from animal cells such as a cell wall that provides support for the plant cell, chloroplasts that harvest sunlight energy to produce sugar, and vacuoles that provide storage and structural support
Question: Which of the following allows a cell to become specialized?
A) location of the cell in the body
B) function of the cell in the body
C) different patterns of gene expression
D) different combinations of genes in the nucleus
|
C) different patterns of gene expression
|
Context:
and measuring radiation levels. the surveyor program conducted uncrewed lunar landings and takeoffs, as well as taking surface and regolith observations. despite the setback caused by the apollo 1 fire, which killed three astronauts, the program proceeded. apollo 8 was the first crewed spacecraft to leave low earth orbit and the first human spaceflight to reach the moon. the crew orbited the moon ten times on december 24 and 25, 1968, and then traveled safely back to earth. the three apollo 8 astronauts β frank borman, james lovell, and william anders β were the first humans to see the earth as a globe in space, the first to witness an earthrise, and the first to see and manually photograph the far side of the moon. the first lunar landing was conducted by apollo 11. commanded by neil armstrong with astronauts buzz aldrin and michael collins, apollo 11 was one of the most significant missions in nasa ' s history, marking the end of the space race when the soviet union gave up its lunar ambitions. as the first human to step on the surface of the moon, neil armstrong uttered the now famous words : that ' s one small step for man, one giant leap for mankind. nasa would conduct six total lunar landings as part of the apollo program, with apollo 17 concluding the program in 1972. = = = = end of apollo = = = = wernher von braun had advocated for nasa to develop a space station since the agency was created. in 1973, following the end of the apollo lunar missions, nasa launched its first space station, skylab, on the final launch of the saturn v. skylab reused a significant amount of apollo and saturn hardware, with a repurposed saturn v third stage serving as the primary module for the space station. damage to skylab during its launch required spacewalks to be performed by the first crew to make it habitable and operational. skylab hosted nine missions and was decommissioned in 1974 and deorbited in 1979, two years prior to the first launch of the space shuttle and any possibility of boosting its orbit. in 1975, the apollo β soyuz mission was the first ever international spaceflight and a major diplomatic accomplishment between the cold war rivals, which also marked the last flight of the apollo capsule. flown in 1975, a us apollo spacecraft docked with a soviet soyuz capsule. = = = interplanetary exploration and space science = = = during the 1960s, nasa started its space science and interplanetary probe program. the mariner program was its flagship
fuel cells instead of batteries, and conducted the first american spacewalks and rendezvous operations. the ranger program was started in the 1950s as a response to soviet lunar exploration, however most missions ended in failure. the lunar orbiter program had greater success, mapping the surface in preparation for apollo landings, conducting meteoroid detection, and measuring radiation levels. the surveyor program conducted uncrewed lunar landings and takeoffs, as well as taking surface and regolith observations. despite the setback caused by the apollo 1 fire, which killed three astronauts, the program proceeded. apollo 8 was the first crewed spacecraft to leave low earth orbit and the first human spaceflight to reach the moon. the crew orbited the moon ten times on december 24 and 25, 1968, and then traveled safely back to earth. the three apollo 8 astronauts β frank borman, james lovell, and william anders β were the first humans to see the earth as a globe in space, the first to witness an earthrise, and the first to see and manually photograph the far side of the moon. the first lunar landing was conducted by apollo 11. commanded by neil armstrong with astronauts buzz aldrin and michael collins, apollo 11 was one of the most significant missions in nasa ' s history, marking the end of the space race when the soviet union gave up its lunar ambitions. as the first human to step on the surface of the moon, neil armstrong uttered the now famous words : that ' s one small step for man, one giant leap for mankind. nasa would conduct six total lunar landings as part of the apollo program, with apollo 17 concluding the program in 1972. = = = = end of apollo = = = = wernher von braun had advocated for nasa to develop a space station since the agency was created. in 1973, following the end of the apollo lunar missions, nasa launched its first space station, skylab, on the final launch of the saturn v. skylab reused a significant amount of apollo and saturn hardware, with a repurposed saturn v third stage serving as the primary module for the space station. damage to skylab during its launch required spacewalks to be performed by the first crew to make it habitable and operational. skylab hosted nine missions and was decommissioned in 1974 and deorbited in 1979, two years prior to the first launch of the space shuttle and any possibility of boosting its orbit. in 1975, the apollo β soyuz mission was the first ever international spaceflight and a major diplomatic accomplishment between the cold war
three of what is called the six simple machines, from which all machines are based. these machines are the inclined plane, the wedge, and the lever, which allowed the ancient egyptians to move millions of limestone blocks which weighed approximately 3. 5 tons ( 7, 000 lbs. ) each into place to create structures like the great pyramid of giza, which is 481 feet ( 147 meters ) high. they also made writing medium similar to paper from papyrus, which joshua mark states is the foundation for modern paper. papyrus is a plant ( cyperus papyrus ) which grew in plentiful amounts in the egyptian delta and throughout the nile river valley during ancient times. the papyrus was harvested by field workers and brought to processing centers where it was cut into thin strips. the strips were then laid - out side by side and covered in plant resin. the second layer of strips was laid on perpendicularly, then both pressed together until the sheet was dry. the sheets were then joined to form a roll and later used for writing. egyptian society made several significant advances during dynastic periods in many areas of technology. according to hossam elanzeery, they were the first civilization to use timekeeping devices such as sundials, shadow clocks, and obelisks and successfully leveraged their knowledge of astronomy to create a calendar model that society still uses today. they developed shipbuilding technology that saw them progress from papyrus reed vessels to cedar wood ships while also pioneering the use of rope trusses and stem - mounted rudders. the egyptians also used their knowledge of anatomy to lay the foundation for many modern medical techniques and practiced the earliest known version of neuroscience. elanzeery also states that they used and furthered mathematical science, as evidenced in the building of the pyramids. ancient egyptians also invented and pioneered many food technologies that have become the basis of modern food technology processes. based on paintings and reliefs found in tombs, as well as archaeological artifacts, scholars like paul t nicholson believe that the ancient egyptians established systematic farming practices, engaged in cereal processing, brewed beer and baked bread, processed meat, practiced viticulture and created the basis for modern wine production, and created condiments to complement, preserve and mask the flavors of their food. = = = = indus valley = = = = the indus valley civilization, situated in a resource - rich area ( in modern pakistan and northwestern india ), is notable for its early application of city planning, sanitation technologies, and plumbing. indus valley construction and architecture, called ' vaastu
in supersymmetric models the mass of the stops can be considered as the naturalness measure of the theory. roughly, the lighter the stops are, the more natural the theory is. both, the absence of supersymmetric signals at experiment and the measurement of the higgs mass, put scenarios with light stops under increasing tension. i will present a supersymmetry breaking mechanism of the scherk - schwarz type that, by introducing extra $ su ( 2 ) _ l $ triplets in the higgs sector, is able to generate the correct higgs mass while keeping stops light.
ultra high energy particles arrive at earth constantly. they provide a beam at energies higher than any man - made accelerator, but at a very low rate. two large experiments, the pierre auger observatory and the telescope array experiment, have been taking data for several years now covering together the whole sky. i summarize the most recent measurements from both experiments, i compare their results and, for a change, i highlight their agreements.
aviation, while the apollo lunar module was designed and built by grumman. to develop the spaceflight skills and equipment required for a lunar mission, nasa initiated project gemini. using a modified air force titan ii launch vehicle, the gemini capsule could hold two astronauts for flights of over two weeks. gemini pioneered the use of fuel cells instead of batteries, and conducted the first american spacewalks and rendezvous operations. the ranger program was started in the 1950s as a response to soviet lunar exploration, however most missions ended in failure. the lunar orbiter program had greater success, mapping the surface in preparation for apollo landings, conducting meteoroid detection, and measuring radiation levels. the surveyor program conducted uncrewed lunar landings and takeoffs, as well as taking surface and regolith observations. despite the setback caused by the apollo 1 fire, which killed three astronauts, the program proceeded. apollo 8 was the first crewed spacecraft to leave low earth orbit and the first human spaceflight to reach the moon. the crew orbited the moon ten times on december 24 and 25, 1968, and then traveled safely back to earth. the three apollo 8 astronauts β frank borman, james lovell, and william anders β were the first humans to see the earth as a globe in space, the first to witness an earthrise, and the first to see and manually photograph the far side of the moon. the first lunar landing was conducted by apollo 11. commanded by neil armstrong with astronauts buzz aldrin and michael collins, apollo 11 was one of the most significant missions in nasa ' s history, marking the end of the space race when the soviet union gave up its lunar ambitions. as the first human to step on the surface of the moon, neil armstrong uttered the now famous words : that ' s one small step for man, one giant leap for mankind. nasa would conduct six total lunar landings as part of the apollo program, with apollo 17 concluding the program in 1972. = = = = end of apollo = = = = wernher von braun had advocated for nasa to develop a space station since the agency was created. in 1973, following the end of the apollo lunar missions, nasa launched its first space station, skylab, on the final launch of the saturn v. skylab reused a significant amount of apollo and saturn hardware, with a repurposed saturn v third stage serving as the primary module for the space station. damage to skylab during its launch required spacewalks to be performed by the first crew to make it habitable and
##angulation from bearings taken by two rdf stations separated geographically, as the point where the two bearing lines cross, this is called a " fix ". military forces use rdf to locate enemy forces by their tactical radio transmissions, counterintelligence services use it to locate clandestine transmitters used by espionage agents, and governments use it to locate unlicensed transmitters or interference sources. older rdf receivers used rotatable loop antennas, the antenna is rotated until the radio signal strength is weakest, indicating the transmitter is in one of the antenna ' s two nulls. the nulls are used since they are sharper than the antenna ' s lobes ( maxima ). more modern receivers use phased array antennas which have a much greater angular resolution. animal migration tracking β a widely used technique in wildlife biology, conservation biology, and wildlife management in which small battery - powered radio transmitters are attached to wild animals so their movements can be tracked with a directional rdf receiver. sometimes the transmitter is implanted in the animal. the vhf band is typically used since antennas in this band are fairly compact. the receiver has a directional antenna ( typically a small yagi ) which is rotated until the received signal is strongest ; at this point the antenna is pointing in the direction of the animal. sophisticated systems used in recent years use satellites to track the animal, or geolocation tags with gps receivers which record and transmit a log of the animal ' s location. = = = = remote control = = = = radio remote control is the use of electronic control signals sent by radio waves from a transmitter to control the actions of a device at a remote location. remote control systems may also include telemetry channels in the other direction, used to transmit real - time information on the state of the device back to the control station. uncrewed spacecraft are an example of remote - controlled machines, controlled by commands transmitted by satellite ground stations. most handheld remote controls used to control consumer electronics products like televisions or dvd players actually operate by infrared light rather than radio waves, so are not examples of radio remote control. a security concern with remote control systems is spoofing, in which an unauthorized person transmits an imitation of the control signal to take control of the device. examples of radio remote control : unmanned aerial vehicle ( uav, drone ) β a drone is an aircraft without an onboard pilot, flown by remote control by a pilot in another location, usually in a piloting station on the ground. they are used by the military for reconnaissance and ground attack, and
the space station is established in intergovernmental treaties and agreements, which divide the station into two areas and allow russia to retain full ownership of the russian orbital segment ( with the exception of zarya ), with the us orbital segment allocated between the other international partners. long - duration missions to the iss are referred to as iss expeditions. expedition crew members typically spend approximately six months on the iss. the initial expedition crew size was three, temporarily decreased to two following the columbia disaster. between may 2009 and until the retirement of the space shuttle, the expedition crew size has been six crew members. as of 2024, though the commercial program ' s crew capsules can allow a crew of up to seven, expeditions using them typically consist of a crew of four. the iss has been continuously occupied for the past 24 years and 202 days, having exceeded the previous record held by mir ; and has been visited by astronauts and cosmonauts from 15 different nations. the station can be seen from the earth with the naked eye and, as of 2025, is the largest artificial satellite in earth orbit with a mass and volume greater than that of any previous space station. the russian soyuz and american dragon and starliner spacecraft are used to send astronauts to and from the iss. several uncrewed cargo spacecraft provide service to the iss ; they are the russian progress spacecraft which has done so since 2000, the european automated transfer vehicle ( atv ) since 2008, the japanese h - ii transfer vehicle ( htv ) since 2009, the ( uncrewed ) dragon since 2012, and the american cygnus spacecraft since 2013. the space shuttle, before its retirement, was also used for cargo transfer and would often switch out expedition crew members, although it did not have the capability to remain docked for the duration of their stay. between the retirement of the shuttle in 2011 and the commencement of crewed dragon flights in 2020, american astronauts exclusively used the soyuz for crew transport to and from the iss the highest number of people occupying the iss has been thirteen ; this occurred three times during the late shuttle iss assembly missions. the iss program is expected to continue until 2030, after which the space station will be retired and destroyed in a controlled de - orbit. = = = = commercial resupply services ( 2008 β present ) = = = = commercial resupply services ( crs ) are a contract solution to deliver cargo and supplies to the international space station on a commercial basis by private companies. nasa signed its first crs contracts in 2008 and awarded $ 1
and other components, which were manufactured in various factories around the world and launched by russian proton and soyuz rockets, and the american space shuttle. the on - orbit assembly began in 1998, the completion of the us orbital segment occurred in 2009 and the completion of the russian orbital segment occurred in 2010. the ownership and use of the space station is established in intergovernmental treaties and agreements, which divide the station into two areas and allow russia to retain full ownership of the russian orbital segment ( with the exception of zarya ), with the us orbital segment allocated between the other international partners. long - duration missions to the iss are referred to as iss expeditions. expedition crew members typically spend approximately six months on the iss. the initial expedition crew size was three, temporarily decreased to two following the columbia disaster. between may 2009 and until the retirement of the space shuttle, the expedition crew size has been six crew members. as of 2024, though the commercial program ' s crew capsules can allow a crew of up to seven, expeditions using them typically consist of a crew of four. the iss has been continuously occupied for the past 24 years and 202 days, having exceeded the previous record held by mir ; and has been visited by astronauts and cosmonauts from 15 different nations. the station can be seen from the earth with the naked eye and, as of 2025, is the largest artificial satellite in earth orbit with a mass and volume greater than that of any previous space station. the russian soyuz and american dragon and starliner spacecraft are used to send astronauts to and from the iss. several uncrewed cargo spacecraft provide service to the iss ; they are the russian progress spacecraft which has done so since 2000, the european automated transfer vehicle ( atv ) since 2008, the japanese h - ii transfer vehicle ( htv ) since 2009, the ( uncrewed ) dragon since 2012, and the american cygnus spacecraft since 2013. the space shuttle, before its retirement, was also used for cargo transfer and would often switch out expedition crew members, although it did not have the capability to remain docked for the duration of their stay. between the retirement of the shuttle in 2011 and the commencement of crewed dragon flights in 2020, american astronauts exclusively used the soyuz for crew transport to and from the iss the highest number of people occupying the iss has been thirteen ; this occurred three times during the late shuttle iss assembly missions. the iss program is expected to continue until 2030, after which the space station will be retired and destroyed in a controlled de -
as iss expeditions. expedition crew members typically spend approximately six months on the iss. the initial expedition crew size was three, temporarily decreased to two following the columbia disaster. between may 2009 and until the retirement of the space shuttle, the expedition crew size has been six crew members. as of 2024, though the commercial program ' s crew capsules can allow a crew of up to seven, expeditions using them typically consist of a crew of four. the iss has been continuously occupied for the past 24 years and 202 days, having exceeded the previous record held by mir ; and has been visited by astronauts and cosmonauts from 15 different nations. the station can be seen from the earth with the naked eye and, as of 2025, is the largest artificial satellite in earth orbit with a mass and volume greater than that of any previous space station. the russian soyuz and american dragon and starliner spacecraft are used to send astronauts to and from the iss. several uncrewed cargo spacecraft provide service to the iss ; they are the russian progress spacecraft which has done so since 2000, the european automated transfer vehicle ( atv ) since 2008, the japanese h - ii transfer vehicle ( htv ) since 2009, the ( uncrewed ) dragon since 2012, and the american cygnus spacecraft since 2013. the space shuttle, before its retirement, was also used for cargo transfer and would often switch out expedition crew members, although it did not have the capability to remain docked for the duration of their stay. between the retirement of the shuttle in 2011 and the commencement of crewed dragon flights in 2020, american astronauts exclusively used the soyuz for crew transport to and from the iss the highest number of people occupying the iss has been thirteen ; this occurred three times during the late shuttle iss assembly missions. the iss program is expected to continue until 2030, after which the space station will be retired and destroyed in a controlled de - orbit. = = = = commercial resupply services ( 2008 β present ) = = = = commercial resupply services ( crs ) are a contract solution to deliver cargo and supplies to the international space station on a commercial basis by private companies. nasa signed its first crs contracts in 2008 and awarded $ 1. 6 billion to spacex for twelve cargo dragon and $ 1. 9 billion to orbital sciences for eight cygnus flights, covering deliveries until 2016. both companies evolved or created their launch vehicle products to launch the spacecrafts ( spacex with the falcon 9 and orbital with the antares ). spacex flew its
Question: Which tool would be best to use for counting the number of legs on an ant?
A) ruler
B) hand lens
C) calculator
D) microscope
|
B) hand lens
|
Context:
( division of the nucleus ) is preceded by the s stage of interphase ( during which the dna is replicated ) and is often followed by telophase and cytokinesis ; which divides the cytoplasm, organelles and cell membrane of one cell into two new cells containing roughly equal shares of these cellular components. the different stages of mitosis all together define the mitotic phase of an animal cell cycle β the division of the mother cell into two genetically identical daughter cells. the cell cycle is a vital process by which a single - celled fertilized egg develops into a mature organism, as well as the process by which hair, skin, blood cells, and some internal organs are renewed. after cell division, each of the daughter cells begin the interphase of a new cycle. in contrast to mitosis, meiosis results in four haploid daughter cells by undergoing one round of dna replication followed by two divisions. homologous chromosomes are separated in the first division ( meiosis i ), and sister chromatids are separated in the second division ( meiosis ii ). both of these cell division cycles are used in the process of sexual reproduction at some point in their life cycle. both are believed to be present in the last eukaryotic common ancestor. prokaryotes ( i. e., archaea and bacteria ) can also undergo cell division ( or binary fission ). unlike the processes of mitosis and meiosis in eukaryotes, binary fission in prokaryotes takes place without the formation of a spindle apparatus on the cell. before binary fission, dna in the bacterium is tightly coiled. after it has uncoiled and duplicated, it is pulled to the separate poles of the bacterium as it increases the size to prepare for splitting. growth of a new cell wall begins to separate the bacterium ( triggered by ftsz polymerization and " z - ring " formation ). the new cell wall ( septum ) fully develops, resulting in the complete split of the bacterium. the new daughter cells have tightly coiled dna rods, ribosomes, and plasmids. = = = sexual reproduction and meiosis = = = meiosis is a central feature of sexual reproduction in eukaryotes, and the most fundamental function of meiosis appears to be conservation of the integrity of the genome that is passed on to progeny by parents. two aspects of sexual reproduction, meiotic recombination and outcrossing, are likely maintained respectively by
of these cellular components. the different stages of mitosis all together define the mitotic phase of an animal cell cycle β the division of the mother cell into two genetically identical daughter cells. the cell cycle is a vital process by which a single - celled fertilized egg develops into a mature organism, as well as the process by which hair, skin, blood cells, and some internal organs are renewed. after cell division, each of the daughter cells begin the interphase of a new cycle. in contrast to mitosis, meiosis results in four haploid daughter cells by undergoing one round of dna replication followed by two divisions. homologous chromosomes are separated in the first division ( meiosis i ), and sister chromatids are separated in the second division ( meiosis ii ). both of these cell division cycles are used in the process of sexual reproduction at some point in their life cycle. both are believed to be present in the last eukaryotic common ancestor. prokaryotes ( i. e., archaea and bacteria ) can also undergo cell division ( or binary fission ). unlike the processes of mitosis and meiosis in eukaryotes, binary fission in prokaryotes takes place without the formation of a spindle apparatus on the cell. before binary fission, dna in the bacterium is tightly coiled. after it has uncoiled and duplicated, it is pulled to the separate poles of the bacterium as it increases the size to prepare for splitting. growth of a new cell wall begins to separate the bacterium ( triggered by ftsz polymerization and " z - ring " formation ). the new cell wall ( septum ) fully develops, resulting in the complete split of the bacterium. the new daughter cells have tightly coiled dna rods, ribosomes, and plasmids. = = = sexual reproduction and meiosis = = = meiosis is a central feature of sexual reproduction in eukaryotes, and the most fundamental function of meiosis appears to be conservation of the integrity of the genome that is passed on to progeny by parents. two aspects of sexual reproduction, meiotic recombination and outcrossing, are likely maintained respectively by the adaptive advantages of recombinational repair of genomic dna damage and genetic complementation which masks the expression of deleterious recessive mutations. the beneficial effect of genetic complementation, derived from outcrossing ( cross - fertilization ) is also referred to as hybrid vigor or heterosis. charles
or removed from the dna during programmed stages of development of the plant, and are responsible, for example, for the differences between anthers, petals and normal leaves, despite the fact that they all have the same underlying genetic code. epigenetic changes may be temporary or may remain through successive cell divisions for the remainder of the cell ' s life. some epigenetic changes have been shown to be heritable, while others are reset in the germ cells. epigenetic changes in eukaryotic biology serve to regulate the process of cellular differentiation. during morphogenesis, totipotent stem cells become the various pluripotent cell lines of the embryo, which in turn become fully differentiated cells. a single fertilised egg cell, the zygote, gives rise to the many different plant cell types including parenchyma, xylem vessel elements, phloem sieve tubes, guard cells of the epidermis, etc. as it continues to divide. the process results from the epigenetic activation of some genes and inhibition of others. unlike animals, many plant cells, particularly those of the parenchyma, do not terminally differentiate, remaining totipotent with the ability to give rise to a new individual plant. exceptions include highly lignified cells, the sclerenchyma and xylem which are dead at maturity, and the phloem sieve tubes which lack nuclei. while plants use many of the same epigenetic mechanisms as animals, such as chromatin remodelling, an alternative hypothesis is that plants set their gene expression patterns using positional information from the environment and surrounding cells to determine their developmental fate. epigenetic changes can lead to paramutations, which do not follow the mendelian heritage rules. these epigenetic marks are carried from one generation to the next, with one allele inducing a change on the other. = = plant evolution = = the chloroplasts of plants have a number of biochemical, structural and genetic similarities to cyanobacteria, ( commonly but incorrectly known as " blue - green algae " ) and are thought to be derived from an ancient endosymbiotic relationship between an ancestral eukaryotic cell and a cyanobacterial resident. the algae are a polyphyletic group and are placed in various divisions, some more closely related to plants than others. there are many differences between them in features such as cell wall composition, biochemistry,
it to divide into two daughter cells. these events include the duplication of its dna and some of its organelles, and the subsequent partitioning of its cytoplasm into two daughter cells in a process called cell division. in eukaryotes ( i. e., animal, plant, fungal, and protist cells ), there are two distinct types of cell division : mitosis and meiosis. mitosis is part of the cell cycle, in which replicated chromosomes are separated into two new nuclei. cell division gives rise to genetically identical cells in which the total number of chromosomes is maintained. in general, mitosis ( division of the nucleus ) is preceded by the s stage of interphase ( during which the dna is replicated ) and is often followed by telophase and cytokinesis ; which divides the cytoplasm, organelles and cell membrane of one cell into two new cells containing roughly equal shares of these cellular components. the different stages of mitosis all together define the mitotic phase of an animal cell cycle β the division of the mother cell into two genetically identical daughter cells. the cell cycle is a vital process by which a single - celled fertilized egg develops into a mature organism, as well as the process by which hair, skin, blood cells, and some internal organs are renewed. after cell division, each of the daughter cells begin the interphase of a new cycle. in contrast to mitosis, meiosis results in four haploid daughter cells by undergoing one round of dna replication followed by two divisions. homologous chromosomes are separated in the first division ( meiosis i ), and sister chromatids are separated in the second division ( meiosis ii ). both of these cell division cycles are used in the process of sexual reproduction at some point in their life cycle. both are believed to be present in the last eukaryotic common ancestor. prokaryotes ( i. e., archaea and bacteria ) can also undergo cell division ( or binary fission ). unlike the processes of mitosis and meiosis in eukaryotes, binary fission in prokaryotes takes place without the formation of a spindle apparatus on the cell. before binary fission, dna in the bacterium is tightly coiled. after it has uncoiled and duplicated, it is pulled to the separate poles of the bacterium as it increases the size to prepare for splitting. growth of a new cell wall begins to separate the bacterium ( triggered by ft
protist cells ), there are two distinct types of cell division : mitosis and meiosis. mitosis is part of the cell cycle, in which replicated chromosomes are separated into two new nuclei. cell division gives rise to genetically identical cells in which the total number of chromosomes is maintained. in general, mitosis ( division of the nucleus ) is preceded by the s stage of interphase ( during which the dna is replicated ) and is often followed by telophase and cytokinesis ; which divides the cytoplasm, organelles and cell membrane of one cell into two new cells containing roughly equal shares of these cellular components. the different stages of mitosis all together define the mitotic phase of an animal cell cycle β the division of the mother cell into two genetically identical daughter cells. the cell cycle is a vital process by which a single - celled fertilized egg develops into a mature organism, as well as the process by which hair, skin, blood cells, and some internal organs are renewed. after cell division, each of the daughter cells begin the interphase of a new cycle. in contrast to mitosis, meiosis results in four haploid daughter cells by undergoing one round of dna replication followed by two divisions. homologous chromosomes are separated in the first division ( meiosis i ), and sister chromatids are separated in the second division ( meiosis ii ). both of these cell division cycles are used in the process of sexual reproduction at some point in their life cycle. both are believed to be present in the last eukaryotic common ancestor. prokaryotes ( i. e., archaea and bacteria ) can also undergo cell division ( or binary fission ). unlike the processes of mitosis and meiosis in eukaryotes, binary fission in prokaryotes takes place without the formation of a spindle apparatus on the cell. before binary fission, dna in the bacterium is tightly coiled. after it has uncoiled and duplicated, it is pulled to the separate poles of the bacterium as it increases the size to prepare for splitting. growth of a new cell wall begins to separate the bacterium ( triggered by ftsz polymerization and " z - ring " formation ). the new cell wall ( septum ) fully develops, resulting in the complete split of the bacterium. the new daughter cells have tightly coiled dna rods, ribosomes, and plasmids. = = = sexual reproduction and meiosis = = = mei
the process by which hair, skin, blood cells, and some internal organs are renewed. after cell division, each of the daughter cells begin the interphase of a new cycle. in contrast to mitosis, meiosis results in four haploid daughter cells by undergoing one round of dna replication followed by two divisions. homologous chromosomes are separated in the first division ( meiosis i ), and sister chromatids are separated in the second division ( meiosis ii ). both of these cell division cycles are used in the process of sexual reproduction at some point in their life cycle. both are believed to be present in the last eukaryotic common ancestor. prokaryotes ( i. e., archaea and bacteria ) can also undergo cell division ( or binary fission ). unlike the processes of mitosis and meiosis in eukaryotes, binary fission in prokaryotes takes place without the formation of a spindle apparatus on the cell. before binary fission, dna in the bacterium is tightly coiled. after it has uncoiled and duplicated, it is pulled to the separate poles of the bacterium as it increases the size to prepare for splitting. growth of a new cell wall begins to separate the bacterium ( triggered by ftsz polymerization and " z - ring " formation ). the new cell wall ( septum ) fully develops, resulting in the complete split of the bacterium. the new daughter cells have tightly coiled dna rods, ribosomes, and plasmids. = = = sexual reproduction and meiosis = = = meiosis is a central feature of sexual reproduction in eukaryotes, and the most fundamental function of meiosis appears to be conservation of the integrity of the genome that is passed on to progeny by parents. two aspects of sexual reproduction, meiotic recombination and outcrossing, are likely maintained respectively by the adaptive advantages of recombinational repair of genomic dna damage and genetic complementation which masks the expression of deleterious recessive mutations. the beneficial effect of genetic complementation, derived from outcrossing ( cross - fertilization ) is also referred to as hybrid vigor or heterosis. charles darwin in his 1878 book the effects of cross and self - fertilization in the vegetable kingdom at the start of chapter xii noted β the first and most important of the conclusions which may be drawn from the observations given in this volume, is that generally cross - fertilisation is beneficial and self - fertilis
and the creation of genetically modified crops. = = = epigenetics = = = epigenetics is the study of heritable changes in gene function that cannot be explained by changes in the underlying dna sequence but cause the organism ' s genes to behave ( or " express themselves " ) differently. one example of epigenetic change is the marking of the genes by dna methylation which determines whether they will be expressed or not. gene expression can also be controlled by repressor proteins that attach to silencer regions of the dna and prevent that region of the dna code from being expressed. epigenetic marks may be added or removed from the dna during programmed stages of development of the plant, and are responsible, for example, for the differences between anthers, petals and normal leaves, despite the fact that they all have the same underlying genetic code. epigenetic changes may be temporary or may remain through successive cell divisions for the remainder of the cell ' s life. some epigenetic changes have been shown to be heritable, while others are reset in the germ cells. epigenetic changes in eukaryotic biology serve to regulate the process of cellular differentiation. during morphogenesis, totipotent stem cells become the various pluripotent cell lines of the embryo, which in turn become fully differentiated cells. a single fertilised egg cell, the zygote, gives rise to the many different plant cell types including parenchyma, xylem vessel elements, phloem sieve tubes, guard cells of the epidermis, etc. as it continues to divide. the process results from the epigenetic activation of some genes and inhibition of others. unlike animals, many plant cells, particularly those of the parenchyma, do not terminally differentiate, remaining totipotent with the ability to give rise to a new individual plant. exceptions include highly lignified cells, the sclerenchyma and xylem which are dead at maturity, and the phloem sieve tubes which lack nuclei. while plants use many of the same epigenetic mechanisms as animals, such as chromatin remodelling, an alternative hypothesis is that plants set their gene expression patterns using positional information from the environment and surrounding cells to determine their developmental fate. epigenetic changes can lead to paramutations, which do not follow the mendelian heritage rules. these epigenetic marks are carried from one generation to the next,
organelles. in terms of their structural composition, the microtubules are made up of tubulin ( e. g., Ξ± - tubulin and Ξ² - tubulin ) whereas intermediate filaments are made up of fibrous proteins. microfilaments are made up of actin molecules that interact with other strands of proteins. = = = metabolism = = = all cells require energy to sustain cellular processes. metabolism is the set of chemical reactions in an organism. the three main purposes of metabolism are : the conversion of food to energy to run cellular processes ; the conversion of food / fuel to monomer building blocks ; and the elimination of metabolic wastes. these enzyme - catalyzed reactions allow organisms to grow and reproduce, maintain their structures, and respond to their environments. metabolic reactions may be categorized as catabolic β the breaking down of compounds ( for example, the breaking down of glucose to pyruvate by cellular respiration ) ; or anabolic β the building up ( synthesis ) of compounds ( such as proteins, carbohydrates, lipids, and nucleic acids ). usually, catabolism releases energy, and anabolism consumes energy. the chemical reactions of metabolism are organized into metabolic pathways, in which one chemical is transformed through a series of steps into another chemical, each step being facilitated by a specific enzyme. enzymes are crucial to metabolism because they allow organisms to drive desirable reactions that require energy that will not occur by themselves, by coupling them to spontaneous reactions that release energy. enzymes act as catalysts β they allow a reaction to proceed more rapidly without being consumed by it β by reducing the amount of activation energy needed to convert reactants into products. enzymes also allow the regulation of the rate of a metabolic reaction, for example in response to changes in the cell ' s environment or to signals from other cells. = = = cellular respiration = = = cellular respiration is a set of metabolic reactions and processes that take place in cells to convert chemical energy from nutrients into adenosine triphosphate ( atp ), and then release waste products. the reactions involved in respiration are catabolic reactions, which break large molecules into smaller ones, releasing energy. respiration is one of the key ways a cell releases chemical energy to fuel cellular activity. the overall reaction occurs in a series of biochemical steps, some of which are redox reactions. although cellular respiration is technically a combustion reaction, it clearly does not resemble one when it occurs in a cell because of the
the cell ' s life. some epigenetic changes have been shown to be heritable, while others are reset in the germ cells. epigenetic changes in eukaryotic biology serve to regulate the process of cellular differentiation. during morphogenesis, totipotent stem cells become the various pluripotent cell lines of the embryo, which in turn become fully differentiated cells. a single fertilised egg cell, the zygote, gives rise to the many different plant cell types including parenchyma, xylem vessel elements, phloem sieve tubes, guard cells of the epidermis, etc. as it continues to divide. the process results from the epigenetic activation of some genes and inhibition of others. unlike animals, many plant cells, particularly those of the parenchyma, do not terminally differentiate, remaining totipotent with the ability to give rise to a new individual plant. exceptions include highly lignified cells, the sclerenchyma and xylem which are dead at maturity, and the phloem sieve tubes which lack nuclei. while plants use many of the same epigenetic mechanisms as animals, such as chromatin remodelling, an alternative hypothesis is that plants set their gene expression patterns using positional information from the environment and surrounding cells to determine their developmental fate. epigenetic changes can lead to paramutations, which do not follow the mendelian heritage rules. these epigenetic marks are carried from one generation to the next, with one allele inducing a change on the other. = = plant evolution = = the chloroplasts of plants have a number of biochemical, structural and genetic similarities to cyanobacteria, ( commonly but incorrectly known as " blue - green algae " ) and are thought to be derived from an ancient endosymbiotic relationship between an ancestral eukaryotic cell and a cyanobacterial resident. the algae are a polyphyletic group and are placed in various divisions, some more closely related to plants than others. there are many differences between them in features such as cell wall composition, biochemistry, pigmentation, chloroplast structure and nutrient reserves. the algal division charophyta, sister to the green algal division chlorophyta, is considered to contain the ancestor of true plants. the charophyte class charophyceae and the land plant sub - kingdom embryophy
pigmentation, chloroplast structure and nutrient reserves. the algal division charophyta, sister to the green algal division chlorophyta, is considered to contain the ancestor of true plants. the charophyte class charophyceae and the land plant sub - kingdom embryophyta together form the monophyletic group or clade streptophytina. nonvascular land plants are embryophytes that lack the vascular tissues xylem and phloem. they include mosses, liverworts and hornworts. pteridophytic vascular plants with true xylem and phloem that reproduced by spores germinating into free - living gametophytes evolved during the silurian period and diversified into several lineages during the late silurian and early devonian. representatives of the lycopods have survived to the present day. by the end of the devonian period, several groups, including the lycopods, sphenophylls and progymnosperms, had independently evolved " megaspory " β their spores were of two distinct sizes, larger megaspores and smaller microspores. their reduced gametophytes developed from megaspores retained within the spore - producing organs ( megasporangia ) of the sporophyte, a condition known as endospory. seeds consist of an endosporic megasporangium surrounded by one or two sheathing layers ( integuments ). the young sporophyte develops within the seed, which on germination splits to release it. the earliest known seed plants date from the latest devonian famennian stage. following the evolution of the seed habit, seed plants diversified, giving rise to a number of now - extinct groups, including seed ferns, as well as the modern gymnosperms and angiosperms. gymnosperms produce " naked seeds " not fully enclosed in an ovary ; modern representatives include conifers, cycads, ginkgo, and gnetales. angiosperms produce seeds enclosed in a structure such as a carpel or an ovary. ongoing research on the molecular phylogenetics of living plants appears to show that the angiosperms are a sister clade to the gymnosperms. = = plant physiology = = plant physiology encompasses all the internal chemical and physical activities of plants associated with life. chemicals obtained from the air, soil and water form
Question: A dragonfly develops through incomplete metamorphosis. Which stages does the dragonfly's life cycle include?
A) egg and larva
B) egg and nymph
C) larva and pupa
D) larva and nymph
|
B) egg and nymph
|
Context:
in one or more of these kinds of structures, it is invariably accompanied by an increase or decrease of energy of the substances involved. some energy is transferred between the surroundings and the reactants of the reaction in the form of heat or light ; thus the products of a reaction may have more or less energy than the reactants. a reaction is said to be exergonic if the final state is lower on the energy scale than the initial state ; in the case of endergonic reactions the situation is the reverse. a reaction is said to be exothermic if the reaction releases heat to the surroundings ; in the case of endothermic reactions, the reaction absorbs heat from the surroundings. chemical reactions are invariably not possible unless the reactants surmount an energy barrier known as the activation energy. the speed of a chemical reaction ( at given temperature t ) is related to the activation energy e, by the boltzmann ' s population factor e β e / k t { \ displaystyle e ^ { - e / kt } } β that is the probability of a molecule to have energy greater than or equal to e at the given temperature t. this exponential dependence of a reaction rate on temperature is known as the arrhenius equation. the activation energy necessary for a chemical reaction to occur can be in the form of heat, light, electricity or mechanical force in the form of ultrasound. a related concept free energy, which also incorporates entropy considerations, is a very useful means for predicting the feasibility of a reaction and determining the state of equilibrium of a chemical reaction, in chemical thermodynamics. a reaction is feasible only if the total change in the gibbs free energy is negative, Ξ΄ g β€ 0 { \ displaystyle \ delta g \ leq 0 \, } ; if it is equal to zero the chemical reaction is said to be at equilibrium. there exist only limited possible states of energy for electrons, atoms and molecules. these are determined by the rules of quantum mechanics, which require quantization of energy of a bound system. the atoms / molecules in a higher energy state are said to be excited. the molecules / atoms of substance in an excited energy state are often much more reactive ; that is, more amenable to chemical reactions. the phase of a substance is invariably determined by its energy and the energy of its surroundings. when the intermolecular forces of a substance are such that the energy of the surroundings is not sufficient to overcome them, it occurs in a more ordered phase like liquid
ambient air ( see lockheed f - 117 nighthawk, rectangular nozzles on the lockheed martin f - 22 raptor, and serrated nozzle flaps on the lockheed martin f - 35 lightning ). often, cool air is deliberately injected into the exhaust flow to boost this process ( see ryan aqm - 91 firefly and northrop b - 2 spirit ). the stefan β boltzmann law shows how this results in less energy ( thermal radiation in infrared spectrum ) being released and thus reduces the heat signature. in some aircraft, the jet exhaust is vented above the wing surface to shield it from observers below, as in the lockheed f - 117 nighthawk, and the unstealthy fairchild republic a - 10 thunderbolt ii. to achieve infrared stealth, the exhaust gas is cooled to the temperatures where the brightest wavelengths it radiates are absorbed by atmospheric carbon dioxide and water vapor, greatly reducing the infrared visibility of the exhaust plume. another way to reduce the exhaust temperature is to circulate coolant fluids such as fuel inside the exhaust pipe, where the fuel tanks serve as heat sinks cooled by the flow of air along the wings. ground combat includes the use of both active and passive infrared sensors. thus, the united states marine corps ( usmc ) ground combat uniform requirements document specifies infrared reflective quality standards. = = reducing radio frequency ( rf ) emissions = = in addition to reducing infrared and acoustic emissions, a stealth vehicle must avoid radiating any other detectable energy, such as from onboard radars, communications systems, or rf leakage from electronics enclosures. the f - 117 uses passive infrared and low light level television sensor systems to aim its weapons and the f - 22 raptor has an advanced lpi radar which can illuminate enemy aircraft without triggering a radar warning receiver response. = = measuring = = the size of a target ' s image on radar is measured by the rcs, often represented by the symbol Ο and expressed in square meters. this does not equal geometric area. a perfectly conducting sphere of projected cross sectional area 1 m2 ( i. e. a diameter of 1. 13 m ) will have an rcs of 1 m2. note that for radar wavelengths much less than the diameter of the sphere, rcs is independent of frequency. conversely, a square flat plate of area 1 m2 will have an rcs of Ο = 4Ο a2 / Ξ»2 ( where a = area, Ξ» = wavelength ), or 13, 982 m2 at 10 ghz if the radar is perpendicular to the flat
. a reaction is said to be exergonic if the final state is lower on the energy scale than the initial state ; in the case of endergonic reactions the situation is the reverse. a reaction is said to be exothermic if the reaction releases heat to the surroundings ; in the case of endothermic reactions, the reaction absorbs heat from the surroundings. chemical reactions are invariably not possible unless the reactants surmount an energy barrier known as the activation energy. the speed of a chemical reaction ( at given temperature t ) is related to the activation energy e, by the boltzmann ' s population factor e β e / k t { \ displaystyle e ^ { - e / kt } } β that is the probability of a molecule to have energy greater than or equal to e at the given temperature t. this exponential dependence of a reaction rate on temperature is known as the arrhenius equation. the activation energy necessary for a chemical reaction to occur can be in the form of heat, light, electricity or mechanical force in the form of ultrasound. a related concept free energy, which also incorporates entropy considerations, is a very useful means for predicting the feasibility of a reaction and determining the state of equilibrium of a chemical reaction, in chemical thermodynamics. a reaction is feasible only if the total change in the gibbs free energy is negative, Ξ΄ g β€ 0 { \ displaystyle \ delta g \ leq 0 \, } ; if it is equal to zero the chemical reaction is said to be at equilibrium. there exist only limited possible states of energy for electrons, atoms and molecules. these are determined by the rules of quantum mechanics, which require quantization of energy of a bound system. the atoms / molecules in a higher energy state are said to be excited. the molecules / atoms of substance in an excited energy state are often much more reactive ; that is, more amenable to chemical reactions. the phase of a substance is invariably determined by its energy and the energy of its surroundings. when the intermolecular forces of a substance are such that the energy of the surroundings is not sufficient to overcome them, it occurs in a more ordered phase like liquid or solid as is the case with water ( h2o ) ; a liquid at room temperature because its molecules are bound by hydrogen bonds. whereas hydrogen sulfide ( h2s ) is a gas at room temperature and standard pressure, as its molecules are bound by weaker dipole β dipole interactions. the transfer of
a state of semi - perpetual convection. this convection process causes the lithospheric plates to move, albeit slowly. the resulting process is known as plate tectonics. areas of the crust where new crust is created are called divergent boundaries, those where it is brought back into the earth are convergent boundaries and those where plates slide past each other, but no new lithospheric material is created or destroyed, are referred to as transform ( or conservative ) boundaries. earthquakes result from the movement of the lithospheric plates, and they often occur near convergent boundaries where parts of the crust are forced into the earth as part of subduction. plate tectonics might be thought of as the process by which the earth is resurfaced. as the result of seafloor spreading, new crust and lithosphere is created by the flow of magma from the mantle to the near surface, through fissures, where it cools and solidifies. through subduction, oceanic crust and lithosphere vehemently returns to the convecting mantle. volcanoes result primarily from the melting of subducted crust material. crust material that is forced into the asthenosphere melts, and some portion of the melted material becomes light enough to rise to the surface β giving birth to volcanoes. = = atmospheric science = = atmospheric science initially developed in the late - 19th century as a means to forecast the weather through meteorology, the study of weather. atmospheric chemistry was developed in the 20th century to measure air pollution and expanded in the 1970s in response to acid rain. climatology studies the climate and climate change. the troposphere, stratosphere, mesosphere, thermosphere, and exosphere are the five layers which make up earth ' s atmosphere. 75 % of the mass in the atmosphere is located within the troposphere, the lowest layer. in all, the atmosphere is made up of about 78. 0 % nitrogen, 20. 9 % oxygen, and 0. 92 % argon, and small amounts of other gases including co2 and water vapor. water vapor and co2 cause the earth ' s atmosphere to catch and hold the sun ' s energy through the greenhouse effect. this makes earth ' s surface warm enough for liquid water and life. in addition to trapping heat, the atmosphere also protects living organisms by shielding the earth ' s surface from cosmic rays. the magnetic field β created by the internal motions of the core β produces the magnetosphere which protects earth '
higher concentrations of atmospheric nitrous oxide ( n2o ) are expected to slightly warm earth ' s surface because of increases in radiative forcing. radiative forcing is the difference in the net upward thermal radiation flux from the earth through a transparent atmosphere and radiation through an otherwise identical atmosphere with greenhouse gases. radiative forcing, normally measured in w / m ^ 2, depends on latitude, longitude and altitude, but it is often quoted for the tropopause, about 11 km of altitude for temperate latitudes, or for the top of the atmosphere at around 90 km. for current concentrations of greenhouse gases, the radiative forcing per added n2o molecule is about 230 times larger than the forcing per added carbon dioxide ( co2 ) molecule. this is due to the heavy saturation of the absorption band of the relatively abundant greenhouse gas, co2, compared to the much smaller saturation of the absorption bands of the trace greenhouse gas n2o. but the rate of increase of co2 molecules, about 2. 5 ppm / year ( ppm = part per million by mole ), is about 3000 times larger than the rate of increase of n2o molecules, which has held steady at around 0. 00085 ppm / year since 1985. so, the contribution of nitrous oxide to the annual increase in forcing is 230 / 3000 or about 1 / 13 that of co2. if the main greenhouse gases, co2, ch4 and n2o have contributed about 0. 1 c / decade of the warming observed over the past few decades, this would correspond to about 0. 00064 k per year or 0. 064 k per century of warming from n2o. proposals to place harsh restrictions on nitrous oxide emissions because of warming fears are not justified by these facts. restrictions would cause serious harm ; for example, by jeopardizing world food supplies.
10 kgy most food, which is ( with regard to warming ) physically equivalent to water, would warm by only about 2. 5 Β°c ( 4. 5 Β°f ). the specialty of processing food by ionizing radiation is the fact, that the energy density per atomic transition is very high, it can cleave molecules and induce ionization ( hence the name ) which cannot be achieved by mere heating. this is the reason for new beneficial effects, however at the same time, for new concerns. the treatment of solid food by ionizing radiation can provide an effect similar to heat pasteurization of liquids, such as milk. however, the use of the term, cold pasteurization, to describe irradiated foods is controversial, because pasteurization and irradiation are fundamentally different processes, although the intended end results can in some cases be similar. detractors of food irradiation have concerns about the health hazards of induced radioactivity. a report for the industry advocacy group american council on science and health entitled " irradiated foods " states : " the types of radiation sources approved for the treatment of foods have specific energy levels well below that which would cause any element in food to become radioactive. food undergoing irradiation does not become any more radioactive than luggage passing through an airport x - ray scanner or teeth that have been x - rayed. " food irradiation is currently permitted by over 40 countries and volumes are estimated to exceed 500, 000 metric tons ( 490, 000 long tons ; 550, 000 short tons ) annually worldwide. food irradiation is essentially a non - nuclear technology ; it relies on the use of ionizing radiation which may be generated by accelerators for electrons and conversion into bremsstrahlung, but which may use also gamma - rays from nuclear decay. there is a worldwide industry for processing by ionizing radiation, the majority by number and by processing power using accelerators. food irradiation is only a niche application compared to medical supplies, plastic materials, raw materials, gemstones, cables and wires, etc. = = accidents = = nuclear accidents, because of the powerful forces involved, are often very dangerous. historically, the first incidents involved fatal radiation exposure. marie curie died from aplastic anemia which resulted from her high levels of exposure. two scientists, an american and canadian respectively, harry daghlian and louis slotin, died after mishandling the same plutonium mass. unlike conventional weapons, the intense light, heat, and explosive force is
current in passing over from one concave bank to the next on the opposite side. the lowering of such a shoal by dredging merely effects a temporary deepening, for it soon forms again from the causes which produced it. the removal, moreover, of the rocky obstructions at rapids, though increasing the depth and equalizing the flow at these places, produces a lowering of the river above the rapids by facilitating the efflux, which may result in the appearance of fresh shoals at the low stage of the river. where, however, narrow rocky reefs or other hard shoals stretch across the bottom of a river and present obstacles to the erosion by the current of the soft materials forming the bed of the river above and below, their removal may result in permanent improvement by enabling the river to deepen its bed by natural scour. the capability of a river to provide a waterway for navigation during the summer or throughout the dry season depends on the depth that can be secured in the channel at the lowest stage. the problem in the dry season is the small discharge and deficiency in scour during this period. a typical solution is to restrict the width of the low - water channel, concentrate all of the flow in it, and also to fix its position so that it is scoured out every year by the floods which follow the deepest part of the bed along the line of the strongest current. this can be effected by closing subsidiary low - water channels with dikes across them, and narrowing the channel at the low stage by low - dipping cross dikes extending from the river banks down the slope and pointing slightly up - stream so as to direct the water flowing over them into a central channel. = = estuarine works = = the needs of navigation may also require that a stable, continuous, navigable channel is prolonged from the navigable river to deep water at the mouth of the estuary. the interaction of river flow and tide needs to be modeled by computer or using scale models, moulded to the configuration of the estuary under consideration and reproducing in miniature the tidal ebb and flow and fresh - water discharge over a bed of fine sand, in which various lines of training walls can be successively inserted. the models should be capable of furnishing valuable indications of the respective effects and comparative merits of the different schemes proposed for works. = = see also = = bridge scour flood control = = references = = = = external links = = u. s. army corps of engineers β civil works program river morphology and stream restoration references
masculinity and warmth. the five phases β fire, earth, metal, wood, and water β described a cycle of transformations in nature. the water turned into wood, which turned into the fire when it burned. the ashes left by fire were earth. using these principles, chinese philosophers and doctors explored human anatomy, characterizing organs as predominantly yin or yang, and understood the relationship between the pulse, the heart, and the flow of blood in the body centuries before it became accepted in the west. little evidence survives of how ancient indian cultures around the indus river understood nature, but some of their perspectives may be reflected in the vedas, a set of sacred hindu texts. they reveal a conception of the universe as ever - expanding and constantly being recycled and reformed. surgeons in the ayurvedic tradition saw health and illness as a combination of three humors : wind, bile and phlegm. a healthy life resulted from a balance among these humors. in ayurvedic thought, the body consisted of five elements : earth, water, fire, wind, and space. ayurvedic surgeons performed complex surgeries and developed a detailed understanding of human anatomy. pre - socratic philosophers in ancient greek culture brought natural philosophy a step closer to direct inquiry about cause and effect in nature between 600 and 400 bc. however, an element of magic and mythology remained. natural phenomena such as earthquakes and eclipses were explained increasingly in the context of nature itself instead of being attributed to angry gods. thales of miletus, an early philosopher who lived from 625 to 546 bc, explained earthquakes by theorizing that the world floated on water and that water was the fundamental element in nature. in the 5th century bc, leucippus was an early exponent of atomism, the idea that the world is made up of fundamental indivisible particles. pythagoras applied greek innovations in mathematics to astronomy and suggested that the earth was spherical. = = = aristotelian natural philosophy ( 400 bc β 1100 ad ) = = = later socratic and platonic thought focused on ethics, morals, and art and did not attempt an investigation of the physical world ; plato criticized pre - socratic thinkers as materialists and anti - religionists. aristotle, however, a student of plato who lived from 384 to 322 bc, paid closer attention to the natural world in his philosophy. in his history of animals, he described the inner workings of 110 species, including the stingray, catfish and
nuclear jets containing relativistic ` ` hot ' ' particles close to the central engine cool dramatically by producing high energy radiation. the radiative dissipation is similar to the famous compton drag acting upon ` ` cold ' ' thermal particles in a relativistic bulk flow. highly relativistic protons induce anisotropic showers raining electromagnetic power down onto the putative accretion disk. thus, the radiative signature of hot hadronic jets is x - ray irradiation of cold thermal matter. the synchrotron radio emission of the accelerated electrons is self - absorbed due to the strong magnetic fields close to the magnetic nozzle.
weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under ordinary conditions, most of the materials brought down from the high lands by torrential water courses are carried forward by the main river to the sea, or partially strewn over flat alluvial plains during floods ; the size of the materials forming the bed of the river or borne along by the stream is gradually reduced on proceeding seawards, so that in the po river in italy, for instance, pebbles and gravel are found for about 140 miles below turin, sand along the next 100 miles, and silt and mud in the last 110 miles ( 176 km ). = = channelization = = the removal of obstructions, natural or artificial
Question: When a cold front moves over a land region that has warm, moist air, a transfer of energy occurs. Cold air sinks and pushes warmer air upward. As energy is removed from the air, which change will most likely take place?
A) Ozone in the atmosphere will decrease.
B) Less ultraviolet radiation will reach the land.
C) More water will be absorbed from the area lakes.
D) Water in the atmosphere will undergo a phase change.
|
D) Water in the atmosphere will undergo a phase change.
|
Context:
can the apparent complexity we observe in the real world be generated from simple initial conditions via simple, deterministic rules?
a letter to the editor shortly summing up ten or so years of research into the h - index.
, tertiary, and quaternary ). the similarities among all known present - day species indicate that they have diverged through the process of evolution from their common ancestor. biologists regard the ubiquity of the genetic code as evidence of universal common descent for all bacteria, archaea, and eukaryotes. microbial mats of coexisting bacteria and archaea were the dominant form of life in the early archean eon and many of the major steps in early evolution are thought to have taken place in this environment. the earliest evidence of eukaryotes dates from 1. 85 billion years ago, and while they may have been present earlier, their diversification accelerated when they started using oxygen in their metabolism. later, around 1. 7 billion years ago, multicellular organisms began to appear, with differentiated cells performing specialised functions. algae - like multicellular land plants are dated back to about 1 billion years ago, although evidence suggests that microorganisms formed the earliest terrestrial ecosystems, at least 2. 7 billion years ago. microorganisms are thought to have paved the way for the inception of land plants in the ordovician period. land plants were so successful that they are thought to have contributed to the late devonian extinction event. ediacara biota appear during the ediacaran period, while vertebrates, along with most other modern phyla originated about 525 million years ago during the cambrian explosion. during the permian period, synapsids, including the ancestors of mammals, dominated the land, but most of this group became extinct in the permian β triassic extinction event 252 million years ago. during the recovery from this catastrophe, archosaurs became the most abundant land vertebrates ; one archosaur group, the dinosaurs, dominated the jurassic and cretaceous periods. after the cretaceous β paleogene extinction event 66 million years ago killed off the non - avian dinosaurs, mammals increased rapidly in size and diversity. such mass extinctions may have accelerated evolution by providing opportunities for new groups of organisms to diversify. = = diversity = = = = = bacteria and archaea = = = bacteria are a type of cell that constitute a large domain of prokaryotic microorganisms. typically a few micrometers in length, bacteria have a number of shapes, ranging from spheres to rods and spirals. bacteria were among the first life forms to appear on earth, and are present in most of its habitats. bacteria inhabit soil, water, acidic hot springs, radioactive
smallest genomes among flowering plants. arabidopsis was the first plant to have its genome sequenced, in 2000. the sequencing of some other relatively small genomes, of rice ( oryza sativa ) and brachypodium distachyon, has made them important model species for understanding the genetics, cellular and molecular biology of cereals, grasses and monocots generally. model plants such as arabidopsis thaliana are used for studying the molecular biology of plant cells and the chloroplast. ideally, these organisms have small genomes that are well known or completely sequenced, small stature and short generation times. corn has been used to study mechanisms of photosynthesis and phloem loading of sugar in c4 plants. the single celled green alga chlamydomonas reinhardtii, while not an embryophyte itself, contains a green - pigmented chloroplast related to that of land plants, making it useful for study. a red alga cyanidioschyzon merolae has also been used to study some basic chloroplast functions. spinach, peas, soybeans and a moss physcomitrella patens are commonly used to study plant cell biology. agrobacterium tumefaciens, a soil rhizosphere bacterium, can attach to plant cells and infect them with a callus - inducing ti plasmid by horizontal gene transfer, causing a callus infection called crown gall disease. schell and van montagu ( 1977 ) hypothesised that the ti plasmid could be a natural vector for introducing the nif gene responsible for nitrogen fixation in the root nodules of legumes and other plant species. today, genetic modification of the ti plasmid is one of the main techniques for introduction of transgenes to plants and the creation of genetically modified crops. = = = epigenetics = = = epigenetics is the study of heritable changes in gene function that cannot be explained by changes in the underlying dna sequence but cause the organism ' s genes to behave ( or " express themselves " ) differently. one example of epigenetic change is the marking of the genes by dna methylation which determines whether they will be expressed or not. gene expression can also be controlled by repressor proteins that attach to silencer regions of the dna and prevent that region of the dna code from being expressed. epigenetic marks may be added
as you read these words you are using a complex biological neural network. you have a highly interconnected set of some neurons to facilitate your reading, breathing, motion and thinking. each of your biological neurons, a rich assembly of tissue and chemistry, has the complexity, if not the speed, of a microprocessor. some of your neural structure was with you at birth. other parts have been established by experience.
the origins of the series of european cosmic - ray symposia are briefly described. the first meeting in the series, on hadronic interactions and extensive air showers, held in lodz, poland in 1968, was attended by the author : some memories are recounted.
by ancestry rather than superficial characteristics. while scientists do not always agree on how to classify organisms, molecular phylogenetics, which uses dna sequences as data, has driven many recent revisions along evolutionary lines and is likely to continue to do so. the dominant classification system is called linnaean taxonomy. it includes ranks and binomial nomenclature. the nomenclature of botanical organisms is codified in the international code of nomenclature for algae, fungi, and plants ( icn ) and administered by the international botanical congress. kingdom plantae belongs to domain eukaryota and is broken down recursively until each species is separately classified. the order is : kingdom ; phylum ( or division ) ; class ; order ; family ; genus ( plural genera ) ; species. the scientific name of a plant represents its genus and its species within the genus, resulting in a single worldwide name for each organism. for example, the tiger lily is lilium columbianum. lilium is the genus, and columbianum the specific epithet. the combination is the name of the species. when writing the scientific name of an organism, it is proper to capitalise the first letter in the genus and put all of the specific epithet in lowercase. additionally, the entire term is ordinarily italicised ( or underlined when italics are not available ). the evolutionary relationships and heredity of a group of organisms is called its phylogeny. phylogenetic studies attempt to discover phylogenies. the basic approach is to use similarities based on shared inheritance to determine relationships. as an example, species of pereskia are trees or bushes with prominent leaves. they do not obviously resemble a typical leafless cactus such as an echinocactus. however, both pereskia and echinocactus have spines produced from areoles ( highly specialised pad - like structures ) suggesting that the two genera are indeed related. judging relationships based on shared characters requires care, since plants may resemble one another through convergent evolution in which characters have arisen independently. some euphorbias have leafless, rounded bodies adapted to water conservation similar to those of globular cacti, but characters such as the structure of their flowers make it clear that the two groups are not closely related. the cladistic method takes a systematic approach to characters, distinguishing between those that carry no information about shared evolutionary history β such as those evolved separately in different groups ( homoplasies ) or those left over from ancestors ( plesiomorphies ) β and derived characters, which
listing of diseases in the family that may impact the patient. a family tree is sometimes used. history of present illness ( hpi ) : the chronological order of events of symptoms and further clarification of each symptom. distinguishable from history of previous illness, often called past medical history ( pmh ). medical history comprises hpi and pmh. medications ( rx ) : what drugs the patient takes including prescribed, over - the - counter, and home remedies, as well as alternative and herbal medicines or remedies. allergies are also recorded. past medical history ( pmh / pmhx ) : concurrent medical problems, past hospitalizations and operations, injuries, past infectious diseases or vaccinations, history of known allergies. review of systems ( ros ) or systems inquiry : a set of additional questions to ask, which may be missed on hpi : a general enquiry ( have you noticed any weight loss, change in sleep quality, fevers, lumps and bumps? etc. ), followed by questions on the body ' s main organ systems ( heart, lungs, digestive tract, urinary tract, etc. ). social history ( sh ) : birthplace, residences, marital history, social and economic status, habits ( including diet, medications, tobacco, alcohol ). the physical examination is the examination of the patient for medical signs of disease that are objective and observable, in contrast to symptoms that are volunteered by the patient and are not necessarily objectively observable. the healthcare provider uses sight, hearing, touch, and sometimes smell ( e. g., in infection, uremia, diabetic ketoacidosis ). four actions are the basis of physical examination : inspection, palpation ( feel ), percussion ( tap to determine resonance characteristics ), and auscultation ( listen ), generally in that order, although auscultation occurs prior to percussion and palpation for abdominal assessments. the clinical examination involves the study of : abdomen and rectum cardiovascular ( heart and blood vessels ) general appearance of the patient and specific indicators of disease ( nutritional status, presence of jaundice, pallor or clubbing ) genitalia ( and pregnancy if the patient is or could be pregnant ) head, eye, ear, nose, and throat ( heent ) musculoskeletal ( including spine and extremities ) neurological ( consciousness, awareness, brain, vision, cranial nerves,
the internal functions and processes within plant organelles, cells, tissues, whole plants, plant populations and plant communities. at each of these levels, a botanist may be concerned with the classification ( taxonomy ), phylogeny and evolution, structure ( anatomy and morphology ), or function ( physiology ) of plant life. the strictest definition of " plant " includes only the " land plants " or embryophytes, which include seed plants ( gymnosperms, including the pines, and flowering plants ) and the free - sporing cryptogams including ferns, clubmosses, liverworts, hornworts and mosses. embryophytes are multicellular eukaryotes descended from an ancestor that obtained its energy from sunlight by photosynthesis. they have life cycles with alternating haploid and diploid phases. the sexual haploid phase of embryophytes, known as the gametophyte, nurtures the developing diploid embryo sporophyte within its tissues for at least part of its life, even in the seed plants, where the gametophyte itself is nurtured by its parent sporophyte. other groups of organisms that were previously studied by botanists include bacteria ( now studied in bacteriology ), fungi ( mycology ) β including lichen - forming fungi ( lichenology ), non - chlorophyte algae ( phycology ), and viruses ( virology ). however, attention is still given to these groups by botanists, and fungi ( including lichens ) and photosynthetic protists are usually covered in introductory botany courses. palaeobotanists study ancient plants in the fossil record to provide information about the evolutionary history of plants. cyanobacteria, the first oxygen - releasing photosynthetic organisms on earth, are thought to have given rise to the ancestor of plants by entering into an endosymbiotic relationship with an early eukaryote, ultimately becoming the chloroplasts in plant cells. the new photosynthetic plants ( along with their algal relatives ) accelerated the rise in atmospheric oxygen started by the cyanobacteria, changing the ancient oxygen - free, reducing, atmosphere to one in which free oxygen has been abundant for more than 2 billion years. among the important botanical questions of the 21st century are the role of plants as primary producers in the global cycling of life ' s basic ingredients : energy, carbon, oxygen, nitrogen and water, and ways
##rozoic eon that began 539 million years ago being subdivided into paleozoic, mesozoic, and cenozoic eras. these three eras together comprise eleven periods ( cambrian, ordovician, silurian, devonian, carboniferous, permian, triassic, jurassic, cretaceous, tertiary, and quaternary ). the similarities among all known present - day species indicate that they have diverged through the process of evolution from their common ancestor. biologists regard the ubiquity of the genetic code as evidence of universal common descent for all bacteria, archaea, and eukaryotes. microbial mats of coexisting bacteria and archaea were the dominant form of life in the early archean eon and many of the major steps in early evolution are thought to have taken place in this environment. the earliest evidence of eukaryotes dates from 1. 85 billion years ago, and while they may have been present earlier, their diversification accelerated when they started using oxygen in their metabolism. later, around 1. 7 billion years ago, multicellular organisms began to appear, with differentiated cells performing specialised functions. algae - like multicellular land plants are dated back to about 1 billion years ago, although evidence suggests that microorganisms formed the earliest terrestrial ecosystems, at least 2. 7 billion years ago. microorganisms are thought to have paved the way for the inception of land plants in the ordovician period. land plants were so successful that they are thought to have contributed to the late devonian extinction event. ediacara biota appear during the ediacaran period, while vertebrates, along with most other modern phyla originated about 525 million years ago during the cambrian explosion. during the permian period, synapsids, including the ancestors of mammals, dominated the land, but most of this group became extinct in the permian β triassic extinction event 252 million years ago. during the recovery from this catastrophe, archosaurs became the most abundant land vertebrates ; one archosaur group, the dinosaurs, dominated the jurassic and cretaceous periods. after the cretaceous β paleogene extinction event 66 million years ago killed off the non - avian dinosaurs, mammals increased rapidly in size and diversity. such mass extinctions may have accelerated evolution by providing opportunities for new groups of organisms to diversify. = = diversity = = = = = bacteria and archaea = = = bacteria are a type of cell that constitute a large domain of prokar
Question: Which sequence lists the levels of organization in the human body from simplest to most complex?
A) organ system -> tissue -> cell -> organ
B) tissue -> cell -> organ -> organ system
C) organ -> organ system -> tissue -> cell
D) cell -> tissue -> organ -> organ system
|
D) cell -> tissue -> organ -> organ system
|
Context:
and peripheral blood. they concluded from the results that immuno - cytochemical staining of bone marrow and peripheral blood is a sensitive and simple way to detect and quantify breast cancer cells. one of the main reasons for metastatic relapse in patients with solid tumours is the early dissemination of malignant cells. the use of monoclonal antibodies ( mabs ) specific for cytokeratins can identify disseminated individual epithelial tumor cells in the bone marrow. one study reports on having developed an immuno - cytochemical procedure for simultaneous labeling of cytokeratin component no. 18 ( ck18 ) and prostate specific antigen ( psa ). this would help in the further characterization of disseminated individual epithelial tumor cells in patients with prostate cancer. the twelve control aspirates from patients with benign prostatic hyperplasia showed negative staining, which further supports the specificity of ck18 in detecting epithelial tumour cells in bone marrow. in most cases of malignant disease complicated by effusion, neoplastic cells can be easily recognized. however, in some cases, malignant cells are not so easily seen or their presence is too doubtful to call it a positive report. the use of immuno - cytochemical techniques increases diagnostic accuracy in these cases. ghosh, mason and spriggs analysed 53 samples of pleural or peritoneal fluid from 41 patients with malignant disease. conventional cytological examination had not revealed any neoplastic cells. three monoclonal antibodies ( anti - cea, ca 1 and hmfg - 2 ) were used to search for malignant cells. immunocytochemical labelling was performed on unstained smears, which had been stored at - 20 Β°c up to 18 months. twelve of the forty - one cases in which immuno - cytochemical staining was performed, revealed malignant cells. the result represented an increase in diagnostic accuracy of approximately 20 %. the study concluded that in patients with suspected malignant disease, immuno - cytochemical labeling should be used routinely in the examination of cytologically negative samples and has important implications with respect to patient management. another application of immuno - cytochemical staining is for the detection of two antigens in the same smear. double staining with light chain antibodies and with t and b cell markers can indicate the neoplastic origin of a lymph
it was the best of times ; it was the worst of times is the way dickens begins the tale of two cities. the line is appropriate to our time in particle physics. it is the best of times because we are in the midst of a revolution in understanding, the third to occur during my career. it is the worst of times because accelerator facilities are shutting down before new ones are opening, restricting the opportunity for experiments, and because of great uncertainty about future funding. my task today is to give you a view of the most important opportunities for our field under a scenario that is constrained by a tight budget. it is a time when we cannot afford the merely good, but must give first priority to the really important.
apoptosis is a complex pathway regulated by the concerted action of multiple pro - and anti - apoptotic molecules. the intrinsic ( mitochondrial ) pathway of apoptosis is governed up - stream of mitochondria, by the family of bcl - 2 proteins, and down - stream of mitochondria, by low - probability events, such as apoptosome formation, and by feedback circuits involving caspases and inhibitor of apoptosis proteins ( iaps ), such as xiap. all these regulatory mechanisms ensure that cells only commit to death once a threshold of damage has been reached and the anti - apoptotic reserve of the cell is overcome. as cancer cells are invariably exposed to strong intracellular and extracellular stress stimuli, they are particularly reliant on the expression of anti - apoptotic proteins. hence, many cancer cells undergo apoptosis when exposed to agents that inhibit anti - apoptotic bcl - 2 molecules, such as bh3 mimetics, while normal cells remain relatively insensitive to single agent treatments with the same class of molecules. targeting different proteins within the apoptotic network with combinatorial treatment approaches often achieves even greater specificity. this led us to investigate the sensitivity of leukemia and lymphoma cells to a pro - apoptotic action of a bh3 mimetic combined with a small molecule inhibitor of xiap. using computational probabilistic model of apoptotic pathway, verified by experimental results from human leukemia and lymphoma cell lines, we show that inhibition of xiap has a non - linear effect on sensitization towards apoptosis induced by the bh3 mimetic ha14 - 1. this study justifies further ex vivo and animal studies on the potential of the treatment of leukemia and lymphoma with a combination of bh3 mimetics and xiap inhibitors.
##rates and peripheral blood, further development of this method is necessary before it can be used routinely. one major drawback of immuno - cytochemistry is that only tumor - associated and not tumor - specific monoclonal antibodies are used, and as a result, some cross - reaction with normal cells can occur. in order to effectively stage breast cancer and assess the efficacy of purging regimens prior to autologous stem cell infusion, it is important to detect even small quantities of breast cancer cells. immuno - histochemical methods are ideal for this purpose because they are simple, sensitive, and quite specific. franklin et al. performed a sensitive immuno - cytochemical assay by using a combination of four monoclonal antibodies ( 260f9, 520c9, 317g5 and bre - 3 ) against tumor cell surface glycoproteins to identify breast tumour cells in bone marrow and peripheral blood. they concluded from the results that immuno - cytochemical staining of bone marrow and peripheral blood is a sensitive and simple way to detect and quantify breast cancer cells. one of the main reasons for metastatic relapse in patients with solid tumours is the early dissemination of malignant cells. the use of monoclonal antibodies ( mabs ) specific for cytokeratins can identify disseminated individual epithelial tumor cells in the bone marrow. one study reports on having developed an immuno - cytochemical procedure for simultaneous labeling of cytokeratin component no. 18 ( ck18 ) and prostate specific antigen ( psa ). this would help in the further characterization of disseminated individual epithelial tumor cells in patients with prostate cancer. the twelve control aspirates from patients with benign prostatic hyperplasia showed negative staining, which further supports the specificity of ck18 in detecting epithelial tumour cells in bone marrow. in most cases of malignant disease complicated by effusion, neoplastic cells can be easily recognized. however, in some cases, malignant cells are not so easily seen or their presence is too doubtful to call it a positive report. the use of immuno - cytochemical techniques increases diagnostic accuracy in these cases. ghosh, mason and spriggs analysed 53 samples of pleural or peritoneal fluid from 41 patients with malignant disease. conventional cytological examination had not revealed any neoplastic cells. three monocl
##tase, human chorionic gonadotrophin, Ξ± - fetoprotein and others are organ - associated antigens and the production of monoclonal antibodies against these antigens helps in determining the nature of a primary tumor. monoclonal antibodies are especially useful in distinguishing morphologically similar lesions, like pleural and peritoneal mesothelioma, adenocarcinoma, and in the determination of the organ or tissue origin of undifferentiated metastases. selected monoclonal antibodies help in the detection of occult metastases ( cancer of unknown primary origin ) by immuno - cytological analysis of bone marrow, other tissue aspirates, as well as lymph nodes and other tissues and can have increased sensitivity over normal histopathological staining. one study performed a sensitive immuno - histochemical assay on bone marrow aspirates of 20 patients with localized prostate cancer. three monoclonal antibodies ( t16, c26, and ae - 1 ), capable of recognizing membrane and cytoskeletal antigens expressed by epithelial cells to detect tumour cells, were used in the assay. bone marrow aspirates of 22 % of patients with localized prostate cancer ( stage b, 0 / 5 ; stage c, 2 / 4 ), and 36 % patients with metastatic prostate cancer ( stage d1, 0 / 7 patients ; stage d2, 4 / 4 patients ) had antigen - positive cells in their bone marrow. it was concluded that immuno - histochemical staining of bone marrow aspirates are very useful to detect occult bone marrow metastases in patients with apparently localized prostate cancer. although immuno - cytochemistry using tumor - associated monoclonal antibodies has led to an improved ability to detect occult breast cancer cells in bone marrow aspirates and peripheral blood, further development of this method is necessary before it can be used routinely. one major drawback of immuno - cytochemistry is that only tumor - associated and not tumor - specific monoclonal antibodies are used, and as a result, some cross - reaction with normal cells can occur. in order to effectively stage breast cancer and assess the efficacy of purging regimens prior to autologous stem cell infusion, it is important to detect even small quantities of breast cancer cells. immuno - histochemical methods are ideal for this purpose because they are simple, sensitive, and quite specific
##gnant cells. the use of monoclonal antibodies ( mabs ) specific for cytokeratins can identify disseminated individual epithelial tumor cells in the bone marrow. one study reports on having developed an immuno - cytochemical procedure for simultaneous labeling of cytokeratin component no. 18 ( ck18 ) and prostate specific antigen ( psa ). this would help in the further characterization of disseminated individual epithelial tumor cells in patients with prostate cancer. the twelve control aspirates from patients with benign prostatic hyperplasia showed negative staining, which further supports the specificity of ck18 in detecting epithelial tumour cells in bone marrow. in most cases of malignant disease complicated by effusion, neoplastic cells can be easily recognized. however, in some cases, malignant cells are not so easily seen or their presence is too doubtful to call it a positive report. the use of immuno - cytochemical techniques increases diagnostic accuracy in these cases. ghosh, mason and spriggs analysed 53 samples of pleural or peritoneal fluid from 41 patients with malignant disease. conventional cytological examination had not revealed any neoplastic cells. three monoclonal antibodies ( anti - cea, ca 1 and hmfg - 2 ) were used to search for malignant cells. immunocytochemical labelling was performed on unstained smears, which had been stored at - 20 Β°c up to 18 months. twelve of the forty - one cases in which immuno - cytochemical staining was performed, revealed malignant cells. the result represented an increase in diagnostic accuracy of approximately 20 %. the study concluded that in patients with suspected malignant disease, immuno - cytochemical labeling should be used routinely in the examination of cytologically negative samples and has important implications with respect to patient management. another application of immuno - cytochemical staining is for the detection of two antigens in the same smear. double staining with light chain antibodies and with t and b cell markers can indicate the neoplastic origin of a lymphoma. one study has reported the isolation of a hybridoma cell line ( clone 1e10 ), which produces a monoclonal antibody ( igm, k isotype ). this monoclonal antibody shows specific immuno - cytochemical staining of nucleoli. tissues and tumours can
weakly established hybridomas. unfused b cells die as they have a short life span. in this way, only the b cell - myeloma hybrids survive, since the hgprt gene coming from the b cells is functional. these cells produce antibodies ( a property of b cells ) and are immortal ( a property of myeloma cells ). the incubated medium is then diluted into multi - well plates to such an extent that each well contains only one cell. since the antibodies in a well are produced by the same b cell, they will be directed towards the same epitope, and are thus monoclonal antibodies. the next stage is a rapid primary screening process, which identifies and selects only those hybridomas that produce antibodies of appropriate specificity. the first screening technique used is called elisa. the hybridoma culture supernatant, secondary enzyme labeled conjugate, and chromogenic substrate, are then incubated, and the formation of a colored product indicates a positive hybridoma. alternatively, immunocytochemical, western blot, and immunoprecipitation - mass spectrometry. unlike western blot assays, immunoprecipitation - mass spectrometry facilitates screening and ranking of clones which bind to the native ( non - denaturated ) forms of antigen proteins. flow cytometry screening has been used for primary screening of a large number ( ~ 1000 ) of hybridoma clones recognizing the native form of the antigen on the cell surface. in the flow cytometry - based screening, a mixture of antigen - negative cells and antigen - positive cells is used as the antigen to be tested for each hybridoma supernatant sample. the b cell that produces the desired antibodies can be cloned to produce many identical daughter clones. supplemental media containing interleukin - 6 ( such as briclone ) are essential for this step. once a hybridoma colony is established, it will continually grow in culture medium like rpmi - 1640 ( with antibiotics and fetal bovine serum ) and produce antibodies. multiwell plates are used initially to grow the hybridomas, and after selection, are changed to larger tissue culture flasks. this maintains the well - being of the hybridomas and provides enough cells for cryopreservation and supernatant for subsequent investigations. the culture supernatant can yield 1 to 60 ΞΌg / ml of monoclonal antibody, which is maintained at -
prostate cancer. three monoclonal antibodies ( t16, c26, and ae - 1 ), capable of recognizing membrane and cytoskeletal antigens expressed by epithelial cells to detect tumour cells, were used in the assay. bone marrow aspirates of 22 % of patients with localized prostate cancer ( stage b, 0 / 5 ; stage c, 2 / 4 ), and 36 % patients with metastatic prostate cancer ( stage d1, 0 / 7 patients ; stage d2, 4 / 4 patients ) had antigen - positive cells in their bone marrow. it was concluded that immuno - histochemical staining of bone marrow aspirates are very useful to detect occult bone marrow metastases in patients with apparently localized prostate cancer. although immuno - cytochemistry using tumor - associated monoclonal antibodies has led to an improved ability to detect occult breast cancer cells in bone marrow aspirates and peripheral blood, further development of this method is necessary before it can be used routinely. one major drawback of immuno - cytochemistry is that only tumor - associated and not tumor - specific monoclonal antibodies are used, and as a result, some cross - reaction with normal cells can occur. in order to effectively stage breast cancer and assess the efficacy of purging regimens prior to autologous stem cell infusion, it is important to detect even small quantities of breast cancer cells. immuno - histochemical methods are ideal for this purpose because they are simple, sensitive, and quite specific. franklin et al. performed a sensitive immuno - cytochemical assay by using a combination of four monoclonal antibodies ( 260f9, 520c9, 317g5 and bre - 3 ) against tumor cell surface glycoproteins to identify breast tumour cells in bone marrow and peripheral blood. they concluded from the results that immuno - cytochemical staining of bone marrow and peripheral blood is a sensitive and simple way to detect and quantify breast cancer cells. one of the main reasons for metastatic relapse in patients with solid tumours is the early dissemination of malignant cells. the use of monoclonal antibodies ( mabs ) specific for cytokeratins can identify disseminated individual epithelial tumor cells in the bone marrow. one study reports on having developed an immuno - cytochemical procedure for simultaneous labeling of cytokeratin component no. 18
. in order to effectively stage breast cancer and assess the efficacy of purging regimens prior to autologous stem cell infusion, it is important to detect even small quantities of breast cancer cells. immuno - histochemical methods are ideal for this purpose because they are simple, sensitive, and quite specific. franklin et al. performed a sensitive immuno - cytochemical assay by using a combination of four monoclonal antibodies ( 260f9, 520c9, 317g5 and bre - 3 ) against tumor cell surface glycoproteins to identify breast tumour cells in bone marrow and peripheral blood. they concluded from the results that immuno - cytochemical staining of bone marrow and peripheral blood is a sensitive and simple way to detect and quantify breast cancer cells. one of the main reasons for metastatic relapse in patients with solid tumours is the early dissemination of malignant cells. the use of monoclonal antibodies ( mabs ) specific for cytokeratins can identify disseminated individual epithelial tumor cells in the bone marrow. one study reports on having developed an immuno - cytochemical procedure for simultaneous labeling of cytokeratin component no. 18 ( ck18 ) and prostate specific antigen ( psa ). this would help in the further characterization of disseminated individual epithelial tumor cells in patients with prostate cancer. the twelve control aspirates from patients with benign prostatic hyperplasia showed negative staining, which further supports the specificity of ck18 in detecting epithelial tumour cells in bone marrow. in most cases of malignant disease complicated by effusion, neoplastic cells can be easily recognized. however, in some cases, malignant cells are not so easily seen or their presence is too doubtful to call it a positive report. the use of immuno - cytochemical techniques increases diagnostic accuracy in these cases. ghosh, mason and spriggs analysed 53 samples of pleural or peritoneal fluid from 41 patients with malignant disease. conventional cytological examination had not revealed any neoplastic cells. three monoclonal antibodies ( anti - cea, ca 1 and hmfg - 2 ) were used to search for malignant cells. immunocytochemical labelling was performed on unstained smears, which had been stored at - 20 Β°c up to 18 months. twelve of the forty - one cases
for nucleotide synthesis. hence, unfused myeloma cells die, as they cannot produce nucleotides by the de novo or salvage pathways because they lack hgprt. removal of the unfused myeloma cells is necessary because they have the potential to outgrow other cells, especially weakly established hybridomas. unfused b cells die as they have a short life span. in this way, only the b cell - myeloma hybrids survive, since the hgprt gene coming from the b cells is functional. these cells produce antibodies ( a property of b cells ) and are immortal ( a property of myeloma cells ). the incubated medium is then diluted into multi - well plates to such an extent that each well contains only one cell. since the antibodies in a well are produced by the same b cell, they will be directed towards the same epitope, and are thus monoclonal antibodies. the next stage is a rapid primary screening process, which identifies and selects only those hybridomas that produce antibodies of appropriate specificity. the first screening technique used is called elisa. the hybridoma culture supernatant, secondary enzyme labeled conjugate, and chromogenic substrate, are then incubated, and the formation of a colored product indicates a positive hybridoma. alternatively, immunocytochemical, western blot, and immunoprecipitation - mass spectrometry. unlike western blot assays, immunoprecipitation - mass spectrometry facilitates screening and ranking of clones which bind to the native ( non - denaturated ) forms of antigen proteins. flow cytometry screening has been used for primary screening of a large number ( ~ 1000 ) of hybridoma clones recognizing the native form of the antigen on the cell surface. in the flow cytometry - based screening, a mixture of antigen - negative cells and antigen - positive cells is used as the antigen to be tested for each hybridoma supernatant sample. the b cell that produces the desired antibodies can be cloned to produce many identical daughter clones. supplemental media containing interleukin - 6 ( such as briclone ) are essential for this step. once a hybridoma colony is established, it will continually grow in culture medium like rpmi - 1640 ( with antibiotics and fetal bovine serum ) and produce antibodies. multiwell plates are used initially to grow the hybridomas, and after selection,
Question: Cancer is best described as
A) a type of genetic engineering
B) the merging of sex cells
C) abnormal cell division
D) tissue that stops growing
|
C) abnormal cell division
|
Context:
blood vessels. mechanical stimuli, such as pressure pulses seem to be beneficial to all kind of cardiovascular tissue such as heart valves, blood vessels or pericardium. = = = bioreactors = = = in tissue engineering, a bioreactor is a device that attempts to simulate a physiological environment in order to promote cell or tissue growth in vitro. a physiological environment can consist of many different parameters such as temperature, pressure, oxygen or carbon dioxide concentration, or osmolality of fluid environment, and it can extend to all kinds of biological, chemical or mechanical stimuli. therefore, there are systems that may include the application of forces such as electromagnetic forces, mechanical pressures, or fluid pressures to the tissue. these systems can be two - or three - dimensional setups. bioreactors can be used in both academic and industry applications. general - use and application - specific bioreactors are also commercially available, which may provide static chemical stimulation or a combination of chemical and mechanical stimulation. cell proliferation and differentiation are largely influenced by mechanical and biochemical cues in the surrounding extracellular matrix environment. bioreactors are typically developed to replicate the specific physiological environment of the tissue being grown ( e. g., flex and fluid shearing for heart tissue growth ). this can allow specialized cell lines to thrive in cultures replicating their native environments, but it also makes bioreactors attractive tools for culturing stem cells. a successful stem - cell - based bioreactor is effective at expanding stem cells with uniform properties and / or promoting controlled, reproducible differentiation into selected mature cell types. there are a variety of bioreactors designed for 3d cell cultures. there are small plastic cylindrical chambers, as well as glass chambers, with regulated internal humidity and moisture specifically engineered for the purpose of growing cells in three dimensions. the bioreactor uses bioactive synthetic materials such as polyethylene terephthalate membranes to surround the spheroid cells in an environment that maintains high levels of nutrients. they are easy to open and close, so that cell spheroids can be removed for testing, yet the chamber is able to maintain 100 % humidity throughout. this humidity is important to achieve maximum cell growth and function. the bioreactor chamber is part of a larger device that rotates to ensure equal cell growth in each direction across three dimensions. quinxell technologies now under quintech life sciences from singapore has developed a bioreactor known as the tisxell biaxial bioreactor which is specially designed for the purpose of
this scaffold and cells were placed in a bioreactor, where it matured to become a partially or fully transplantable organ. the work was called a " landmark ". the lab first stripped the cells away from a rat heart ( a process called " decellularization " ) and then injected rat stem cells into the decellularized rat heart. tissue - engineered blood vessels : blood vessels that have been grown in a lab and can be used to repair damaged blood vessels without eliciting an immune response. tissue engineered blood vessels have been developed by many different approaches. they could be implanted as pre - seeded cellularized blood vessels, as acellular vascular grafts made with decellularized vessels or synthetic vascular grafts. artificial skin constructed from human skin cells embedded in a hydrogel, such as in the case of bio - printed constructs for battlefield burn repairs. artificial bone marrow : bone marrow cultured in vitro to be transplanted serves as a " just cells " approach to tissue engineering. tissue engineered bone : a structural matrix can be composed of metals such as titanium, polymers of varying degradation rates, or certain types of ceramics. materials are often chosen to recruit osteoblasts to aid in reforming the bone and returning biological function. various types of cells can be added directly into the matrix to expedite the process. laboratory - grown penis : decellularized scaffolds of rabbit penises were recellularised with smooth muscle and endothelial cells. the organ was then transplanted to live rabbits and functioned comparably to the native organ, suggesting potential as treatment for genital trauma. oral mucosa tissue engineering uses a cells and scaffold approach to replicate the 3 dimensional structure and function of oral mucosa. = = cells as building blocks = = cells are one of the main components for the success of tissue engineering approaches. tissue engineering uses cells as strategies for creation / replacement of new tissue. examples include fibroblasts used for skin repair or renewal, chondrocytes used for cartilage repair ( maci β fda approved product ), and hepatocytes used in liver support systems cells can be used alone or with support matrices for tissue engineering applications. an adequate environment for promoting cell growth, differentiation, and integration with the existing tissue is a critical factor for cell - based building blocks. manipulation of any of these cell processes create alternative avenues for the development of new tissue ( e. g., cell reprogramming - somatic
einstein, when he began working on the general theory of relativity, believed that energy of any kind is the source of the gravitational field. therefore, the energy of gravity, like any energy, must be the source of the field. it was previously discovered that the energy - momentum tensor of the gravitational field is already contained in the ricci tensor. this hypothesis is used to construct a new equation of the gravitational field.
##ilage generated without the use of exogenous scaffold material. in this methodology, all material in the construct is cellular produced directly by the cells. bioartificial heart : doris taylor ' s lab constructed a biocompatible rat heart by re - cellularising a de - cellularised rat heart. this scaffold and cells were placed in a bioreactor, where it matured to become a partially or fully transplantable organ. the work was called a " landmark ". the lab first stripped the cells away from a rat heart ( a process called " decellularization " ) and then injected rat stem cells into the decellularized rat heart. tissue - engineered blood vessels : blood vessels that have been grown in a lab and can be used to repair damaged blood vessels without eliciting an immune response. tissue engineered blood vessels have been developed by many different approaches. they could be implanted as pre - seeded cellularized blood vessels, as acellular vascular grafts made with decellularized vessels or synthetic vascular grafts. artificial skin constructed from human skin cells embedded in a hydrogel, such as in the case of bio - printed constructs for battlefield burn repairs. artificial bone marrow : bone marrow cultured in vitro to be transplanted serves as a " just cells " approach to tissue engineering. tissue engineered bone : a structural matrix can be composed of metals such as titanium, polymers of varying degradation rates, or certain types of ceramics. materials are often chosen to recruit osteoblasts to aid in reforming the bone and returning biological function. various types of cells can be added directly into the matrix to expedite the process. laboratory - grown penis : decellularized scaffolds of rabbit penises were recellularised with smooth muscle and endothelial cells. the organ was then transplanted to live rabbits and functioned comparably to the native organ, suggesting potential as treatment for genital trauma. oral mucosa tissue engineering uses a cells and scaffold approach to replicate the 3 dimensional structure and function of oral mucosa. = = cells as building blocks = = cells are one of the main components for the success of tissue engineering approaches. tissue engineering uses cells as strategies for creation / replacement of new tissue. examples include fibroblasts used for skin repair or renewal, chondrocytes used for cartilage repair ( maci β fda approved product ), and hepatocytes used in liver support systems cells can be used alone or with
such as electromagnetic forces, mechanical pressures, or fluid pressures to the tissue. these systems can be two - or three - dimensional setups. bioreactors can be used in both academic and industry applications. general - use and application - specific bioreactors are also commercially available, which may provide static chemical stimulation or a combination of chemical and mechanical stimulation. cell proliferation and differentiation are largely influenced by mechanical and biochemical cues in the surrounding extracellular matrix environment. bioreactors are typically developed to replicate the specific physiological environment of the tissue being grown ( e. g., flex and fluid shearing for heart tissue growth ). this can allow specialized cell lines to thrive in cultures replicating their native environments, but it also makes bioreactors attractive tools for culturing stem cells. a successful stem - cell - based bioreactor is effective at expanding stem cells with uniform properties and / or promoting controlled, reproducible differentiation into selected mature cell types. there are a variety of bioreactors designed for 3d cell cultures. there are small plastic cylindrical chambers, as well as glass chambers, with regulated internal humidity and moisture specifically engineered for the purpose of growing cells in three dimensions. the bioreactor uses bioactive synthetic materials such as polyethylene terephthalate membranes to surround the spheroid cells in an environment that maintains high levels of nutrients. they are easy to open and close, so that cell spheroids can be removed for testing, yet the chamber is able to maintain 100 % humidity throughout. this humidity is important to achieve maximum cell growth and function. the bioreactor chamber is part of a larger device that rotates to ensure equal cell growth in each direction across three dimensions. quinxell technologies now under quintech life sciences from singapore has developed a bioreactor known as the tisxell biaxial bioreactor which is specially designed for the purpose of tissue engineering. it is the first bioreactor in the world to have a spherical glass chamber with biaxial rotation ; specifically to mimic the rotation of the fetus in the womb ; which provides a conducive environment for the growth of tissues. multiple forms of mechanical stimulation have also been combined into a single bioreactor. using gene expression analysis, one academic study found that applying a combination of cyclic strain and ultrasound stimulation to pre - osteoblast cells in a bioreactor accelerated matrix maturation and differentiation. the technology of this combined stimulation bioreactor could be used to grow bone cells more quickly and effectively
into seven out of approximately 20 human test subjects as part of a long - term experiment. cartilage : lab - grown cartilage, cultured in vitro on a scaffold, was successfully used as an autologous transplant to repair patients ' knees. scaffold - free cartilage : cartilage generated without the use of exogenous scaffold material. in this methodology, all material in the construct is cellular produced directly by the cells. bioartificial heart : doris taylor ' s lab constructed a biocompatible rat heart by re - cellularising a de - cellularised rat heart. this scaffold and cells were placed in a bioreactor, where it matured to become a partially or fully transplantable organ. the work was called a " landmark ". the lab first stripped the cells away from a rat heart ( a process called " decellularization " ) and then injected rat stem cells into the decellularized rat heart. tissue - engineered blood vessels : blood vessels that have been grown in a lab and can be used to repair damaged blood vessels without eliciting an immune response. tissue engineered blood vessels have been developed by many different approaches. they could be implanted as pre - seeded cellularized blood vessels, as acellular vascular grafts made with decellularized vessels or synthetic vascular grafts. artificial skin constructed from human skin cells embedded in a hydrogel, such as in the case of bio - printed constructs for battlefield burn repairs. artificial bone marrow : bone marrow cultured in vitro to be transplanted serves as a " just cells " approach to tissue engineering. tissue engineered bone : a structural matrix can be composed of metals such as titanium, polymers of varying degradation rates, or certain types of ceramics. materials are often chosen to recruit osteoblasts to aid in reforming the bone and returning biological function. various types of cells can be added directly into the matrix to expedite the process. laboratory - grown penis : decellularized scaffolds of rabbit penises were recellularised with smooth muscle and endothelial cells. the organ was then transplanted to live rabbits and functioned comparably to the native organ, suggesting potential as treatment for genital trauma. oral mucosa tissue engineering uses a cells and scaffold approach to replicate the 3 dimensional structure and function of oral mucosa. = = cells as building blocks = = cells are one of the main components for the success of tissue engineering approaches
cells into the decellularized rat heart. tissue - engineered blood vessels : blood vessels that have been grown in a lab and can be used to repair damaged blood vessels without eliciting an immune response. tissue engineered blood vessels have been developed by many different approaches. they could be implanted as pre - seeded cellularized blood vessels, as acellular vascular grafts made with decellularized vessels or synthetic vascular grafts. artificial skin constructed from human skin cells embedded in a hydrogel, such as in the case of bio - printed constructs for battlefield burn repairs. artificial bone marrow : bone marrow cultured in vitro to be transplanted serves as a " just cells " approach to tissue engineering. tissue engineered bone : a structural matrix can be composed of metals such as titanium, polymers of varying degradation rates, or certain types of ceramics. materials are often chosen to recruit osteoblasts to aid in reforming the bone and returning biological function. various types of cells can be added directly into the matrix to expedite the process. laboratory - grown penis : decellularized scaffolds of rabbit penises were recellularised with smooth muscle and endothelial cells. the organ was then transplanted to live rabbits and functioned comparably to the native organ, suggesting potential as treatment for genital trauma. oral mucosa tissue engineering uses a cells and scaffold approach to replicate the 3 dimensional structure and function of oral mucosa. = = cells as building blocks = = cells are one of the main components for the success of tissue engineering approaches. tissue engineering uses cells as strategies for creation / replacement of new tissue. examples include fibroblasts used for skin repair or renewal, chondrocytes used for cartilage repair ( maci β fda approved product ), and hepatocytes used in liver support systems cells can be used alone or with support matrices for tissue engineering applications. an adequate environment for promoting cell growth, differentiation, and integration with the existing tissue is a critical factor for cell - based building blocks. manipulation of any of these cell processes create alternative avenues for the development of new tissue ( e. g., cell reprogramming - somatic cells, vascularization ). = = = isolation = = = techniques for cell isolation depend on the cell source. centrifugation and apheresis are techniques used for extracting cells from biofluids ( e. g., blood ). whereas digestion processes, typically using enzymes to remove the extra
defective body parts. inside the body, artificial heart valves are in common use with artificial hearts and lungs seeing less common use but under active technology development. other medical devices and aids that can be considered prosthetics include hearing aids, artificial eyes, palatal obturator, gastric bands, and dentures. prostheses are specifically not orthoses, although given certain circumstances a prosthesis might end up performing some or all of the same functionary benefits as an orthosis. prostheses are technically the complete finished item. for instance, a c - leg knee alone is not a prosthesis, but only a prosthetic component. the complete prosthesis would consist of the attachment system to the residual limb β usually a " socket ", and all the attachment hardware components all the way down to and including the terminal device. despite the technical difference, the terms are often used interchangeably. the terms " prosthetic " and " orthotic " are adjectives used to describe devices such as a prosthetic knee. the terms " prosthetics " and " orthotics " are used to describe the respective allied health fields. an occupational therapist ' s role in prosthetics include therapy, training and evaluations. prosthetic training includes orientation to prosthetics components and terminology, donning and doffing, wearing schedule, and how to care for residual limb and the prosthesis. = = = exoskeletons = = = a powered exoskeleton is a wearable mobile machine that is powered by a system of electric motors, pneumatics, levers, hydraulics, or a combination of technologies that allow for limb movement with increased strength and endurance. its design aims to provide back support, sense the user ' s motion, and send a signal to motors which manage the gears. the exoskeleton supports the shoulder, waist and thigh, and assists movement for lifting and holding heavy items, while lowering back stress. = = = adaptive seating and positioning = = = people with balance and motor function challenges often need specialized equipment to sit or stand safely and securely. this equipment is frequently specialized for specific settings such as in a classroom or nursing home. positioning is often important in seating arrangements to ensure that user ' s body pressure is distributed equally without inhibiting movement in a desired way. positioning devices have been developed to aid in allowing people to stand and bear weight on their legs without risk of a fall.
so mars below means blood and war ", is a false cause fallacy. : 26 many astrologers claim that astrology is scientific. if one were to attempt to try to explain it scientifically, there are only four fundamental forces ( conventionally ), limiting the choice of possible natural mechanisms. : 65 some astrologers have proposed conventional causal agents such as electromagnetism and gravity. the strength of these forces drops off with distance. : 65 scientists reject these proposed mechanisms as implausible since, for example, the magnetic field, when measured from earth, of a large but distant planet such as jupiter is far smaller than that produced by ordinary household appliances. astronomer phil plait noted that in terms of magnitude, the sun is the only object with an electromagnetic field of note, but astrology isn ' t based just off the sun alone. : 65 while astrologers could try to suggest a fifth force, this is inconsistent with the trends in physics with the unification of electromagnetism and the weak force into the electroweak force. if the astrologer insisted on being inconsistent with the current understanding and evidential basis of physics, that would be an extraordinary claim. : 65 it would also be inconsistent with the other forces which drop off with distance. : 65 if distance is irrelevant, then, logically, all objects in space should be taken into account. : 66 carl jung sought to invoke synchronicity, the claim that two events have some sort of acausal connection, to explain the lack of statistically significant results on astrology from a single study he conducted. however, synchronicity itself is considered neither testable nor falsifiable. the study was subsequently heavily criticised for its non - random sample and its use of statistics and also its lack of consistency with astrology. = = psychology = = psychological studies have not found any robust relationship between astrological signs and life outcomes. for example, a study showed that zodiac signs are no more effective than random numbers in predicting subjective well - being and quality of life. it has also been shown that confirmation bias is a psychological factor that contributes to belief in astrology. : 344 : 180 β 181 : 42 β 48 confirmation bias is a form of cognitive bias. : 553 from the literature, astrology believers often tend to selectively remember those predictions that turned out to be true and do not remember those that turned out false. another, separate, form of confirmation bias also plays a role, where believers often fail to
, characterizing organs as predominantly yin or yang, and understood the relationship between the pulse, the heart, and the flow of blood in the body centuries before it became accepted in the west. little evidence survives of how ancient indian cultures around the indus river understood nature, but some of their perspectives may be reflected in the vedas, a set of sacred hindu texts. they reveal a conception of the universe as ever - expanding and constantly being recycled and reformed. surgeons in the ayurvedic tradition saw health and illness as a combination of three humors : wind, bile and phlegm. a healthy life resulted from a balance among these humors. in ayurvedic thought, the body consisted of five elements : earth, water, fire, wind, and space. ayurvedic surgeons performed complex surgeries and developed a detailed understanding of human anatomy. pre - socratic philosophers in ancient greek culture brought natural philosophy a step closer to direct inquiry about cause and effect in nature between 600 and 400 bc. however, an element of magic and mythology remained. natural phenomena such as earthquakes and eclipses were explained increasingly in the context of nature itself instead of being attributed to angry gods. thales of miletus, an early philosopher who lived from 625 to 546 bc, explained earthquakes by theorizing that the world floated on water and that water was the fundamental element in nature. in the 5th century bc, leucippus was an early exponent of atomism, the idea that the world is made up of fundamental indivisible particles. pythagoras applied greek innovations in mathematics to astronomy and suggested that the earth was spherical. = = = aristotelian natural philosophy ( 400 bc β 1100 ad ) = = = later socratic and platonic thought focused on ethics, morals, and art and did not attempt an investigation of the physical world ; plato criticized pre - socratic thinkers as materialists and anti - religionists. aristotle, however, a student of plato who lived from 384 to 322 bc, paid closer attention to the natural world in his philosophy. in his history of animals, he described the inner workings of 110 species, including the stingray, catfish and bee. he investigated chick embryos by breaking open eggs and observing them at various stages of development. aristotle ' s works were influential through the 16th century, and he is considered to be the father of biology for his pioneering work in that science. he also presented philosophies about physics, nature, and astronomy using
Question: To pump blood through the body, the heart must overcome the force of gravity. Which body position would require the least work from the heart?
A) sitting
B) standing
C) kneeling
D) lying down
|
D) lying down
|
Context:
of several types of apomixis that occur in plants. apomixis can also happen in a seed, producing a seed that contains an embryo genetically identical to the parent. most sexually reproducing organisms are diploid, with paired chromosomes, but doubling of their chromosome number may occur due to errors in cytokinesis. this can occur early in development to produce an autopolyploid or partly autopolyploid organism, or during normal processes of cellular differentiation to produce some cell types that are polyploid ( endopolyploidy ), or during gamete formation. an allopolyploid plant may result from a hybridisation event between two different species. both autopolyploid and allopolyploid plants can often reproduce normally, but may be unable to cross - breed successfully with the parent population because there is a mismatch in chromosome numbers. these plants that are reproductively isolated from the parent species but live within the same geographical area, may be sufficiently successful to form a new species. some otherwise sterile plant polyploids can still reproduce vegetatively or by seed apomixis, forming clonal populations of identical individuals. durum wheat is a fertile tetraploid allopolyploid, while bread wheat is a fertile hexaploid. the commercial banana is an example of a sterile, seedless triploid hybrid. common dandelion is a triploid that produces viable seeds by apomictic seed. as in other eukaryotes, the inheritance of endosymbiotic organelles like mitochondria and chloroplasts in plants is non - mendelian. chloroplasts are inherited through the male parent in gymnosperms but often through the female parent in flowering plants. = = = molecular genetics = = = a considerable amount of new knowledge about plant function comes from studies of the molecular genetics of model plants such as the thale cress, arabidopsis thaliana, a weedy species in the mustard family ( brassicaceae ). the genome or hereditary information contained in the genes of this species is encoded by about 135 million base pairs of dna, forming one of the smallest genomes among flowering plants. arabidopsis was the first plant to have its genome sequenced, in 2000. the sequencing of some other relatively small genomes, of rice ( oryza sativa ) and brachypodium distachyon, has made them important model species for understanding the genetics,
tubers in potato is one example. particularly in arctic or alpine habitats, where opportunities for fertilisation of flowers by animals are rare, plantlets or bulbs, may develop instead of flowers, replacing sexual reproduction with asexual reproduction and giving rise to clonal populations genetically identical to the parent. this is one of several types of apomixis that occur in plants. apomixis can also happen in a seed, producing a seed that contains an embryo genetically identical to the parent. most sexually reproducing organisms are diploid, with paired chromosomes, but doubling of their chromosome number may occur due to errors in cytokinesis. this can occur early in development to produce an autopolyploid or partly autopolyploid organism, or during normal processes of cellular differentiation to produce some cell types that are polyploid ( endopolyploidy ), or during gamete formation. an allopolyploid plant may result from a hybridisation event between two different species. both autopolyploid and allopolyploid plants can often reproduce normally, but may be unable to cross - breed successfully with the parent population because there is a mismatch in chromosome numbers. these plants that are reproductively isolated from the parent species but live within the same geographical area, may be sufficiently successful to form a new species. some otherwise sterile plant polyploids can still reproduce vegetatively or by seed apomixis, forming clonal populations of identical individuals. durum wheat is a fertile tetraploid allopolyploid, while bread wheat is a fertile hexaploid. the commercial banana is an example of a sterile, seedless triploid hybrid. common dandelion is a triploid that produces viable seeds by apomictic seed. as in other eukaryotes, the inheritance of endosymbiotic organelles like mitochondria and chloroplasts in plants is non - mendelian. chloroplasts are inherited through the male parent in gymnosperms but often through the female parent in flowering plants. = = = molecular genetics = = = a considerable amount of new knowledge about plant function comes from studies of the molecular genetics of model plants such as the thale cress, arabidopsis thaliana, a weedy species in the mustard family ( brassicaceae ). the genome or hereditary information contained in the genes of this species is encoded by about 135 million base pairs of dna, forming one of the
, subsequent switching to inbreeding becomes disadvantageous since it allows expression of the previously masked deleterious recessive mutations, commonly referred to as inbreeding depression. unlike in higher animals, where parthenogenesis is rare, asexual reproduction may occur in plants by several different mechanisms. the formation of stem tubers in potato is one example. particularly in arctic or alpine habitats, where opportunities for fertilisation of flowers by animals are rare, plantlets or bulbs, may develop instead of flowers, replacing sexual reproduction with asexual reproduction and giving rise to clonal populations genetically identical to the parent. this is one of several types of apomixis that occur in plants. apomixis can also happen in a seed, producing a seed that contains an embryo genetically identical to the parent. most sexually reproducing organisms are diploid, with paired chromosomes, but doubling of their chromosome number may occur due to errors in cytokinesis. this can occur early in development to produce an autopolyploid or partly autopolyploid organism, or during normal processes of cellular differentiation to produce some cell types that are polyploid ( endopolyploidy ), or during gamete formation. an allopolyploid plant may result from a hybridisation event between two different species. both autopolyploid and allopolyploid plants can often reproduce normally, but may be unable to cross - breed successfully with the parent population because there is a mismatch in chromosome numbers. these plants that are reproductively isolated from the parent species but live within the same geographical area, may be sufficiently successful to form a new species. some otherwise sterile plant polyploids can still reproduce vegetatively or by seed apomixis, forming clonal populations of identical individuals. durum wheat is a fertile tetraploid allopolyploid, while bread wheat is a fertile hexaploid. the commercial banana is an example of a sterile, seedless triploid hybrid. common dandelion is a triploid that produces viable seeds by apomictic seed. as in other eukaryotes, the inheritance of endosymbiotic organelles like mitochondria and chloroplasts in plants is non - mendelian. chloroplasts are inherited through the male parent in gymnosperms but often through the female parent in flowering plants. = = = molecular genetics = = = a considerable amount of new knowledge about plant function comes from
often injurious, at least with the plants on which i experimented. " an important adaptive benefit of outcrossing is that it allows the masking of deleterious mutations in the genome of progeny. this beneficial effect is also known as hybrid vigor or heterosis. once outcrossing is established, subsequent switching to inbreeding becomes disadvantageous since it allows expression of the previously masked deleterious recessive mutations, commonly referred to as inbreeding depression. unlike in higher animals, where parthenogenesis is rare, asexual reproduction may occur in plants by several different mechanisms. the formation of stem tubers in potato is one example. particularly in arctic or alpine habitats, where opportunities for fertilisation of flowers by animals are rare, plantlets or bulbs, may develop instead of flowers, replacing sexual reproduction with asexual reproduction and giving rise to clonal populations genetically identical to the parent. this is one of several types of apomixis that occur in plants. apomixis can also happen in a seed, producing a seed that contains an embryo genetically identical to the parent. most sexually reproducing organisms are diploid, with paired chromosomes, but doubling of their chromosome number may occur due to errors in cytokinesis. this can occur early in development to produce an autopolyploid or partly autopolyploid organism, or during normal processes of cellular differentiation to produce some cell types that are polyploid ( endopolyploidy ), or during gamete formation. an allopolyploid plant may result from a hybridisation event between two different species. both autopolyploid and allopolyploid plants can often reproduce normally, but may be unable to cross - breed successfully with the parent population because there is a mismatch in chromosome numbers. these plants that are reproductively isolated from the parent species but live within the same geographical area, may be sufficiently successful to form a new species. some otherwise sterile plant polyploids can still reproduce vegetatively or by seed apomixis, forming clonal populations of identical individuals. durum wheat is a fertile tetraploid allopolyploid, while bread wheat is a fertile hexaploid. the commercial banana is an example of a sterile, seedless triploid hybrid. common dandelion is a triploid that produces viable seeds by apomictic seed. as in other eukaryotes, the inheritance of endosymbiotic organelles like
cytokinesis. this can occur early in development to produce an autopolyploid or partly autopolyploid organism, or during normal processes of cellular differentiation to produce some cell types that are polyploid ( endopolyploidy ), or during gamete formation. an allopolyploid plant may result from a hybridisation event between two different species. both autopolyploid and allopolyploid plants can often reproduce normally, but may be unable to cross - breed successfully with the parent population because there is a mismatch in chromosome numbers. these plants that are reproductively isolated from the parent species but live within the same geographical area, may be sufficiently successful to form a new species. some otherwise sterile plant polyploids can still reproduce vegetatively or by seed apomixis, forming clonal populations of identical individuals. durum wheat is a fertile tetraploid allopolyploid, while bread wheat is a fertile hexaploid. the commercial banana is an example of a sterile, seedless triploid hybrid. common dandelion is a triploid that produces viable seeds by apomictic seed. as in other eukaryotes, the inheritance of endosymbiotic organelles like mitochondria and chloroplasts in plants is non - mendelian. chloroplasts are inherited through the male parent in gymnosperms but often through the female parent in flowering plants. = = = molecular genetics = = = a considerable amount of new knowledge about plant function comes from studies of the molecular genetics of model plants such as the thale cress, arabidopsis thaliana, a weedy species in the mustard family ( brassicaceae ). the genome or hereditary information contained in the genes of this species is encoded by about 135 million base pairs of dna, forming one of the smallest genomes among flowering plants. arabidopsis was the first plant to have its genome sequenced, in 2000. the sequencing of some other relatively small genomes, of rice ( oryza sativa ) and brachypodium distachyon, has made them important model species for understanding the genetics, cellular and molecular biology of cereals, grasses and monocots generally. model plants such as arabidopsis thaliana are used for studying the molecular biology of plant cells and the chloroplast. ideally, these organisms have small genomes that are well known or completely sequenced, small stature and short
stems mainly provide support to the leaves and reproductive structures, but can store water in succulent plants such as cacti, food as in potato tubers, or reproduce vegetatively as in the stolons of strawberry plants or in the process of layering. leaves gather sunlight and carry out photosynthesis. large, flat, flexible, green leaves are called foliage leaves. gymnosperms, such as conifers, cycads, ginkgo, and gnetophytes are seed - producing plants with open seeds. angiosperms are seed - producing plants that produce flowers and have enclosed seeds. woody plants, such as azaleas and oaks, undergo a secondary growth phase resulting in two additional types of tissues : wood ( secondary xylem ) and bark ( secondary phloem and cork ). all gymnosperms and many angiosperms are woody plants. some plants reproduce sexually, some asexually, and some via both means. although reference to major morphological categories such as root, stem, leaf, and trichome are useful, one has to keep in mind that these categories are linked through intermediate forms so that a continuum between the categories results. furthermore, structures can be seen as processes, that is, process combinations. = = systematic botany = = systematic botany is part of systematic biology, which is concerned with the range and diversity of organisms and their relationships, particularly as determined by their evolutionary history. it involves, or is related to, biological classification, scientific taxonomy and phylogenetics. biological classification is the method by which botanists group organisms into categories such as genera or species. biological classification is a form of scientific taxonomy. modern taxonomy is rooted in the work of carl linnaeus, who grouped species according to shared physical characteristics. these groupings have since been revised to align better with the darwinian principle of common descent β grouping organisms by ancestry rather than superficial characteristics. while scientists do not always agree on how to classify organisms, molecular phylogenetics, which uses dna sequences as data, has driven many recent revisions along evolutionary lines and is likely to continue to do so. the dominant classification system is called linnaean taxonomy. it includes ranks and binomial nomenclature. the nomenclature of botanical organisms is codified in the international code of nomenclature for algae, fungi, and plants ( icn ) and administered by the international botanical congress. kingdom plantae belongs to domain eukaryota and is broken down recursively until each species is separately classified. the order is :
inherited traits such as shape in pisum sativum ( peas ). what mendel learned from studying plants has had far - reaching benefits outside of botany. similarly, " jumping genes " were discovered by barbara mcclintock while she was studying maize. nevertheless, there are some distinctive genetic differences between plants and other organisms. species boundaries in plants may be weaker than in animals, and cross species hybrids are often possible. a familiar example is peppermint, mentha Γ piperita, a sterile hybrid between mentha aquatica and spearmint, mentha spicata. the many cultivated varieties of wheat are the result of multiple inter - and intra - specific crosses between wild species and their hybrids. angiosperms with monoecious flowers often have self - incompatibility mechanisms that operate between the pollen and stigma so that the pollen either fails to reach the stigma or fails to germinate and produce male gametes. this is one of several methods used by plants to promote outcrossing. in many land plants the male and female gametes are produced by separate individuals. these species are said to be dioecious when referring to vascular plant sporophytes and dioicous when referring to bryophyte gametophytes. charles darwin in his 1878 book the effects of cross and self - fertilization in the vegetable kingdom at the start of chapter xii noted " the first and most important of the conclusions which may be drawn from the observations given in this volume, is that generally cross - fertilisation is beneficial and self - fertilisation often injurious, at least with the plants on which i experimented. " an important adaptive benefit of outcrossing is that it allows the masking of deleterious mutations in the genome of progeny. this beneficial effect is also known as hybrid vigor or heterosis. once outcrossing is established, subsequent switching to inbreeding becomes disadvantageous since it allows expression of the previously masked deleterious recessive mutations, commonly referred to as inbreeding depression. unlike in higher animals, where parthenogenesis is rare, asexual reproduction may occur in plants by several different mechanisms. the formation of stem tubers in potato is one example. particularly in arctic or alpine habitats, where opportunities for fertilisation of flowers by animals are rare, plantlets or bulbs, may develop instead of flowers, replacing sexual reproduction with asexual reproduction and giving rise to clonal populations genetically identical to the parent. this is one
, the other can often regrow it. in fact it is possible to grow an entire plant from a single leaf, as is the case with plants in streptocarpus sect. saintpaulia, or even a single cell β which can dedifferentiate into a callus ( a mass of unspecialised cells ) that can grow into a new plant. in vascular plants, the xylem and phloem are the conductive tissues that transport resources between shoots and roots. roots are often adapted to store food such as sugars or starch, as in sugar beets and carrots. stems mainly provide support to the leaves and reproductive structures, but can store water in succulent plants such as cacti, food as in potato tubers, or reproduce vegetatively as in the stolons of strawberry plants or in the process of layering. leaves gather sunlight and carry out photosynthesis. large, flat, flexible, green leaves are called foliage leaves. gymnosperms, such as conifers, cycads, ginkgo, and gnetophytes are seed - producing plants with open seeds. angiosperms are seed - producing plants that produce flowers and have enclosed seeds. woody plants, such as azaleas and oaks, undergo a secondary growth phase resulting in two additional types of tissues : wood ( secondary xylem ) and bark ( secondary phloem and cork ). all gymnosperms and many angiosperms are woody plants. some plants reproduce sexually, some asexually, and some via both means. although reference to major morphological categories such as root, stem, leaf, and trichome are useful, one has to keep in mind that these categories are linked through intermediate forms so that a continuum between the categories results. furthermore, structures can be seen as processes, that is, process combinations. = = systematic botany = = systematic botany is part of systematic biology, which is concerned with the range and diversity of organisms and their relationships, particularly as determined by their evolutionary history. it involves, or is related to, biological classification, scientific taxonomy and phylogenetics. biological classification is the method by which botanists group organisms into categories such as genera or species. biological classification is a form of scientific taxonomy. modern taxonomy is rooted in the work of carl linnaeus, who grouped species according to shared physical characteristics. these groupings have since been revised to align better with the darwinian principle of common descent β grouping organisms
of several methods used by plants to promote outcrossing. in many land plants the male and female gametes are produced by separate individuals. these species are said to be dioecious when referring to vascular plant sporophytes and dioicous when referring to bryophyte gametophytes. charles darwin in his 1878 book the effects of cross and self - fertilization in the vegetable kingdom at the start of chapter xii noted " the first and most important of the conclusions which may be drawn from the observations given in this volume, is that generally cross - fertilisation is beneficial and self - fertilisation often injurious, at least with the plants on which i experimented. " an important adaptive benefit of outcrossing is that it allows the masking of deleterious mutations in the genome of progeny. this beneficial effect is also known as hybrid vigor or heterosis. once outcrossing is established, subsequent switching to inbreeding becomes disadvantageous since it allows expression of the previously masked deleterious recessive mutations, commonly referred to as inbreeding depression. unlike in higher animals, where parthenogenesis is rare, asexual reproduction may occur in plants by several different mechanisms. the formation of stem tubers in potato is one example. particularly in arctic or alpine habitats, where opportunities for fertilisation of flowers by animals are rare, plantlets or bulbs, may develop instead of flowers, replacing sexual reproduction with asexual reproduction and giving rise to clonal populations genetically identical to the parent. this is one of several types of apomixis that occur in plants. apomixis can also happen in a seed, producing a seed that contains an embryo genetically identical to the parent. most sexually reproducing organisms are diploid, with paired chromosomes, but doubling of their chromosome number may occur due to errors in cytokinesis. this can occur early in development to produce an autopolyploid or partly autopolyploid organism, or during normal processes of cellular differentiation to produce some cell types that are polyploid ( endopolyploidy ), or during gamete formation. an allopolyploid plant may result from a hybridisation event between two different species. both autopolyploid and allopolyploid plants can often reproduce normally, but may be unable to cross - breed successfully with the parent population because there is a mismatch in chromosome numbers. these plants that are reproductively isolated from the parent
. species boundaries in plants may be weaker than in animals, and cross species hybrids are often possible. a familiar example is peppermint, mentha Γ piperita, a sterile hybrid between mentha aquatica and spearmint, mentha spicata. the many cultivated varieties of wheat are the result of multiple inter - and intra - specific crosses between wild species and their hybrids. angiosperms with monoecious flowers often have self - incompatibility mechanisms that operate between the pollen and stigma so that the pollen either fails to reach the stigma or fails to germinate and produce male gametes. this is one of several methods used by plants to promote outcrossing. in many land plants the male and female gametes are produced by separate individuals. these species are said to be dioecious when referring to vascular plant sporophytes and dioicous when referring to bryophyte gametophytes. charles darwin in his 1878 book the effects of cross and self - fertilization in the vegetable kingdom at the start of chapter xii noted " the first and most important of the conclusions which may be drawn from the observations given in this volume, is that generally cross - fertilisation is beneficial and self - fertilisation often injurious, at least with the plants on which i experimented. " an important adaptive benefit of outcrossing is that it allows the masking of deleterious mutations in the genome of progeny. this beneficial effect is also known as hybrid vigor or heterosis. once outcrossing is established, subsequent switching to inbreeding becomes disadvantageous since it allows expression of the previously masked deleterious recessive mutations, commonly referred to as inbreeding depression. unlike in higher animals, where parthenogenesis is rare, asexual reproduction may occur in plants by several different mechanisms. the formation of stem tubers in potato is one example. particularly in arctic or alpine habitats, where opportunities for fertilisation of flowers by animals are rare, plantlets or bulbs, may develop instead of flowers, replacing sexual reproduction with asexual reproduction and giving rise to clonal populations genetically identical to the parent. this is one of several types of apomixis that occur in plants. apomixis can also happen in a seed, producing a seed that contains an embryo genetically identical to the parent. most sexually reproducing organisms are diploid, with paired chromosomes, but doubling of their chromosome number may occur due to errors in
Question: In order to grow two of the exact same plants, a person would reproduce them asexually because the
A) genetic diversity is greater.
B) offspring would be genetically identical.
C) recessive traits would emerge.
D) negative traits are suppressed.
|
B) offspring would be genetically identical.
|
Context:
, the other can often regrow it. in fact it is possible to grow an entire plant from a single leaf, as is the case with plants in streptocarpus sect. saintpaulia, or even a single cell β which can dedifferentiate into a callus ( a mass of unspecialised cells ) that can grow into a new plant. in vascular plants, the xylem and phloem are the conductive tissues that transport resources between shoots and roots. roots are often adapted to store food such as sugars or starch, as in sugar beets and carrots. stems mainly provide support to the leaves and reproductive structures, but can store water in succulent plants such as cacti, food as in potato tubers, or reproduce vegetatively as in the stolons of strawberry plants or in the process of layering. leaves gather sunlight and carry out photosynthesis. large, flat, flexible, green leaves are called foliage leaves. gymnosperms, such as conifers, cycads, ginkgo, and gnetophytes are seed - producing plants with open seeds. angiosperms are seed - producing plants that produce flowers and have enclosed seeds. woody plants, such as azaleas and oaks, undergo a secondary growth phase resulting in two additional types of tissues : wood ( secondary xylem ) and bark ( secondary phloem and cork ). all gymnosperms and many angiosperms are woody plants. some plants reproduce sexually, some asexually, and some via both means. although reference to major morphological categories such as root, stem, leaf, and trichome are useful, one has to keep in mind that these categories are linked through intermediate forms so that a continuum between the categories results. furthermore, structures can be seen as processes, that is, process combinations. = = systematic botany = = systematic botany is part of systematic biology, which is concerned with the range and diversity of organisms and their relationships, particularly as determined by their evolutionary history. it involves, or is related to, biological classification, scientific taxonomy and phylogenetics. biological classification is the method by which botanists group organisms into categories such as genera or species. biological classification is a form of scientific taxonomy. modern taxonomy is rooted in the work of carl linnaeus, who grouped species according to shared physical characteristics. these groupings have since been revised to align better with the darwinian principle of common descent β grouping organisms
unspecialised cells ) that can grow into a new plant. in vascular plants, the xylem and phloem are the conductive tissues that transport resources between shoots and roots. roots are often adapted to store food such as sugars or starch, as in sugar beets and carrots. stems mainly provide support to the leaves and reproductive structures, but can store water in succulent plants such as cacti, food as in potato tubers, or reproduce vegetatively as in the stolons of strawberry plants or in the process of layering. leaves gather sunlight and carry out photosynthesis. large, flat, flexible, green leaves are called foliage leaves. gymnosperms, such as conifers, cycads, ginkgo, and gnetophytes are seed - producing plants with open seeds. angiosperms are seed - producing plants that produce flowers and have enclosed seeds. woody plants, such as azaleas and oaks, undergo a secondary growth phase resulting in two additional types of tissues : wood ( secondary xylem ) and bark ( secondary phloem and cork ). all gymnosperms and many angiosperms are woody plants. some plants reproduce sexually, some asexually, and some via both means. although reference to major morphological categories such as root, stem, leaf, and trichome are useful, one has to keep in mind that these categories are linked through intermediate forms so that a continuum between the categories results. furthermore, structures can be seen as processes, that is, process combinations. = = systematic botany = = systematic botany is part of systematic biology, which is concerned with the range and diversity of organisms and their relationships, particularly as determined by their evolutionary history. it involves, or is related to, biological classification, scientific taxonomy and phylogenetics. biological classification is the method by which botanists group organisms into categories such as genera or species. biological classification is a form of scientific taxonomy. modern taxonomy is rooted in the work of carl linnaeus, who grouped species according to shared physical characteristics. these groupings have since been revised to align better with the darwinian principle of common descent β grouping organisms by ancestry rather than superficial characteristics. while scientists do not always agree on how to classify organisms, molecular phylogenetics, which uses dna sequences as data, has driven many recent revisions along evolutionary lines and is likely to continue to do so. the dominant classification system is called linnaean taxonomy. it includes ranks and binomi
studies of the molecular genetics of model plants such as the thale cress, arabidopsis thaliana, a weedy species in the mustard family ( brassicaceae ). the genome or hereditary information contained in the genes of this species is encoded by about 135 million base pairs of dna, forming one of the smallest genomes among flowering plants. arabidopsis was the first plant to have its genome sequenced, in 2000. the sequencing of some other relatively small genomes, of rice ( oryza sativa ) and brachypodium distachyon, has made them important model species for understanding the genetics, cellular and molecular biology of cereals, grasses and monocots generally. model plants such as arabidopsis thaliana are used for studying the molecular biology of plant cells and the chloroplast. ideally, these organisms have small genomes that are well known or completely sequenced, small stature and short generation times. corn has been used to study mechanisms of photosynthesis and phloem loading of sugar in c4 plants. the single celled green alga chlamydomonas reinhardtii, while not an embryophyte itself, contains a green - pigmented chloroplast related to that of land plants, making it useful for study. a red alga cyanidioschyzon merolae has also been used to study some basic chloroplast functions. spinach, peas, soybeans and a moss physcomitrella patens are commonly used to study plant cell biology. agrobacterium tumefaciens, a soil rhizosphere bacterium, can attach to plant cells and infect them with a callus - inducing ti plasmid by horizontal gene transfer, causing a callus infection called crown gall disease. schell and van montagu ( 1977 ) hypothesised that the ti plasmid could be a natural vector for introducing the nif gene responsible for nitrogen fixation in the root nodules of legumes and other plant species. today, genetic modification of the ti plasmid is one of the main techniques for introduction of transgenes to plants and the creation of genetically modified crops. = = = epigenetics = = = epigenetics is the study of heritable changes in gene function that cannot be explained by changes in the underlying dna sequence but cause the organism ' s genes to behave ( or " express themselves " ) differently. one example
in plants the dna is often inserted using agrobacterium - mediated transformation, taking advantage of the agrobacteriums t - dna sequence that allows natural insertion of genetic material into plant cells. other methods include biolistics, where particles of gold or tungsten are coated with dna and then shot into young plant cells, and electroporation, which involves using an electric shock to make the cell membrane permeable to plasmid dna. as only a single cell is transformed with genetic material, the organism must be regenerated from that single cell. in plants this is accomplished through the use of tissue culture. in animals it is necessary to ensure that the inserted dna is present in the embryonic stem cells. bacteria consist of a single cell and reproduce clonally so regeneration is not necessary. selectable markers are used to easily differentiate transformed from untransformed cells. these markers are usually present in the transgenic organism, although a number of strategies have been developed that can remove the selectable marker from the mature transgenic plant. further testing using pcr, southern hybridization, and dna sequencing is conducted to confirm that an organism contains the new gene. these tests can also confirm the chromosomal location and copy number of the inserted gene. the presence of the gene does not guarantee it will be expressed at appropriate levels in the target tissue so methods that look for and measure the gene products ( rna and protein ) are also used. these include northern hybridisation, quantitative rt - pcr, western blot, immunofluorescence, elisa and phenotypic analysis. the new genetic material can be inserted randomly within the host genome or targeted to a specific location. the technique of gene targeting uses homologous recombination to make desired changes to a specific endogenous gene. this tends to occur at a relatively low frequency in plants and animals and generally requires the use of selectable markers. the frequency of gene targeting can be greatly enhanced through genome editing. genome editing uses artificially engineered nucleases that create specific double - stranded breaks at desired locations in the genome, and use the cell ' s endogenous mechanisms to repair the induced break by the natural processes of homologous recombination and nonhomologous end - joining. there are four families of engineered nucleases : meganucleases, zinc finger nucleases, transcription activator - like effector nucleases ( talens ), and the cas9 - guide
plant cells and tissues, whereas plant morphology is the study of their external form. all plants are multicellular eukaryotes, their dna stored in nuclei. the characteristic features of plant cells that distinguish them from those of animals and fungi include a primary cell wall composed of the polysaccharides cellulose, hemicellulose and pectin, larger vacuoles than in animal cells and the presence of plastids with unique photosynthetic and biosynthetic functions as in the chloroplasts. other plastids contain storage products such as starch ( amyloplasts ) or lipids ( elaioplasts ). uniquely, streptophyte cells and those of the green algal order trentepohliales divide by construction of a phragmoplast as a template for building a cell plate late in cell division. the bodies of vascular plants including clubmosses, ferns and seed plants ( gymnosperms and angiosperms ) generally have aerial and subterranean subsystems. the shoots consist of stems bearing green photosynthesising leaves and reproductive structures. the underground vascularised roots bear root hairs at their tips and generally lack chlorophyll. non - vascular plants, the liverworts, hornworts and mosses do not produce ground - penetrating vascular roots and most of the plant participates in photosynthesis. the sporophyte generation is nonphotosynthetic in liverworts but may be able to contribute part of its energy needs by photosynthesis in mosses and hornworts. the root system and the shoot system are interdependent β the usually nonphotosynthetic root system depends on the shoot system for food, and the usually photosynthetic shoot system depends on water and minerals from the root system. cells in each system are capable of creating cells of the other and producing adventitious shoots or roots. stolons and tubers are examples of shoots that can grow roots. roots that spread out close to the surface, such as those of willows, can produce shoots and ultimately new plants. in the event that one of the systems is lost, the other can often regrow it. in fact it is possible to grow an entire plant from a single leaf, as is the case with plants in streptocarpus sect. saintpaulia, or even a single cell β which can dedifferentiate into a callus ( a mass of
inherited traits such as shape in pisum sativum ( peas ). what mendel learned from studying plants has had far - reaching benefits outside of botany. similarly, " jumping genes " were discovered by barbara mcclintock while she was studying maize. nevertheless, there are some distinctive genetic differences between plants and other organisms. species boundaries in plants may be weaker than in animals, and cross species hybrids are often possible. a familiar example is peppermint, mentha Γ piperita, a sterile hybrid between mentha aquatica and spearmint, mentha spicata. the many cultivated varieties of wheat are the result of multiple inter - and intra - specific crosses between wild species and their hybrids. angiosperms with monoecious flowers often have self - incompatibility mechanisms that operate between the pollen and stigma so that the pollen either fails to reach the stigma or fails to germinate and produce male gametes. this is one of several methods used by plants to promote outcrossing. in many land plants the male and female gametes are produced by separate individuals. these species are said to be dioecious when referring to vascular plant sporophytes and dioicous when referring to bryophyte gametophytes. charles darwin in his 1878 book the effects of cross and self - fertilization in the vegetable kingdom at the start of chapter xii noted " the first and most important of the conclusions which may be drawn from the observations given in this volume, is that generally cross - fertilisation is beneficial and self - fertilisation often injurious, at least with the plants on which i experimented. " an important adaptive benefit of outcrossing is that it allows the masking of deleterious mutations in the genome of progeny. this beneficial effect is also known as hybrid vigor or heterosis. once outcrossing is established, subsequent switching to inbreeding becomes disadvantageous since it allows expression of the previously masked deleterious recessive mutations, commonly referred to as inbreeding depression. unlike in higher animals, where parthenogenesis is rare, asexual reproduction may occur in plants by several different mechanisms. the formation of stem tubers in potato is one example. particularly in arctic or alpine habitats, where opportunities for fertilisation of flowers by animals are rare, plantlets or bulbs, may develop instead of flowers, replacing sexual reproduction with asexual reproduction and giving rise to clonal populations genetically identical to the parent. this is one
from the oil of jasminum grandiflorum which regulates wound responses in plants by unblocking the expression of genes required in the systemic acquired resistance response to pathogen attack. in addition to being the primary energy source for plants, light functions as a signalling device, providing information to the plant, such as how much sunlight the plant receives each day. this can result in adaptive changes in a process known as photomorphogenesis. phytochromes are the photoreceptors in a plant that are sensitive to light. = = plant anatomy and morphology = = plant anatomy is the study of the structure of plant cells and tissues, whereas plant morphology is the study of their external form. all plants are multicellular eukaryotes, their dna stored in nuclei. the characteristic features of plant cells that distinguish them from those of animals and fungi include a primary cell wall composed of the polysaccharides cellulose, hemicellulose and pectin, larger vacuoles than in animal cells and the presence of plastids with unique photosynthetic and biosynthetic functions as in the chloroplasts. other plastids contain storage products such as starch ( amyloplasts ) or lipids ( elaioplasts ). uniquely, streptophyte cells and those of the green algal order trentepohliales divide by construction of a phragmoplast as a template for building a cell plate late in cell division. the bodies of vascular plants including clubmosses, ferns and seed plants ( gymnosperms and angiosperms ) generally have aerial and subterranean subsystems. the shoots consist of stems bearing green photosynthesising leaves and reproductive structures. the underground vascularised roots bear root hairs at their tips and generally lack chlorophyll. non - vascular plants, the liverworts, hornworts and mosses do not produce ground - penetrating vascular roots and most of the plant participates in photosynthesis. the sporophyte generation is nonphotosynthetic in liverworts but may be able to contribute part of its energy needs by photosynthesis in mosses and hornworts. the root system and the shoot system are interdependent β the usually nonphotosynthetic root system depends on the shoot system for food, and the usually photosynthetic shoot system depends on water and minerals from the root system. cells in each system are capable
stems mainly provide support to the leaves and reproductive structures, but can store water in succulent plants such as cacti, food as in potato tubers, or reproduce vegetatively as in the stolons of strawberry plants or in the process of layering. leaves gather sunlight and carry out photosynthesis. large, flat, flexible, green leaves are called foliage leaves. gymnosperms, such as conifers, cycads, ginkgo, and gnetophytes are seed - producing plants with open seeds. angiosperms are seed - producing plants that produce flowers and have enclosed seeds. woody plants, such as azaleas and oaks, undergo a secondary growth phase resulting in two additional types of tissues : wood ( secondary xylem ) and bark ( secondary phloem and cork ). all gymnosperms and many angiosperms are woody plants. some plants reproduce sexually, some asexually, and some via both means. although reference to major morphological categories such as root, stem, leaf, and trichome are useful, one has to keep in mind that these categories are linked through intermediate forms so that a continuum between the categories results. furthermore, structures can be seen as processes, that is, process combinations. = = systematic botany = = systematic botany is part of systematic biology, which is concerned with the range and diversity of organisms and their relationships, particularly as determined by their evolutionary history. it involves, or is related to, biological classification, scientific taxonomy and phylogenetics. biological classification is the method by which botanists group organisms into categories such as genera or species. biological classification is a form of scientific taxonomy. modern taxonomy is rooted in the work of carl linnaeus, who grouped species according to shared physical characteristics. these groupings have since been revised to align better with the darwinian principle of common descent β grouping organisms by ancestry rather than superficial characteristics. while scientists do not always agree on how to classify organisms, molecular phylogenetics, which uses dna sequences as data, has driven many recent revisions along evolutionary lines and is likely to continue to do so. the dominant classification system is called linnaean taxonomy. it includes ranks and binomial nomenclature. the nomenclature of botanical organisms is codified in the international code of nomenclature for algae, fungi, and plants ( icn ) and administered by the international botanical congress. kingdom plantae belongs to domain eukaryota and is broken down recursively until each species is separately classified. the order is :
##trophs including all animals, all fungi, all completely parasitic plants, and non - photosynthetic bacteria take in organic molecules produced by photoautotrophs and respire them or use them in the construction of cells and tissues. respiration is the oxidation of carbon compounds by breaking them down into simpler structures to release the energy they contain, essentially the opposite of photosynthesis. molecules are moved within plants by transport processes that operate at a variety of spatial scales. subcellular transport of ions, electrons and molecules such as water and enzymes occurs across cell membranes. minerals and water are transported from roots to other parts of the plant in the transpiration stream. diffusion, osmosis, and active transport and mass flow are all different ways transport can occur. examples of elements that plants need to transport are nitrogen, phosphorus, potassium, calcium, magnesium, and sulfur. in vascular plants, these elements are extracted from the soil as soluble ions by the roots and transported throughout the plant in the xylem. most of the elements required for plant nutrition come from the chemical breakdown of soil minerals. sucrose produced by photosynthesis is transported from the leaves to other parts of the plant in the phloem and plant hormones are transported by a variety of processes. = = = plant hormones = = = plants are not passive, but respond to external signals such as light, touch, and injury by moving or growing towards or away from the stimulus, as appropriate. tangible evidence of touch sensitivity is the almost instantaneous collapse of leaflets of mimosa pudica, the insect traps of venus flytrap and bladderworts, and the pollinia of orchids. the hypothesis that plant growth and development is coordinated by plant hormones or plant growth regulators first emerged in the late 19th century. darwin experimented on the movements of plant shoots and roots towards light and gravity, and concluded " it is hardly an exaggeration to say that the tip of the radicle.. acts like the brain of one of the lower animals.. directing the several movements ". about the same time, the role of auxins ( from the greek auxein, to grow ) in control of plant growth was first outlined by the dutch scientist frits went. the first known auxin, indole - 3 - acetic acid ( iaa ), which promotes cell growth, was only isolated from plants about 50 years later. this compound mediates the tropic responses of shoots and roots towards light and gravity. the finding in 1939 that plant callus
smallest genomes among flowering plants. arabidopsis was the first plant to have its genome sequenced, in 2000. the sequencing of some other relatively small genomes, of rice ( oryza sativa ) and brachypodium distachyon, has made them important model species for understanding the genetics, cellular and molecular biology of cereals, grasses and monocots generally. model plants such as arabidopsis thaliana are used for studying the molecular biology of plant cells and the chloroplast. ideally, these organisms have small genomes that are well known or completely sequenced, small stature and short generation times. corn has been used to study mechanisms of photosynthesis and phloem loading of sugar in c4 plants. the single celled green alga chlamydomonas reinhardtii, while not an embryophyte itself, contains a green - pigmented chloroplast related to that of land plants, making it useful for study. a red alga cyanidioschyzon merolae has also been used to study some basic chloroplast functions. spinach, peas, soybeans and a moss physcomitrella patens are commonly used to study plant cell biology. agrobacterium tumefaciens, a soil rhizosphere bacterium, can attach to plant cells and infect them with a callus - inducing ti plasmid by horizontal gene transfer, causing a callus infection called crown gall disease. schell and van montagu ( 1977 ) hypothesised that the ti plasmid could be a natural vector for introducing the nif gene responsible for nitrogen fixation in the root nodules of legumes and other plant species. today, genetic modification of the ti plasmid is one of the main techniques for introduction of transgenes to plants and the creation of genetically modified crops. = = = epigenetics = = = epigenetics is the study of heritable changes in gene function that cannot be explained by changes in the underlying dna sequence but cause the organism ' s genes to behave ( or " express themselves " ) differently. one example of epigenetic change is the marking of the genes by dna methylation which determines whether they will be expressed or not. gene expression can also be controlled by repressor proteins that attach to silencer regions of the dna and prevent that region of the dna code from being expressed. epigenetic marks may be added
Question: In flowering plants, what structure containing DNA is transported from one plant to another?
A) nectar
B) chlorophyll
C) glucose
D) pollen
|
D) pollen
|
Context:
and bad nvh qualities. the nvh engineer works to either eliminate bad nvh or change the " bad nvh " to good ( i. e., exhaust tones ). vehicle electronics : automotive electronics is an increasingly important aspect of automotive engineering. modern vehicles employ dozens of electronic systems. these systems are responsible for operational controls such as the throttle, brake and steering controls ; as well as many comfort - and - convenience systems such as the hvac, infotainment, and lighting systems. it would not be possible for automobiles to meet modern safety and fuel - economy requirements without electronic controls. performance : performance is a measurable and testable value of a vehicle ' s ability to perform in various conditions. performance can be considered in a wide variety of tasks, but it generally considers how quickly a car can accelerate ( e. g. standing start 1 / 4 mile elapsed time, 0 β 60 mph, etc. ), its top speed, how short and quickly a car can come to a complete stop from a set speed ( e. g. 70 - 0 mph ), how much g - force a car can generate without losing grip, recorded lap - times, cornering speed, brake fade, etc. performance can also reflect the amount of control in inclement weather ( snow, ice, rain ). shift quality : shift quality is the driver ' s perception of the vehicle to an automatic transmission shift event. this is influenced by the powertrain ( internal combustion engine, transmission ), and the vehicle ( driveline, suspension, engine and powertrain mounts, etc. ) shift feel is both a tactile ( felt ) and audible ( heard ) response of the vehicle. shift quality is experienced as various events : transmission shifts are felt as an upshift at acceleration ( 1 β 2 ), or a downshift maneuver in passing ( 4 β 2 ). shift engagements of the vehicle are also evaluated, as in park to reverse, etc. durability / corrosion engineering : durability and corrosion engineering is the evaluation testing of a vehicle for its useful life. tests include mileage accumulation, severe driving conditions, and corrosive salt baths. drivability : drivability is the vehicle ' s response to general driving conditions. cold starts and stalls, rpm dips, idle response, launch hesitations and stumbles, and performance levels all contribute to the overall drivability of any given vehicle. cost : the cost of a vehicle program is typically split into the effect
defective body parts. inside the body, artificial heart valves are in common use with artificial hearts and lungs seeing less common use but under active technology development. other medical devices and aids that can be considered prosthetics include hearing aids, artificial eyes, palatal obturator, gastric bands, and dentures. prostheses are specifically not orthoses, although given certain circumstances a prosthesis might end up performing some or all of the same functionary benefits as an orthosis. prostheses are technically the complete finished item. for instance, a c - leg knee alone is not a prosthesis, but only a prosthetic component. the complete prosthesis would consist of the attachment system to the residual limb β usually a " socket ", and all the attachment hardware components all the way down to and including the terminal device. despite the technical difference, the terms are often used interchangeably. the terms " prosthetic " and " orthotic " are adjectives used to describe devices such as a prosthetic knee. the terms " prosthetics " and " orthotics " are used to describe the respective allied health fields. an occupational therapist ' s role in prosthetics include therapy, training and evaluations. prosthetic training includes orientation to prosthetics components and terminology, donning and doffing, wearing schedule, and how to care for residual limb and the prosthesis. = = = exoskeletons = = = a powered exoskeleton is a wearable mobile machine that is powered by a system of electric motors, pneumatics, levers, hydraulics, or a combination of technologies that allow for limb movement with increased strength and endurance. its design aims to provide back support, sense the user ' s motion, and send a signal to motors which manage the gears. the exoskeleton supports the shoulder, waist and thigh, and assists movement for lifting and holding heavy items, while lowering back stress. = = = adaptive seating and positioning = = = people with balance and motor function challenges often need specialized equipment to sit or stand safely and securely. this equipment is frequently specialized for specific settings such as in a classroom or nursing home. positioning is often important in seating arrangements to ensure that user ' s body pressure is distributed equally without inhibiting movement in a desired way. positioning devices have been developed to aid in allowing people to stand and bear weight on their legs without risk of a fall.
response can be seat vibration or a buzz in the steering wheel. this feedback is generated by components either rubbing, vibrating, or rotating. nvh response can be classified in various ways : powertrain nvh, road noise, wind noise, component noise, and squeak and rattle. note, there are both good and bad nvh qualities. the nvh engineer works to either eliminate bad nvh or change the " bad nvh " to good ( i. e., exhaust tones ). vehicle electronics : automotive electronics is an increasingly important aspect of automotive engineering. modern vehicles employ dozens of electronic systems. these systems are responsible for operational controls such as the throttle, brake and steering controls ; as well as many comfort - and - convenience systems such as the hvac, infotainment, and lighting systems. it would not be possible for automobiles to meet modern safety and fuel - economy requirements without electronic controls. performance : performance is a measurable and testable value of a vehicle ' s ability to perform in various conditions. performance can be considered in a wide variety of tasks, but it generally considers how quickly a car can accelerate ( e. g. standing start 1 / 4 mile elapsed time, 0 β 60 mph, etc. ), its top speed, how short and quickly a car can come to a complete stop from a set speed ( e. g. 70 - 0 mph ), how much g - force a car can generate without losing grip, recorded lap - times, cornering speed, brake fade, etc. performance can also reflect the amount of control in inclement weather ( snow, ice, rain ). shift quality : shift quality is the driver ' s perception of the vehicle to an automatic transmission shift event. this is influenced by the powertrain ( internal combustion engine, transmission ), and the vehicle ( driveline, suspension, engine and powertrain mounts, etc. ) shift feel is both a tactile ( felt ) and audible ( heard ) response of the vehicle. shift quality is experienced as various events : transmission shifts are felt as an upshift at acceleration ( 1 β 2 ), or a downshift maneuver in passing ( 4 β 2 ). shift engagements of the vehicle are also evaluated, as in park to reverse, etc. durability / corrosion engineering : durability and corrosion engineering is the evaluation testing of a vehicle for its useful life. tests include mileage accumulation, severe driving conditions, and corrosive salt baths. drivability
material. silicon nitride parts are used in ceramic ball bearings. their higher hardness means that they are much less susceptible to wear and can offer more than triple lifetimes. they also deform less under load meaning they have less contact with the bearing retainer walls and can roll faster. in very high speed applications, heat from friction during rolling can cause problems for metal bearings ; problems which are reduced by the use of ceramics. ceramics are also more chemically resistant and can be used in wet environments where steel bearings would rust. the major drawback to using ceramics is a significantly higher cost. in many cases their electrically insulating properties may also be valuable in bearings. in the early 1980s, toyota researched production of an adiabatic ceramic engine which can run at a temperature of over 6000 Β°f ( 3300 Β°c ). ceramic engines do not require a cooling system and hence allow a major weight reduction and therefore greater fuel efficiency. fuel efficiency of the engine is also higher at high temperature, as shown by carnot ' s theorem. in a conventional metallic engine, much of the energy released from the fuel must be dissipated as waste heat in order to prevent a meltdown of the metallic parts. despite all of these desirable properties, such engines are not in production because the manufacturing of ceramic parts in the requisite precision and durability is difficult. imperfection in the ceramic leads to cracks, which can lead to potentially dangerous equipment failure. such engines are possible in laboratory settings, but mass - production is not feasible with current technology. work is being done in developing ceramic parts for gas turbine engines. currently, even blades made of advanced metal alloys used in the engines ' hot section require cooling and careful limiting of operating temperatures. turbine engines made with ceramics could operate more efficiently, giving aircraft greater range and payload for a set amount of fuel. recently, there have been advances in ceramics which include bio - ceramics, such as dental implants and synthetic bones. hydroxyapatite, the natural mineral component of bone, has been made synthetically from a number of biological and chemical sources and can be formed into ceramic materials. orthopedic implants made from these materials bond readily to bone and other tissues in the body without rejection or inflammatory reactions. because of this, they are of great interest for gene delivery and tissue engineering scaffolds. most hydroxyapatite ceramics are very porous and lack mechanical strength and are used to coat metal orthopedic devices to aid in forming a bond to bone or as bone fillers. they are
be used at high latitudes because of terrestrial interference. cordless phone β a landline telephone in which the handset is portable and communicates with the rest of the phone by a short - range full duplex radio link, instead of being attached by a cord. both the handset and the base station have low - power radio transceivers that handle the short - range bidirectional radio link. as of 2022, cordless phones in most nations use the dect transmission standard. land mobile radio system β short - range mobile or portable half - duplex radio transceivers operating in the vhf or uhf band that can be used without a license. they are often installed in vehicles, with the mobile units communicating with a dispatcher at a fixed base station. special systems with reserved frequencies are used by first responder services ; police, fire, ambulance, and emergency services, and other government services. other systems are made for use by commercial firms such as taxi and delivery services. vhf systems use channels in the range 30 β 50 mhz and 150 β 172 mhz. uhf systems use the 450 β 470 mhz band and in some areas the 470 β 512 mhz range. in general, vhf systems have a longer range than uhf but require longer antennas. am or fm modulation is mainly used, but digital systems such as dmr are being introduced. the radiated power is typically limited to 4 watts. these systems have a fairly limited range, usually 3 to 20 miles ( 4. 8 to 32 km ) depending on terrain. repeaters installed on tall buildings, hills, or mountain peaks are often used to increase the range when it is desired to cover a larger area than line - of - sight. examples of land mobile systems are cb, frs, gmrs, and murs. modern digital systems, called trunked radio systems, have a digital channel management system using a control channel that automatically assigns frequency channels to user groups. walkie - talkie β a battery - powered portable handheld half - duplex two - way radio, used in land mobile radio systems. airband β half - duplex radio system used by aircraft pilots to talk to other aircraft and ground - based air traffic controllers. this vital system is the main communication channel for air traffic control. for most communication in overland flights in air corridors a vhf - am system using channels between 108 and 137 mhz in the vhf band is used. this system has a typical transmission range of 200 miles ( 320 km ) for aircraft flying at cruising altitude. for flights in more
and evaporative emissions. nvh engineering ( noise, vibration, and harshness ) : nvh involves customer feedback ( both tactile [ felt ] and audible [ heard ] ) concerning a vehicle. while sound can be interpreted as a rattle, squeal, or hot, a tactile response can be seat vibration or a buzz in the steering wheel. this feedback is generated by components either rubbing, vibrating, or rotating. nvh response can be classified in various ways : powertrain nvh, road noise, wind noise, component noise, and squeak and rattle. note, there are both good and bad nvh qualities. the nvh engineer works to either eliminate bad nvh or change the " bad nvh " to good ( i. e., exhaust tones ). vehicle electronics : automotive electronics is an increasingly important aspect of automotive engineering. modern vehicles employ dozens of electronic systems. these systems are responsible for operational controls such as the throttle, brake and steering controls ; as well as many comfort - and - convenience systems such as the hvac, infotainment, and lighting systems. it would not be possible for automobiles to meet modern safety and fuel - economy requirements without electronic controls. performance : performance is a measurable and testable value of a vehicle ' s ability to perform in various conditions. performance can be considered in a wide variety of tasks, but it generally considers how quickly a car can accelerate ( e. g. standing start 1 / 4 mile elapsed time, 0 β 60 mph, etc. ), its top speed, how short and quickly a car can come to a complete stop from a set speed ( e. g. 70 - 0 mph ), how much g - force a car can generate without losing grip, recorded lap - times, cornering speed, brake fade, etc. performance can also reflect the amount of control in inclement weather ( snow, ice, rain ). shift quality : shift quality is the driver ' s perception of the vehicle to an automatic transmission shift event. this is influenced by the powertrain ( internal combustion engine, transmission ), and the vehicle ( driveline, suspension, engine and powertrain mounts, etc. ) shift feel is both a tactile ( felt ) and audible ( heard ) response of the vehicle. shift quality is experienced as various events : transmission shifts are felt as an upshift at acceleration ( 1 β 2 ), or a downshift maneuver in passing ( 4 β 2
##tronics, the science of using mechanical devices with human muscular, musculoskeletal, and nervous systems to assist or enhance motor control lost by trauma, disease, or defect. prostheses are typically used to replace parts lost by injury ( traumatic ) or missing from birth ( congenital ) or to supplement defective body parts. inside the body, artificial heart valves are in common use with artificial hearts and lungs seeing less common use but under active technology development. other medical devices and aids that can be considered prosthetics include hearing aids, artificial eyes, palatal obturator, gastric bands, and dentures. prostheses are specifically not orthoses, although given certain circumstances a prosthesis might end up performing some or all of the same functionary benefits as an orthosis. prostheses are technically the complete finished item. for instance, a c - leg knee alone is not a prosthesis, but only a prosthetic component. the complete prosthesis would consist of the attachment system to the residual limb β usually a " socket ", and all the attachment hardware components all the way down to and including the terminal device. despite the technical difference, the terms are often used interchangeably. the terms " prosthetic " and " orthotic " are adjectives used to describe devices such as a prosthetic knee. the terms " prosthetics " and " orthotics " are used to describe the respective allied health fields. an occupational therapist ' s role in prosthetics include therapy, training and evaluations. prosthetic training includes orientation to prosthetics components and terminology, donning and doffing, wearing schedule, and how to care for residual limb and the prosthesis. = = = exoskeletons = = = a powered exoskeleton is a wearable mobile machine that is powered by a system of electric motors, pneumatics, levers, hydraulics, or a combination of technologies that allow for limb movement with increased strength and endurance. its design aims to provide back support, sense the user ' s motion, and send a signal to motors which manage the gears. the exoskeleton supports the shoulder, waist and thigh, and assists movement for lifting and holding heavy items, while lowering back stress. = = = adaptive seating and positioning = = = people with balance and motor function challenges often need specialized equipment to sit or stand safely and securely. this equipment is frequently
, heat from friction during rolling can cause problems for metal bearings ; problems which are reduced by the use of ceramics. ceramics are also more chemically resistant and can be used in wet environments where steel bearings would rust. the major drawback to using ceramics is a significantly higher cost. in many cases their electrically insulating properties may also be valuable in bearings. in the early 1980s, toyota researched production of an adiabatic ceramic engine which can run at a temperature of over 6000 Β°f ( 3300 Β°c ). ceramic engines do not require a cooling system and hence allow a major weight reduction and therefore greater fuel efficiency. fuel efficiency of the engine is also higher at high temperature, as shown by carnot ' s theorem. in a conventional metallic engine, much of the energy released from the fuel must be dissipated as waste heat in order to prevent a meltdown of the metallic parts. despite all of these desirable properties, such engines are not in production because the manufacturing of ceramic parts in the requisite precision and durability is difficult. imperfection in the ceramic leads to cracks, which can lead to potentially dangerous equipment failure. such engines are possible in laboratory settings, but mass - production is not feasible with current technology. work is being done in developing ceramic parts for gas turbine engines. currently, even blades made of advanced metal alloys used in the engines ' hot section require cooling and careful limiting of operating temperatures. turbine engines made with ceramics could operate more efficiently, giving aircraft greater range and payload for a set amount of fuel. recently, there have been advances in ceramics which include bio - ceramics, such as dental implants and synthetic bones. hydroxyapatite, the natural mineral component of bone, has been made synthetically from a number of biological and chemical sources and can be formed into ceramic materials. orthopedic implants made from these materials bond readily to bone and other tissues in the body without rejection or inflammatory reactions. because of this, they are of great interest for gene delivery and tissue engineering scaffolds. most hydroxyapatite ceramics are very porous and lack mechanical strength and are used to coat metal orthopedic devices to aid in forming a bond to bone or as bone fillers. they are also used as fillers for orthopedic plastic screws to aid in reducing the inflammation and increase absorption of these plastic materials. work is being done to make strong, fully dense nano crystalline hydroxyapatite ceramic materials for orthopedic weight bearing devices, replacing foreign metal and plastic orthopedic materials
more recently by the civilian world for news reporting and aerial photography. the pilot uses aircraft controls like a joystick or steering wheel, which create control signals which are transmitted to the drone by radio to control the flight surfaces and engine. a telemetry system transmits back a video image from a camera in the drone to allow the pilot to see where the aircraft is going, and data from a gps receiver giving the real - time position of the aircraft. uavs have sophisticated onboard automatic pilot systems that maintain stable flight and only require manual control to change directions. keyless entry system β a short - range handheld battery powered key fob transmitter, included with most modern cars, which can lock and unlock the doors of a vehicle from outside, eliminating the need to use a key. when a button is pressed, the transmitter sends a coded radio signal to a receiver in the vehicle, operating the locks. the fob must be close to the vehicle, typically within 5 to 20 meters. north america and japan use a frequency of 315 mhz, while europe uses 433. 92 and 868 mhz. some models can also remotely start the engine, to warm up the car. a security concern with all keyless entry systems is a replay attack, in which a thief uses a special receiver ( " code grabber " ) to record the radio signal during opening, which can later be replayed to open the door. to prevent this, keyless systems use a rolling code system in which a pseudorandom number generator in the remote control generates a different random key each time it is used. to prevent thieves from simulating the pseudorandom generator to calculate the next key, the radio signal is also encrypted. garage door opener β a short - range handheld transmitter which can open or close a building ' s electrically operated garage door from outside, so the owner can open the door upon arrival, and close it after departure. when a button is pressed the control transmits a coded fsk radio signal to a receiver in the opener, raising or lowering the door. modern openers use 310, 315 or 390 mhz. to prevent a thief using a replay attack, modern openers use a rolling code system. radio - controlled models β a popular hobby is playing with radio - controlled model boats, cars, airplanes, and helicopters ( quadcopters ) which are controlled by radio signals from a handheld console with a joystick. most recent transmitters use the 2. 4 ghz ism band with multiple control channels modulated with pwm, pc
which offer more ergonomic layouts of the keys. assistive technology devices have been created to enable disabled people to use modern touch screen mobile computers such as the ipad, iphone and ipod touch. the pererro is a plug and play adapter for ios devices which uses the built in apple voiceover feature in combination with a basic switch. this brings touch screen technology to those who were previously unable to use it. apple, with the release of ios 7 had introduced the ability to navigate apps using switch control. switch access could be activated either through an external bluetooth connected switch, single touch of the screen, or use of right and left head turns using the device ' s camera. additional accessibility features include the use of assistive touch which allows a user to access multi - touch gestures through pre - programmed onscreen buttons. for users with physical disabilities a large variety of switches are available and customizable to the user ' s needs varying in size, shape, or amount of pressure required for activation. switch access may be placed near any area of the body which has consistent and reliable mobility and less subject to fatigue. common sites include the hands, head, and feet. eye gaze and head mouse systems can also be used as an alternative mouse navigation. a user may use single or multiple switch sites and the process often involves a scanning through items on a screen and activating the switch once the desired object is highlighted. = = home automation = = the form of home automation called assistive domotics focuses on making it possible for elderly and disabled people to live independently. home automation is becoming a viable option for the elderly and disabled who would prefer to stay in their own homes rather than move to a healthcare facility. this field uses much of the same technology and equipment as home automation for security, entertainment, and energy conservation but tailors it towards elderly and disabled users. for example, automated prompts and reminders use motion sensors and pre - recorded audio messages ; an automated prompt in the kitchen may remind the resident to turn off the oven, and one by the front door may remind the resident to lock the door. = = assistive technology and innovation = = innovation is happening in assistive technology either through improvements to existing devices or the creation of new products. in the wipo published 2021 report on technology trends, assistive products are grouped into either conventional or emerging technologies. conventional assisting technology tracks innovation within well - established assistive products, whereas emerging assistive technology refers to more advanced products. these identified advanced assistive products are distinguished from the
Question: Several interchangeable parts are used in the manufacture of automobiles. Some examples of these parts are batteries, windshield wiper blades, spark plugs, and tires. Which of the following is an advantage of interchangeable parts over non-interchangeable parts on automobiles?
A) Interchangeable parts are generally more expensive than custom-made parts for automobiles.
B) Interchangeable parts break more often now than in years past on most automobiles.
C) Interchangeable parts are plentiful for custom-made automobiles.
D) Interchangeable parts can fit many kinds of automobiles.
|
D) Interchangeable parts can fit many kinds of automobiles.
|
Context:
is said to have occurred. a chemical reaction is therefore a concept related to the " reaction " of a substance when it comes in close contact with another, whether as a mixture or a solution ; exposure to some form of energy, or both. it results in some energy exchange between the constituents of the reaction as well as with the system environment, which may be designed vessels β often laboratory glassware. chemical reactions can result in the formation or dissociation of molecules, that is, molecules breaking apart to form two or more molecules or rearrangement of atoms within or across molecules. chemical reactions usually involve the making or breaking of chemical bonds. oxidation, reduction, dissociation, acid β base neutralization and molecular rearrangement are some examples of common chemical reactions. a chemical reaction can be symbolically depicted through a chemical equation. while in a non - nuclear chemical reaction the number and kind of atoms on both sides of the equation are equal, for a nuclear reaction this holds true only for the nuclear particles viz. protons and neutrons. the sequence of steps in which the reorganization of chemical bonds may be taking place in the course of a chemical reaction is called its mechanism. a chemical reaction can be envisioned to take place in a number of steps, each of which may have a different speed. many reaction intermediates with variable stability can thus be envisaged during the course of a reaction. reaction mechanisms are proposed to explain the kinetics and the relative product mix of a reaction. many physical chemists specialize in exploring and proposing the mechanisms of various chemical reactions. several empirical rules, like the woodward β hoffmann rules often come in handy while proposing a mechanism for a chemical reaction. according to the iupac gold book, a chemical reaction is " a process that results in the interconversion of chemical species. " accordingly, a chemical reaction may be an elementary reaction or a stepwise reaction. an additional caveat is made, in that this definition includes cases where the interconversion of conformers is experimentally observable. such detectable chemical reactions normally involve sets of molecular entities as indicated by this definition, but it is often conceptually convenient to use the term also for changes involving single molecular entities ( i. e. ' microscopic chemical events ' ). = = = ions and salts = = = an ion is a charged species, an atom or a molecule, that has lost or gained one or more electrons. when an atom loses an electron and thus has more protons than electrons, the atom is a positively charged
analyzing their radiation spectra. the term chemical energy is often used to indicate the potential of a chemical substance to undergo a transformation through a chemical reaction or to transform other chemical substances. = = = reaction = = = when a chemical substance is transformed as a result of its interaction with another substance or with energy, a chemical reaction is said to have occurred. a chemical reaction is therefore a concept related to the " reaction " of a substance when it comes in close contact with another, whether as a mixture or a solution ; exposure to some form of energy, or both. it results in some energy exchange between the constituents of the reaction as well as with the system environment, which may be designed vessels β often laboratory glassware. chemical reactions can result in the formation or dissociation of molecules, that is, molecules breaking apart to form two or more molecules or rearrangement of atoms within or across molecules. chemical reactions usually involve the making or breaking of chemical bonds. oxidation, reduction, dissociation, acid β base neutralization and molecular rearrangement are some examples of common chemical reactions. a chemical reaction can be symbolically depicted through a chemical equation. while in a non - nuclear chemical reaction the number and kind of atoms on both sides of the equation are equal, for a nuclear reaction this holds true only for the nuclear particles viz. protons and neutrons. the sequence of steps in which the reorganization of chemical bonds may be taking place in the course of a chemical reaction is called its mechanism. a chemical reaction can be envisioned to take place in a number of steps, each of which may have a different speed. many reaction intermediates with variable stability can thus be envisaged during the course of a reaction. reaction mechanisms are proposed to explain the kinetics and the relative product mix of a reaction. many physical chemists specialize in exploring and proposing the mechanisms of various chemical reactions. several empirical rules, like the woodward β hoffmann rules often come in handy while proposing a mechanism for a chemical reaction. according to the iupac gold book, a chemical reaction is " a process that results in the interconversion of chemical species. " accordingly, a chemical reaction may be an elementary reaction or a stepwise reaction. an additional caveat is made, in that this definition includes cases where the interconversion of conformers is experimentally observable. such detectable chemical reactions normally involve sets of molecular entities as indicated by this definition, but it is often conceptually convenient to use the term also for changes involving single molecular entities (
with the system environment, which may be designed vessels β often laboratory glassware. chemical reactions can result in the formation or dissociation of molecules, that is, molecules breaking apart to form two or more molecules or rearrangement of atoms within or across molecules. chemical reactions usually involve the making or breaking of chemical bonds. oxidation, reduction, dissociation, acid β base neutralization and molecular rearrangement are some examples of common chemical reactions. a chemical reaction can be symbolically depicted through a chemical equation. while in a non - nuclear chemical reaction the number and kind of atoms on both sides of the equation are equal, for a nuclear reaction this holds true only for the nuclear particles viz. protons and neutrons. the sequence of steps in which the reorganization of chemical bonds may be taking place in the course of a chemical reaction is called its mechanism. a chemical reaction can be envisioned to take place in a number of steps, each of which may have a different speed. many reaction intermediates with variable stability can thus be envisaged during the course of a reaction. reaction mechanisms are proposed to explain the kinetics and the relative product mix of a reaction. many physical chemists specialize in exploring and proposing the mechanisms of various chemical reactions. several empirical rules, like the woodward β hoffmann rules often come in handy while proposing a mechanism for a chemical reaction. according to the iupac gold book, a chemical reaction is " a process that results in the interconversion of chemical species. " accordingly, a chemical reaction may be an elementary reaction or a stepwise reaction. an additional caveat is made, in that this definition includes cases where the interconversion of conformers is experimentally observable. such detectable chemical reactions normally involve sets of molecular entities as indicated by this definition, but it is often conceptually convenient to use the term also for changes involving single molecular entities ( i. e. ' microscopic chemical events ' ). = = = ions and salts = = = an ion is a charged species, an atom or a molecule, that has lost or gained one or more electrons. when an atom loses an electron and thus has more protons than electrons, the atom is a positively charged ion or cation. when an atom gains an electron and thus has more electrons than protons, the atom is a negatively charged ion or anion. cations and anions can form a crystalline lattice of neutral salts, such as the na + and clβ ions forming sodium chloride, or nacl. examples of
. oxidation, reduction, dissociation, acid β base neutralization and molecular rearrangement are some examples of common chemical reactions. a chemical reaction can be symbolically depicted through a chemical equation. while in a non - nuclear chemical reaction the number and kind of atoms on both sides of the equation are equal, for a nuclear reaction this holds true only for the nuclear particles viz. protons and neutrons. the sequence of steps in which the reorganization of chemical bonds may be taking place in the course of a chemical reaction is called its mechanism. a chemical reaction can be envisioned to take place in a number of steps, each of which may have a different speed. many reaction intermediates with variable stability can thus be envisaged during the course of a reaction. reaction mechanisms are proposed to explain the kinetics and the relative product mix of a reaction. many physical chemists specialize in exploring and proposing the mechanisms of various chemical reactions. several empirical rules, like the woodward β hoffmann rules often come in handy while proposing a mechanism for a chemical reaction. according to the iupac gold book, a chemical reaction is " a process that results in the interconversion of chemical species. " accordingly, a chemical reaction may be an elementary reaction or a stepwise reaction. an additional caveat is made, in that this definition includes cases where the interconversion of conformers is experimentally observable. such detectable chemical reactions normally involve sets of molecular entities as indicated by this definition, but it is often conceptually convenient to use the term also for changes involving single molecular entities ( i. e. ' microscopic chemical events ' ). = = = ions and salts = = = an ion is a charged species, an atom or a molecule, that has lost or gained one or more electrons. when an atom loses an electron and thus has more protons than electrons, the atom is a positively charged ion or cation. when an atom gains an electron and thus has more electrons than protons, the atom is a negatively charged ion or anion. cations and anions can form a crystalline lattice of neutral salts, such as the na + and clβ ions forming sodium chloride, or nacl. examples of polyatomic ions that do not split up during acid β base reactions are hydroxide ( ohβ ) and phosphate ( po43β ). plasma is composed of gaseous matter that has been completely ionized, usually through high temperature. = = = acidity and basicity = = = a substance can often be
a nuclear reaction this holds true only for the nuclear particles viz. protons and neutrons. the sequence of steps in which the reorganization of chemical bonds may be taking place in the course of a chemical reaction is called its mechanism. a chemical reaction can be envisioned to take place in a number of steps, each of which may have a different speed. many reaction intermediates with variable stability can thus be envisaged during the course of a reaction. reaction mechanisms are proposed to explain the kinetics and the relative product mix of a reaction. many physical chemists specialize in exploring and proposing the mechanisms of various chemical reactions. several empirical rules, like the woodward β hoffmann rules often come in handy while proposing a mechanism for a chemical reaction. according to the iupac gold book, a chemical reaction is " a process that results in the interconversion of chemical species. " accordingly, a chemical reaction may be an elementary reaction or a stepwise reaction. an additional caveat is made, in that this definition includes cases where the interconversion of conformers is experimentally observable. such detectable chemical reactions normally involve sets of molecular entities as indicated by this definition, but it is often conceptually convenient to use the term also for changes involving single molecular entities ( i. e. ' microscopic chemical events ' ). = = = ions and salts = = = an ion is a charged species, an atom or a molecule, that has lost or gained one or more electrons. when an atom loses an electron and thus has more protons than electrons, the atom is a positively charged ion or cation. when an atom gains an electron and thus has more electrons than protons, the atom is a negatively charged ion or anion. cations and anions can form a crystalline lattice of neutral salts, such as the na + and clβ ions forming sodium chloride, or nacl. examples of polyatomic ions that do not split up during acid β base reactions are hydroxide ( ohβ ) and phosphate ( po43β ). plasma is composed of gaseous matter that has been completely ionized, usually through high temperature. = = = acidity and basicity = = = a substance can often be classified as an acid or a base. there are several different theories which explain acid β base behavior. the simplest is arrhenius theory, which states that an acid is a substance that produces hydronium ions when it is dissolved in water, and a base is one that produces hydroxide ions when dissolved in water.
are studied in chemistry are usually the result of interactions between atoms, leading to rearrangements of the chemical bonds which hold atoms together. such behaviors are studied in a chemistry laboratory. the chemistry laboratory stereotypically uses various forms of laboratory glassware. however glassware is not central to chemistry, and a great deal of experimental ( as well as applied / industrial ) chemistry is done without it. a chemical reaction is a transformation of some substances into one or more different substances. the basis of such a chemical transformation is the rearrangement of electrons in the chemical bonds between atoms. it can be symbolically depicted through a chemical equation, which usually involves atoms as subjects. the number of atoms on the left and the right in the equation for a chemical transformation is equal. ( when the number of atoms on either side is unequal, the transformation is referred to as a nuclear reaction or radioactive decay. ) the type of chemical reactions a substance may undergo and the energy changes that may accompany it are constrained by certain basic rules, known as chemical laws. energy and entropy considerations are invariably important in almost all chemical studies. chemical substances are classified in terms of their structure, phase, as well as their chemical compositions. they can be analyzed using the tools of chemical analysis, e. g. spectroscopy and chromatography. scientists engaged in chemical research are known as chemists. most chemists specialize in one or more sub - disciplines. several concepts are essential for the study of chemistry ; some of them are : = = = matter = = = in chemistry, matter is defined as anything that has rest mass and volume ( it takes up space ) and is made up of particles. the particles that make up matter have rest mass as well β not all particles have rest mass, such as the photon. matter can be a pure chemical substance or a mixture of substances. = = = = atom = = = = the atom is the basic unit of chemistry. it consists of a dense core called the atomic nucleus surrounded by a space occupied by an electron cloud. the nucleus is made up of positively charged protons and uncharged neutrons ( together called nucleons ), while the electron cloud consists of negatively charged electrons which orbit the nucleus. in a neutral atom, the negatively charged electrons balance out the positive charge of the protons. the nucleus is dense ; the mass of a nucleon is approximately 1, 836 times that of an electron, yet the radius of an atom is about 10, 000 times that of its nucleus. the atom
energy levels for different chemical substances is useful for their identification by the analysis of spectral lines. different kinds of spectra are often used in chemical spectroscopy, e. g. ir, microwave, nmr, esr, etc. spectroscopy is also used to identify the composition of remote objects β like stars and distant galaxies β by analyzing their radiation spectra. the term chemical energy is often used to indicate the potential of a chemical substance to undergo a transformation through a chemical reaction or to transform other chemical substances. = = = reaction = = = when a chemical substance is transformed as a result of its interaction with another substance or with energy, a chemical reaction is said to have occurred. a chemical reaction is therefore a concept related to the " reaction " of a substance when it comes in close contact with another, whether as a mixture or a solution ; exposure to some form of energy, or both. it results in some energy exchange between the constituents of the reaction as well as with the system environment, which may be designed vessels β often laboratory glassware. chemical reactions can result in the formation or dissociation of molecules, that is, molecules breaking apart to form two or more molecules or rearrangement of atoms within or across molecules. chemical reactions usually involve the making or breaking of chemical bonds. oxidation, reduction, dissociation, acid β base neutralization and molecular rearrangement are some examples of common chemical reactions. a chemical reaction can be symbolically depicted through a chemical equation. while in a non - nuclear chemical reaction the number and kind of atoms on both sides of the equation are equal, for a nuclear reaction this holds true only for the nuclear particles viz. protons and neutrons. the sequence of steps in which the reorganization of chemical bonds may be taking place in the course of a chemical reaction is called its mechanism. a chemical reaction can be envisioned to take place in a number of steps, each of which may have a different speed. many reaction intermediates with variable stability can thus be envisaged during the course of a reaction. reaction mechanisms are proposed to explain the kinetics and the relative product mix of a reaction. many physical chemists specialize in exploring and proposing the mechanisms of various chemical reactions. several empirical rules, like the woodward β hoffmann rules often come in handy while proposing a mechanism for a chemical reaction. according to the iupac gold book, a chemical reaction is " a process that results in the interconversion of chemical species. " accordingly, a chemical reaction may be an elementary reaction or a stepwise reaction.
in one or more of these kinds of structures, it is invariably accompanied by an increase or decrease of energy of the substances involved. some energy is transferred between the surroundings and the reactants of the reaction in the form of heat or light ; thus the products of a reaction may have more or less energy than the reactants. a reaction is said to be exergonic if the final state is lower on the energy scale than the initial state ; in the case of endergonic reactions the situation is the reverse. a reaction is said to be exothermic if the reaction releases heat to the surroundings ; in the case of endothermic reactions, the reaction absorbs heat from the surroundings. chemical reactions are invariably not possible unless the reactants surmount an energy barrier known as the activation energy. the speed of a chemical reaction ( at given temperature t ) is related to the activation energy e, by the boltzmann ' s population factor e β e / k t { \ displaystyle e ^ { - e / kt } } β that is the probability of a molecule to have energy greater than or equal to e at the given temperature t. this exponential dependence of a reaction rate on temperature is known as the arrhenius equation. the activation energy necessary for a chemical reaction to occur can be in the form of heat, light, electricity or mechanical force in the form of ultrasound. a related concept free energy, which also incorporates entropy considerations, is a very useful means for predicting the feasibility of a reaction and determining the state of equilibrium of a chemical reaction, in chemical thermodynamics. a reaction is feasible only if the total change in the gibbs free energy is negative, Ξ΄ g β€ 0 { \ displaystyle \ delta g \ leq 0 \, } ; if it is equal to zero the chemical reaction is said to be at equilibrium. there exist only limited possible states of energy for electrons, atoms and molecules. these are determined by the rules of quantum mechanics, which require quantization of energy of a bound system. the atoms / molecules in a higher energy state are said to be excited. the molecules / atoms of substance in an excited energy state are often much more reactive ; that is, more amenable to chemical reactions. the phase of a substance is invariably determined by its energy and the energy of its surroundings. when the intermolecular forces of a substance are such that the energy of the surroundings is not sufficient to overcome them, it occurs in a more ordered phase like liquid
endothermic reactions, the reaction absorbs heat from the surroundings. chemical reactions are invariably not possible unless the reactants surmount an energy barrier known as the activation energy. the speed of a chemical reaction ( at given temperature t ) is related to the activation energy e, by the boltzmann ' s population factor e β e / k t { \ displaystyle e ^ { - e / kt } } β that is the probability of a molecule to have energy greater than or equal to e at the given temperature t. this exponential dependence of a reaction rate on temperature is known as the arrhenius equation. the activation energy necessary for a chemical reaction to occur can be in the form of heat, light, electricity or mechanical force in the form of ultrasound. a related concept free energy, which also incorporates entropy considerations, is a very useful means for predicting the feasibility of a reaction and determining the state of equilibrium of a chemical reaction, in chemical thermodynamics. a reaction is feasible only if the total change in the gibbs free energy is negative, Ξ΄ g β€ 0 { \ displaystyle \ delta g \ leq 0 \, } ; if it is equal to zero the chemical reaction is said to be at equilibrium. there exist only limited possible states of energy for electrons, atoms and molecules. these are determined by the rules of quantum mechanics, which require quantization of energy of a bound system. the atoms / molecules in a higher energy state are said to be excited. the molecules / atoms of substance in an excited energy state are often much more reactive ; that is, more amenable to chemical reactions. the phase of a substance is invariably determined by its energy and the energy of its surroundings. when the intermolecular forces of a substance are such that the energy of the surroundings is not sufficient to overcome them, it occurs in a more ordered phase like liquid or solid as is the case with water ( h2o ) ; a liquid at room temperature because its molecules are bound by hydrogen bonds. whereas hydrogen sulfide ( h2s ) is a gas at room temperature and standard pressure, as its molecules are bound by weaker dipole β dipole interactions. the transfer of energy from one chemical substance to another depends on the size of energy quanta emitted from one substance. however, heat energy is often transferred more easily from almost any substance to another because the phonons responsible for vibrational and rotational energy levels in a substance have much less energy than photons invoked for the electronic energy transfer
reaction to proceed more rapidly without being consumed by it β by reducing the amount of activation energy needed to convert reactants into products. enzymes also allow the regulation of the rate of a metabolic reaction, for example in response to changes in the cell ' s environment or to signals from other cells. = = = cellular respiration = = = cellular respiration is a set of metabolic reactions and processes that take place in cells to convert chemical energy from nutrients into adenosine triphosphate ( atp ), and then release waste products. the reactions involved in respiration are catabolic reactions, which break large molecules into smaller ones, releasing energy. respiration is one of the key ways a cell releases chemical energy to fuel cellular activity. the overall reaction occurs in a series of biochemical steps, some of which are redox reactions. although cellular respiration is technically a combustion reaction, it clearly does not resemble one when it occurs in a cell because of the slow, controlled release of energy from the series of reactions. sugar in the form of glucose is the main nutrient used by animal and plant cells in respiration. cellular respiration involving oxygen is called aerobic respiration, which has four stages : glycolysis, citric acid cycle ( or krebs cycle ), electron transport chain, and oxidative phosphorylation. glycolysis is a metabolic process that occurs in the cytoplasm whereby glucose is converted into two pyruvates, with two net molecules of atp being produced at the same time. each pyruvate is then oxidized into acetyl - coa by the pyruvate dehydrogenase complex, which also generates nadh and carbon dioxide. acetyl - coa enters the citric acid cycle, which takes places inside the mitochondrial matrix. at the end of the cycle, the total yield from 1 glucose ( or 2 pyruvates ) is 6 nadh, 2 fadh2, and 2 atp molecules. finally, the next stage is oxidative phosphorylation, which in eukaryotes, occurs in the mitochondrial cristae. oxidative phosphorylation comprises the electron transport chain, which is a series of four protein complexes that transfer electrons from one complex to another, thereby releasing energy from nadh and fadh2 that is coupled to the pumping of protons ( hydrogen ions ) across the inner mitochondrial membrane ( chemiosmosis ), which generates a proton motive force. energy
Question: Which of the following represents a chemical reaction?
A) a sugar cube dissolving in water.
B) ice cubes forming in a freezer.
C) ice cream melting in a bowl.
D) a cake baking in an oven.
|
D) a cake baking in an oven.
|
Context:
in which case individual particles retain their shape ) or pulverization ( which involves grinding the particles themselves to a smaller size ). milling is generally done by mechanical means, including attrition ( which is particle - to - particle collision that results in agglomerate break up or particle shearing ), compression ( which applies a forces that results in fracturing ), and impact ( which employs a milling medium or the particles themselves to cause fracturing ). attrition milling equipment includes the wet scrubber ( also called the planetary mill or wet attrition mill ), which has paddles in water creating vortexes in which the material collides and break up. compression mills include the jaw crusher, roller crusher and cone crusher. impact mills include the ball mill, which has media that tumble and fracture the material, or the resonantacoustic mixer. shaft impactors cause particle - to particle attrition and compression. batching is the process of weighing the oxides according to recipes, and preparing them for mixing and drying. mixing occurs after batching and is performed with various machines, such as dry mixing ribbon mixers ( a type of cement mixer ), resonantacoustic mixers, mueller mixers, and pug mills. wet mixing generally involves the same equipment. forming is making the mixed material into shapes, ranging from toilet bowls to spark plug insulators. forming can involve : ( 1 ) extrusion, such as extruding " slugs " to make bricks, ( 2 ) pressing to make shaped parts, ( 3 ) slip casting, as in making toilet bowls, wash basins and ornamentals like ceramic statues. forming produces a " green " part, ready for drying. green parts are soft, pliable, and over time will lose shape. handling the green product will change its shape. for example, a green brick can be " squeezed ", and after squeezing it will stay that way. drying is removing the water or binder from the formed material. spray drying is widely used to prepare powder for pressing operations. other dryers are tunnel dryers and periodic dryers. controlled heat is applied in this two - stage process. first, heat removes water. this step needs careful control, as rapid heating causes cracks and surface defects. the dried part is smaller than the green part, and is brittle, necessitating careful handling, since a small impact will cause crumbling and breaking. sintering is where the dried parts pass through a controlled heating process, and
of tool usage was found in ethiopia within the great rift valley, dating back to 2. 5 million years ago. the earliest methods of stone tool making, known as the oldowan " industry ", date back to at least 2. 3 million years ago. this era of stone tool use is called the paleolithic, or " old stone age ", and spans all of human history up to the development of agriculture approximately 12, 000 years ago. to make a stone tool, a " core " of hard stone with specific flaking properties ( such as flint ) was struck with a hammerstone. this flaking produced sharp edges which could be used as tools, primarily in the form of choppers or scrapers. these tools greatly aided the early humans in their hunter - gatherer lifestyle to perform a variety of tasks including butchering carcasses ( and breaking bones to get at the marrow ) ; chopping wood ; cracking open nuts ; skinning an animal for its hide, and even forming other tools out of softer materials such as bone and wood. the earliest stone tools were irrelevant, being little more than a fractured rock. in the acheulian era, beginning approximately 1. 65 million years ago, methods of working these stones into specific shapes, such as hand axes emerged. this early stone age is described as the lower paleolithic. the middle paleolithic, approximately 300, 000 years ago, saw the introduction of the prepared - core technique, where multiple blades could be rapidly formed from a single core stone. the upper paleolithic, beginning approximately 40, 000 years ago, saw the introduction of pressure flaking, where a wood, bone, or antler punch could be used to shape a stone very finely. the end of the last ice age about 10, 000 years ago is taken as the end point of the upper paleolithic and the beginning of the epipaleolithic / mesolithic. the mesolithic technology included the use of microliths as composite stone tools, along with wood, bone, and antler tools. the later stone age, during which the rudiments of agricultural technology were developed, is called the neolithic period. during this period, polished stone tools were made from a variety of hard rocks such as flint, jade, jadeite, and greenstone, largely by working exposures as quarries, but later the valuable rocks were pursued by tunneling underground, the first steps in mining technology. the polished axes were used for forest clearance and the establishment of crop
##thic, or " old stone age ", and spans all of human history up to the development of agriculture approximately 12, 000 years ago. to make a stone tool, a " core " of hard stone with specific flaking properties ( such as flint ) was struck with a hammerstone. this flaking produced sharp edges which could be used as tools, primarily in the form of choppers or scrapers. these tools greatly aided the early humans in their hunter - gatherer lifestyle to perform a variety of tasks including butchering carcasses ( and breaking bones to get at the marrow ) ; chopping wood ; cracking open nuts ; skinning an animal for its hide, and even forming other tools out of softer materials such as bone and wood. the earliest stone tools were irrelevant, being little more than a fractured rock. in the acheulian era, beginning approximately 1. 65 million years ago, methods of working these stones into specific shapes, such as hand axes emerged. this early stone age is described as the lower paleolithic. the middle paleolithic, approximately 300, 000 years ago, saw the introduction of the prepared - core technique, where multiple blades could be rapidly formed from a single core stone. the upper paleolithic, beginning approximately 40, 000 years ago, saw the introduction of pressure flaking, where a wood, bone, or antler punch could be used to shape a stone very finely. the end of the last ice age about 10, 000 years ago is taken as the end point of the upper paleolithic and the beginning of the epipaleolithic / mesolithic. the mesolithic technology included the use of microliths as composite stone tools, along with wood, bone, and antler tools. the later stone age, during which the rudiments of agricultural technology were developed, is called the neolithic period. during this period, polished stone tools were made from a variety of hard rocks such as flint, jade, jadeite, and greenstone, largely by working exposures as quarries, but later the valuable rocks were pursued by tunneling underground, the first steps in mining technology. the polished axes were used for forest clearance and the establishment of crop farming and were so effective as to remain in use when bronze and iron appeared. these stone axes were used alongside a continued use of stone tools such as a range of projectiles, knives, and scrapers, as well as tools, made from organic materials such as wood, bone, and antler. stone age cultures
which applies a forces that results in fracturing ), and impact ( which employs a milling medium or the particles themselves to cause fracturing ). attrition milling equipment includes the wet scrubber ( also called the planetary mill or wet attrition mill ), which has paddles in water creating vortexes in which the material collides and break up. compression mills include the jaw crusher, roller crusher and cone crusher. impact mills include the ball mill, which has media that tumble and fracture the material, or the resonantacoustic mixer. shaft impactors cause particle - to particle attrition and compression. batching is the process of weighing the oxides according to recipes, and preparing them for mixing and drying. mixing occurs after batching and is performed with various machines, such as dry mixing ribbon mixers ( a type of cement mixer ), resonantacoustic mixers, mueller mixers, and pug mills. wet mixing generally involves the same equipment. forming is making the mixed material into shapes, ranging from toilet bowls to spark plug insulators. forming can involve : ( 1 ) extrusion, such as extruding " slugs " to make bricks, ( 2 ) pressing to make shaped parts, ( 3 ) slip casting, as in making toilet bowls, wash basins and ornamentals like ceramic statues. forming produces a " green " part, ready for drying. green parts are soft, pliable, and over time will lose shape. handling the green product will change its shape. for example, a green brick can be " squeezed ", and after squeezing it will stay that way. drying is removing the water or binder from the formed material. spray drying is widely used to prepare powder for pressing operations. other dryers are tunnel dryers and periodic dryers. controlled heat is applied in this two - stage process. first, heat removes water. this step needs careful control, as rapid heating causes cracks and surface defects. the dried part is smaller than the green part, and is brittle, necessitating careful handling, since a small impact will cause crumbling and breaking. sintering is where the dried parts pass through a controlled heating process, and the oxides are chemically changed to cause bonding and densification. the fired part will be smaller than the dried part. = = forming methods = = ceramic forming techniques include throwing, slipcasting, tape casting, freeze - casting, injection molding, dry pressing, isostatic pressing, hot isostatic pressing
temperature changes up to 1000 Β°c. = = processing steps = = the traditional ceramic process generally follows this sequence : milling β batching β mixing β forming β drying β firing β assembly. milling is the process by which materials are reduced from a large size to a smaller size. milling may involve breaking up cemented material ( in which case individual particles retain their shape ) or pulverization ( which involves grinding the particles themselves to a smaller size ). milling is generally done by mechanical means, including attrition ( which is particle - to - particle collision that results in agglomerate break up or particle shearing ), compression ( which applies a forces that results in fracturing ), and impact ( which employs a milling medium or the particles themselves to cause fracturing ). attrition milling equipment includes the wet scrubber ( also called the planetary mill or wet attrition mill ), which has paddles in water creating vortexes in which the material collides and break up. compression mills include the jaw crusher, roller crusher and cone crusher. impact mills include the ball mill, which has media that tumble and fracture the material, or the resonantacoustic mixer. shaft impactors cause particle - to particle attrition and compression. batching is the process of weighing the oxides according to recipes, and preparing them for mixing and drying. mixing occurs after batching and is performed with various machines, such as dry mixing ribbon mixers ( a type of cement mixer ), resonantacoustic mixers, mueller mixers, and pug mills. wet mixing generally involves the same equipment. forming is making the mixed material into shapes, ranging from toilet bowls to spark plug insulators. forming can involve : ( 1 ) extrusion, such as extruding " slugs " to make bricks, ( 2 ) pressing to make shaped parts, ( 3 ) slip casting, as in making toilet bowls, wash basins and ornamentals like ceramic statues. forming produces a " green " part, ready for drying. green parts are soft, pliable, and over time will lose shape. handling the green product will change its shape. for example, a green brick can be " squeezed ", and after squeezing it will stay that way. drying is removing the water or binder from the formed material. spray drying is widely used to prepare powder for pressing operations. other dryers are tunnel dryers and periodic dryers. controlled heat is applied in this two - stage process. first,
, lightning strikes, tornadoes, building fires, wildfires, and mass shootings disabling most of the system if not the entirety of it. geographic redundancy locations can be more than 621 miles ( 999 km ) continental, more than 62 miles apart and less than 93 miles ( 150 km ) apart, less than 62 miles apart, but not on the same campus, or different buildings that are more than 300 feet ( 91 m ) apart on the same campus. the following methods can reduce the risks of damage by a fire conflagration : large buildings at least 80 feet ( 24 m ) to 110 feet ( 34 m ) apart, but sometimes a minimum of 210 feet ( 64 m ) apart. : 9 high - rise buildings at least 82 feet ( 25 m ) apart : 12 open spaces clear of flammable vegetation within 200 feet ( 61 m ) on each side of objects different wings on the same building, in rooms that are separated by more than 300 feet ( 91 m ) different floors on the same wing of a building in rooms that are horizontally offset by a minimum of 70 feet ( 21 m ) with fire walls between the rooms that are on different floors two rooms separated by another room, leaving at least a 70 - foot gap between the two rooms there should be a minimum of two separated fire walls and on opposite sides of a corridor geographic redundancy is used by amazon web services ( aws ), google cloud platform ( gcp ), microsoft azure, netflix, dropbox, salesforce, linkedin, paypal, twitter, facebook, apple icloud, cisco meraki, and many others to provide geographic redundancy, high availability, fault tolerance and to ensure availability and reliability for their cloud services. as another example, to minimize risk of damage from severe windstorms or water damage, buildings can be located at least 2 miles ( 3. 2 km ) away from the shore, with an elevation of at least 5 feet ( 1. 5 m ) above sea level. for additional protection, they can be located at least 100 feet ( 30 m ) away from flood plain areas. = = functions of redundancy = = the two functions of redundancy are passive redundancy and active redundancy. both functions prevent performance decline from exceeding specification limits without human intervention using extra capacity. passive redundancy uses excess capacity to reduce the impact of component failures. one common form of passive redundancy is the extra strength of cabling and struts used in bridges.
. the first major technologies were tied to survival, hunting, and food preparation. stone tools and weapons, fire, and clothing were technological developments of major importance during this period. human ancestors have been using stone and other tools since long before the emergence of homo sapiens approximately 300, 000 years ago. the earliest direct evidence of tool usage was found in ethiopia within the great rift valley, dating back to 2. 5 million years ago. the earliest methods of stone tool making, known as the oldowan " industry ", date back to at least 2. 3 million years ago. this era of stone tool use is called the paleolithic, or " old stone age ", and spans all of human history up to the development of agriculture approximately 12, 000 years ago. to make a stone tool, a " core " of hard stone with specific flaking properties ( such as flint ) was struck with a hammerstone. this flaking produced sharp edges which could be used as tools, primarily in the form of choppers or scrapers. these tools greatly aided the early humans in their hunter - gatherer lifestyle to perform a variety of tasks including butchering carcasses ( and breaking bones to get at the marrow ) ; chopping wood ; cracking open nuts ; skinning an animal for its hide, and even forming other tools out of softer materials such as bone and wood. the earliest stone tools were irrelevant, being little more than a fractured rock. in the acheulian era, beginning approximately 1. 65 million years ago, methods of working these stones into specific shapes, such as hand axes emerged. this early stone age is described as the lower paleolithic. the middle paleolithic, approximately 300, 000 years ago, saw the introduction of the prepared - core technique, where multiple blades could be rapidly formed from a single core stone. the upper paleolithic, beginning approximately 40, 000 years ago, saw the introduction of pressure flaking, where a wood, bone, or antler punch could be used to shape a stone very finely. the end of the last ice age about 10, 000 years ago is taken as the end point of the upper paleolithic and the beginning of the epipaleolithic / mesolithic. the mesolithic technology included the use of microliths as composite stone tools, along with wood, bone, and antler tools. the later stone age, during which the rudiments of agricultural technology were developed, is called the neolithic period. during this period,
##ning an animal for its hide, and even forming other tools out of softer materials such as bone and wood. the earliest stone tools were irrelevant, being little more than a fractured rock. in the acheulian era, beginning approximately 1. 65 million years ago, methods of working these stones into specific shapes, such as hand axes emerged. this early stone age is described as the lower paleolithic. the middle paleolithic, approximately 300, 000 years ago, saw the introduction of the prepared - core technique, where multiple blades could be rapidly formed from a single core stone. the upper paleolithic, beginning approximately 40, 000 years ago, saw the introduction of pressure flaking, where a wood, bone, or antler punch could be used to shape a stone very finely. the end of the last ice age about 10, 000 years ago is taken as the end point of the upper paleolithic and the beginning of the epipaleolithic / mesolithic. the mesolithic technology included the use of microliths as composite stone tools, along with wood, bone, and antler tools. the later stone age, during which the rudiments of agricultural technology were developed, is called the neolithic period. during this period, polished stone tools were made from a variety of hard rocks such as flint, jade, jadeite, and greenstone, largely by working exposures as quarries, but later the valuable rocks were pursued by tunneling underground, the first steps in mining technology. the polished axes were used for forest clearance and the establishment of crop farming and were so effective as to remain in use when bronze and iron appeared. these stone axes were used alongside a continued use of stone tools such as a range of projectiles, knives, and scrapers, as well as tools, made from organic materials such as wood, bone, and antler. stone age cultures developed music and engaged in organized warfare. stone age humans developed ocean - worthy outrigger canoe technology, leading to migration across the malay archipelago, across the indian ocean to madagascar and also across the pacific ocean, which required knowledge of the ocean currents, weather patterns, sailing, and celestial navigation. although paleolithic cultures left no written records, the shift from nomadic life to settlement and agriculture can be inferred from a range of archaeological evidence. such evidence includes ancient tools, cave paintings, and other prehistoric art, such as the venus of willendorf. human remains also provide direct evidence, both through the examination of bones, and
to that of a flat crack through the plain matrix. the magnitude of the toughening is determined by the mismatch strain caused by thermal contraction incompatibility and the microfracture resistance of the particle / matrix interface. the toughening becomes noticeable with a narrow size distribution of appropriately sized particles, and researchers typically accept that deflection effects in materials with roughly equiaxial grains may increase the fracture toughness by about twice the grain boundary value. the model reveals that the increase in toughness is dependent on particle shape and the volume fraction of the second phase, with the most effective morphology being the rod of high aspect ratio, which can account for a fourfold increase in fracture toughness. the toughening arises primarily from the twist of the crack front between particles, as indicated by deflection profiles. disc - shaped particles and spheres are less effective in toughening. fracture toughness, regardless of morphology, is determined by the twist of the crack front at its most severe configuration, rather than the initial tilt of the crack front. only for disc - shaped particles does the initial tilting of the crack front provide significant toughening ; however, the twist component still overrides the tilt - derived toughening. additional important features of the deflection analysis include the appearance of asymptotic toughening for the three morphologies at volume fractions in excess of 0. 2. it is also noted that a significant influence on the toughening by spherical particles is exerted by the interparticle spacing distribution ; greater toughening is afforded when spheres are nearly contacting such that twist angles approach Ο / 2. these predictions provide the basis for the design of high - toughness two - phase ceramic materials. the ideal second phase, in addition to maintaining chemical compatibility, should be present in amounts of 10 to 20 volume percent. greater amounts may diminish the toughness increase due to overlapping particles. particles with high aspect ratios, especially those with rod - shaped morphologies, are most suitable for maximum toughening. this model is often used to determine the factors that contribute to the increase in fracture toughness in ceramics which is ultimately useful in the development of advanced ceramic materials with improved performance. = = theory of chemical processing = = = = = microstructural uniformity = = = in the processing of fine ceramics, the irregular particle sizes and shapes in a typical powder often lead to non - uniform packing morphologies that result in packing density variations in the powder compact. uncontrolled aggl
which could be used as tools, primarily in the form of choppers or scrapers. these tools greatly aided the early humans in their hunter - gatherer lifestyle to perform a variety of tasks including butchering carcasses ( and breaking bones to get at the marrow ) ; chopping wood ; cracking open nuts ; skinning an animal for its hide, and even forming other tools out of softer materials such as bone and wood. the earliest stone tools were irrelevant, being little more than a fractured rock. in the acheulian era, beginning approximately 1. 65 million years ago, methods of working these stones into specific shapes, such as hand axes emerged. this early stone age is described as the lower paleolithic. the middle paleolithic, approximately 300, 000 years ago, saw the introduction of the prepared - core technique, where multiple blades could be rapidly formed from a single core stone. the upper paleolithic, beginning approximately 40, 000 years ago, saw the introduction of pressure flaking, where a wood, bone, or antler punch could be used to shape a stone very finely. the end of the last ice age about 10, 000 years ago is taken as the end point of the upper paleolithic and the beginning of the epipaleolithic / mesolithic. the mesolithic technology included the use of microliths as composite stone tools, along with wood, bone, and antler tools. the later stone age, during which the rudiments of agricultural technology were developed, is called the neolithic period. during this period, polished stone tools were made from a variety of hard rocks such as flint, jade, jadeite, and greenstone, largely by working exposures as quarries, but later the valuable rocks were pursued by tunneling underground, the first steps in mining technology. the polished axes were used for forest clearance and the establishment of crop farming and were so effective as to remain in use when bronze and iron appeared. these stone axes were used alongside a continued use of stone tools such as a range of projectiles, knives, and scrapers, as well as tools, made from organic materials such as wood, bone, and antler. stone age cultures developed music and engaged in organized warfare. stone age humans developed ocean - worthy outrigger canoe technology, leading to migration across the malay archipelago, across the indian ocean to madagascar and also across the pacific ocean, which required knowledge of the ocean currents, weather patterns, sailing, and celestial navigation. although paleolithic cultures
Question: When a hammer strikes a nail, the nail
A) exerts balanced forces that helps it from moving.
B) exerts a force equal to the weight of the hammer.
C) exerts an unbalanced force that changes its velocity.
D) exerts an equal and opposite force back on the hammer.
|
D) exerts an equal and opposite force back on the hammer.
|
Context:
it was the best of times ; it was the worst of times is the way dickens begins the tale of two cities. the line is appropriate to our time in particle physics. it is the best of times because we are in the midst of a revolution in understanding, the third to occur during my career. it is the worst of times because accelerator facilities are shutting down before new ones are opening, restricting the opportunity for experiments, and because of great uncertainty about future funding. my task today is to give you a view of the most important opportunities for our field under a scenario that is constrained by a tight budget. it is a time when we cannot afford the merely good, but must give first priority to the really important.
the extremely small probability of tunneling through an almost classical potential barrier may become not small under the action of the specially adapted non - stationary signal which selects the certain particle energy e _ r. for particle energies close to this value, the tunneling rate is not small during a finite interval of time and has a very sharp peak at the energy e _ r. after entering inside the barrier, the particle emits electromagnetic quanta and exits the barrier with a lower energy. the signal amplitude can be much less compared to the field of the static barrier. this phenomenon can be called the euclidean resonance since the under - barrier motion occurs in imaginary time. the resonance may stimulate chemical and biochemical reactions in a selective way by adapting the signal to a certain particular chemical bond. the resonance may be used in search of the soft alpha - decay for which a conventional observation is impossible due to an extremely small decay rate.
have evolved from the earliest emergence of life to present day. earth formed about 4. 5 billion years ago and all life on earth, both living and extinct, descended from a last universal common ancestor that lived about 3. 5 billion years ago. geologists have developed a geologic time scale that divides the history of the earth into major divisions, starting with four eons ( hadean, archean, proterozoic, and phanerozoic ), the first three of which are collectively known as the precambrian, which lasted approximately 4 billion years. each eon can be divided into eras, with the phanerozoic eon that began 539 million years ago being subdivided into paleozoic, mesozoic, and cenozoic eras. these three eras together comprise eleven periods ( cambrian, ordovician, silurian, devonian, carboniferous, permian, triassic, jurassic, cretaceous, tertiary, and quaternary ). the similarities among all known present - day species indicate that they have diverged through the process of evolution from their common ancestor. biologists regard the ubiquity of the genetic code as evidence of universal common descent for all bacteria, archaea, and eukaryotes. microbial mats of coexisting bacteria and archaea were the dominant form of life in the early archean eon and many of the major steps in early evolution are thought to have taken place in this environment. the earliest evidence of eukaryotes dates from 1. 85 billion years ago, and while they may have been present earlier, their diversification accelerated when they started using oxygen in their metabolism. later, around 1. 7 billion years ago, multicellular organisms began to appear, with differentiated cells performing specialised functions. algae - like multicellular land plants are dated back to about 1 billion years ago, although evidence suggests that microorganisms formed the earliest terrestrial ecosystems, at least 2. 7 billion years ago. microorganisms are thought to have paved the way for the inception of land plants in the ordovician period. land plants were so successful that they are thought to have contributed to the late devonian extinction event. ediacara biota appear during the ediacaran period, while vertebrates, along with most other modern phyla originated about 525 million years ago during the cambrian explosion. during the permian period, synapsids, including the ancestors of mammals, dominated the land, but most of this group became
in 2023, 639, 300 people died in france, 35, 900 fewer than in 2022, a year of high mortality. over the last twenty years, from 2004 to 2023, january 3rd was the deadliest day, while august 15th was the least deadly one. elderly people die significantly less often in the summer. deaths are also less frequent on public holidays and sundays. finally, the risk of dying is higher on one ' s birthday, especially for young people.
displaystyle \ mathbb { r } } that are both open and closed. a degenerate interval is any set consisting of a single real number ( i. e., an interval of the form [ a, a ] ). some authors include the empty set in this definition. a real interval that is neither empty nor degenerate is said to be proper, and has infinitely many elements. an interval is said to be left - bounded or right - bounded, if there is some real number that is, respectively, smaller than or larger than all its elements. an interval is said to be bounded, if it is both left - and right - bounded ; and is said to be unbounded otherwise. intervals that are bounded at only one end are said to be half - bounded. the empty set is bounded, and the set of all reals is the only interval that is unbounded at both ends. bounded intervals are also commonly known as finite intervals. bounded intervals are bounded sets, in the sense that their diameter ( which is equal to the absolute difference between the endpoints ) is finite. the diameter may be called the length, width, measure, range, or size of the interval. the size of unbounded intervals is usually defined as + β, and the size of the empty interval may be defined as 0 ( or left undefined ). the centre ( midpoint ) of a bounded interval with endpoints a and b is ( a + b ) / 2, and its radius is the half - length | a β b | / 2. these concepts are undefined for empty or unbounded intervals. an interval is said to be left - open if and only if it contains no minimum ( an element that is smaller than all other elements ) ; right - open if it contains no maximum ; and open if it contains neither. the interval [ 0, 1 ) = { x | 0 β€ x < 1 }, for example, is left - closed and right - open. the empty set and the set of all reals are both open and closed intervals, while the set of non - negative reals, is a closed interval that is right - open but not left - open. the open intervals are open sets of the real line in its standard topology, and form a base of the open sets. an interval is said to be left - closed if it has a minimum element or is left - unbounded, right - closed if it has a maximum or is right unbounded ; it is
- and right - bounded ; and is said to be unbounded otherwise. intervals that are bounded at only one end are said to be half - bounded. the empty set is bounded, and the set of all reals is the only interval that is unbounded at both ends. bounded intervals are also commonly known as finite intervals. bounded intervals are bounded sets, in the sense that their diameter ( which is equal to the absolute difference between the endpoints ) is finite. the diameter may be called the length, width, measure, range, or size of the interval. the size of unbounded intervals is usually defined as + β, and the size of the empty interval may be defined as 0 ( or left undefined ). the centre ( midpoint ) of a bounded interval with endpoints a and b is ( a + b ) / 2, and its radius is the half - length | a β b | / 2. these concepts are undefined for empty or unbounded intervals. an interval is said to be left - open if and only if it contains no minimum ( an element that is smaller than all other elements ) ; right - open if it contains no maximum ; and open if it contains neither. the interval [ 0, 1 ) = { x | 0 β€ x < 1 }, for example, is left - closed and right - open. the empty set and the set of all reals are both open and closed intervals, while the set of non - negative reals, is a closed interval that is right - open but not left - open. the open intervals are open sets of the real line in its standard topology, and form a base of the open sets. an interval is said to be left - closed if it has a minimum element or is left - unbounded, right - closed if it has a maximum or is right unbounded ; it is simply closed if it is both left - closed and right closed. so, the closed intervals coincide with the closed sets in that topology. the interior of an interval i is the largest open interval that is contained in i ; it is also the set of points in i which are not endpoints of i. the closure of i is the smallest closed interval that contains i ; which is also the set i augmented with its finite endpoints. for any set x of real numbers, the interval enclosure or interval span of x is the unique interval that contains x, and does not properly contain any other interval that also contains x. an interval i is
ultra high energy particles arrive at earth constantly. they provide a beam at energies higher than any man - made accelerator, but at a very low rate. two large experiments, the pierre auger observatory and the telescope array experiment, have been taking data for several years now covering together the whole sky. i summarize the most recent measurements from both experiments, i compare their results and, for a change, i highlight their agreements.
empty nor degenerate is said to be proper, and has infinitely many elements. an interval is said to be left - bounded or right - bounded, if there is some real number that is, respectively, smaller than or larger than all its elements. an interval is said to be bounded, if it is both left - and right - bounded ; and is said to be unbounded otherwise. intervals that are bounded at only one end are said to be half - bounded. the empty set is bounded, and the set of all reals is the only interval that is unbounded at both ends. bounded intervals are also commonly known as finite intervals. bounded intervals are bounded sets, in the sense that their diameter ( which is equal to the absolute difference between the endpoints ) is finite. the diameter may be called the length, width, measure, range, or size of the interval. the size of unbounded intervals is usually defined as + β, and the size of the empty interval may be defined as 0 ( or left undefined ). the centre ( midpoint ) of a bounded interval with endpoints a and b is ( a + b ) / 2, and its radius is the half - length | a β b | / 2. these concepts are undefined for empty or unbounded intervals. an interval is said to be left - open if and only if it contains no minimum ( an element that is smaller than all other elements ) ; right - open if it contains no maximum ; and open if it contains neither. the interval [ 0, 1 ) = { x | 0 β€ x < 1 }, for example, is left - closed and right - open. the empty set and the set of all reals are both open and closed intervals, while the set of non - negative reals, is a closed interval that is right - open but not left - open. the open intervals are open sets of the real line in its standard topology, and form a base of the open sets. an interval is said to be left - closed if it has a minimum element or is left - unbounded, right - closed if it has a maximum or is right unbounded ; it is simply closed if it is both left - closed and right closed. so, the closed intervals coincide with the closed sets in that topology. the interior of an interval i is the largest open interval that is contained in i ; it is also the set of points in i which are not endpoints of i. the closure of
becomes quite gentle. accordingly, in large basins, rivers in most cases begin as torrents with a variable flow, and end as gently flowing rivers with a comparatively regular discharge. the irregular flow of rivers throughout their course forms one of the main difficulties in devising works for mitigating inundations or for increasing the navigable capabilities of rivers. in tropical countries subject to periodical rains, the rivers are in flood during the rainy season and have hardly any flow during the rest of the year, while in temperate regions, where the rainfall is more evenly distributed throughout the year, evaporation causes the available rainfall to be much less in hot summer weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under
time interval between the incident and scattered photon in raman effect and absorption of photon and emission of electron in photoelectric effect has not been determined till now. this is because there is no such high level instrument discovered till now to detect time interval to such a small level. but this can be calculated theoretically by applying a basic principle of physics like impulse is equal to the change in momentum. considering the collision between electron and photon as perfect inelastic collision in photoelectric effect, elastic and inelastic collision in raman effect and elastic collision in plane mirror reflection and the interaction between electron and photon as strong gravitational interaction we calculate the required time interval. during these phenomena there is lattice vibration which can be quantized as phonon particles.
Question: Which of these natural events occurs at time intervals that are the most regular?
A) drought and flooding
B) high tides and low tides
C) clear skies and cloudy skies
D) hot weather and cool weather
|
B) high tides and low tides
|
Context:
temperature changes up to 1000 Β°c. = = processing steps = = the traditional ceramic process generally follows this sequence : milling β batching β mixing β forming β drying β firing β assembly. milling is the process by which materials are reduced from a large size to a smaller size. milling may involve breaking up cemented material ( in which case individual particles retain their shape ) or pulverization ( which involves grinding the particles themselves to a smaller size ). milling is generally done by mechanical means, including attrition ( which is particle - to - particle collision that results in agglomerate break up or particle shearing ), compression ( which applies a forces that results in fracturing ), and impact ( which employs a milling medium or the particles themselves to cause fracturing ). attrition milling equipment includes the wet scrubber ( also called the planetary mill or wet attrition mill ), which has paddles in water creating vortexes in which the material collides and break up. compression mills include the jaw crusher, roller crusher and cone crusher. impact mills include the ball mill, which has media that tumble and fracture the material, or the resonantacoustic mixer. shaft impactors cause particle - to particle attrition and compression. batching is the process of weighing the oxides according to recipes, and preparing them for mixing and drying. mixing occurs after batching and is performed with various machines, such as dry mixing ribbon mixers ( a type of cement mixer ), resonantacoustic mixers, mueller mixers, and pug mills. wet mixing generally involves the same equipment. forming is making the mixed material into shapes, ranging from toilet bowls to spark plug insulators. forming can involve : ( 1 ) extrusion, such as extruding " slugs " to make bricks, ( 2 ) pressing to make shaped parts, ( 3 ) slip casting, as in making toilet bowls, wash basins and ornamentals like ceramic statues. forming produces a " green " part, ready for drying. green parts are soft, pliable, and over time will lose shape. handling the green product will change its shape. for example, a green brick can be " squeezed ", and after squeezing it will stay that way. drying is removing the water or binder from the formed material. spray drying is widely used to prepare powder for pressing operations. other dryers are tunnel dryers and periodic dryers. controlled heat is applied in this two - stage process. first,
which applies a forces that results in fracturing ), and impact ( which employs a milling medium or the particles themselves to cause fracturing ). attrition milling equipment includes the wet scrubber ( also called the planetary mill or wet attrition mill ), which has paddles in water creating vortexes in which the material collides and break up. compression mills include the jaw crusher, roller crusher and cone crusher. impact mills include the ball mill, which has media that tumble and fracture the material, or the resonantacoustic mixer. shaft impactors cause particle - to particle attrition and compression. batching is the process of weighing the oxides according to recipes, and preparing them for mixing and drying. mixing occurs after batching and is performed with various machines, such as dry mixing ribbon mixers ( a type of cement mixer ), resonantacoustic mixers, mueller mixers, and pug mills. wet mixing generally involves the same equipment. forming is making the mixed material into shapes, ranging from toilet bowls to spark plug insulators. forming can involve : ( 1 ) extrusion, such as extruding " slugs " to make bricks, ( 2 ) pressing to make shaped parts, ( 3 ) slip casting, as in making toilet bowls, wash basins and ornamentals like ceramic statues. forming produces a " green " part, ready for drying. green parts are soft, pliable, and over time will lose shape. handling the green product will change its shape. for example, a green brick can be " squeezed ", and after squeezing it will stay that way. drying is removing the water or binder from the formed material. spray drying is widely used to prepare powder for pressing operations. other dryers are tunnel dryers and periodic dryers. controlled heat is applied in this two - stage process. first, heat removes water. this step needs careful control, as rapid heating causes cracks and surface defects. the dried part is smaller than the green part, and is brittle, necessitating careful handling, since a small impact will cause crumbling and breaking. sintering is where the dried parts pass through a controlled heating process, and the oxides are chemically changed to cause bonding and densification. the fired part will be smaller than the dried part. = = forming methods = = ceramic forming techniques include throwing, slipcasting, tape casting, freeze - casting, injection molding, dry pressing, isostatic pressing, hot isostatic pressing
building block. ceramics β not to be confused with raw, unfired clay β are usually seen in crystalline form. the vast majority of commercial glasses contain a metal oxide fused with silica. at the high temperatures used to prepare glass, the material is a viscous liquid which solidifies into a disordered state upon cooling. windowpanes and eyeglasses are important examples. fibers of glass are also used for long - range telecommunication and optical transmission. scratch resistant corning gorilla glass is a well - known example of the application of materials science to drastically improve the properties of common components. engineering ceramics are known for their stiffness and stability under high temperatures, compression and electrical stress. alumina, silicon carbide, and tungsten carbide are made from a fine powder of their constituents in a process of sintering with a binder. hot pressing provides higher density material. chemical vapor deposition can place a film of a ceramic on another material. cermets are ceramic particles containing some metals. the wear resistance of tools is derived from cemented carbides with the metal phase of cobalt and nickel typically added to modify properties. ceramics can be significantly strengthened for engineering applications using the principle of crack deflection. this process involves the strategic addition of second - phase particles within a ceramic matrix, optimizing their shape, size, and distribution to direct and control crack propagation. this approach enhances fracture toughness, paving the way for the creation of advanced, high - performance ceramics in various industries. = = = composites = = = another application of materials science in industry is making composite materials. these are structured materials composed of two or more macroscopic phases. applications range from structural elements such as steel - reinforced concrete, to the thermal insulating tiles, which play a key and integral role in nasa ' s space shuttle thermal protection system, which is used to protect the surface of the shuttle from the heat of re - entry into the earth ' s atmosphere. one example is reinforced carbon - carbon ( rcc ), the light gray material, which withstands re - entry temperatures up to 1, 510 Β°c ( 2, 750 Β°f ) and protects the space shuttle ' s wing leading edges and nose cap. rcc is a laminated composite material made from graphite rayon cloth and impregnated with a phenolic resin. after curing at high temperature in an autoclave, the laminate is pyrolized to convert the resin to carbon, impregnated with furfuryl alcohol in a
the broad definition of " utilizing a biotechnological system to make products ". indeed, the cultivation of plants may be viewed as the earliest biotechnological enterprise. agriculture has been theorized to have become the dominant way of producing food since the neolithic revolution. through early biotechnology, the earliest farmers selected and bred the best - suited crops ( e. g., those with the highest yields ) to produce enough food to support a growing population. as crops and fields became increasingly large and difficult to maintain, it was discovered that specific organisms and their by - products could effectively fertilize, restore nitrogen, and control pests. throughout the history of agriculture, farmers have inadvertently altered the genetics of their crops through introducing them to new environments and breeding them with other plants β one of the first forms of biotechnology. these processes also were included in early fermentation of beer. these processes were introduced in early mesopotamia, egypt, china and india, and still use the same basic biological methods. in brewing, malted grains ( containing enzymes ) convert starch from grains into sugar and then adding specific yeasts to produce beer. in this process, carbohydrates in the grains broke down into alcohols, such as ethanol. later, other cultures produced the process of lactic acid fermentation, which produced other preserved foods, such as soy sauce. fermentation was also used in this time period to produce leavened bread. although the process of fermentation was not fully understood until louis pasteur ' s work in 1857, it is still the first use of biotechnology to convert a food source into another form. before the time of charles darwin ' s work and life, animal and plant scientists had already used selective breeding. darwin added to that body of work with his scientific observations about the ability of science to change species. these accounts contributed to darwin ' s theory of natural selection. for thousands of years, humans have used selective breeding to improve the production of crops and livestock to use them for food. in selective breeding, organisms with desirable characteristics are mated to produce offspring with the same characteristics. for example, this technique was used with corn to produce the largest and sweetest crops. in the early twentieth century scientists gained a greater understanding of microbiology and explored ways of manufacturing specific products. in 1917, chaim weizmann first used a pure microbiological culture in an industrial process, that of manufacturing corn starch using clostridium acetobutylicum, to produce acetone, which the united
the material collides and break up. compression mills include the jaw crusher, roller crusher and cone crusher. impact mills include the ball mill, which has media that tumble and fracture the material, or the resonantacoustic mixer. shaft impactors cause particle - to particle attrition and compression. batching is the process of weighing the oxides according to recipes, and preparing them for mixing and drying. mixing occurs after batching and is performed with various machines, such as dry mixing ribbon mixers ( a type of cement mixer ), resonantacoustic mixers, mueller mixers, and pug mills. wet mixing generally involves the same equipment. forming is making the mixed material into shapes, ranging from toilet bowls to spark plug insulators. forming can involve : ( 1 ) extrusion, such as extruding " slugs " to make bricks, ( 2 ) pressing to make shaped parts, ( 3 ) slip casting, as in making toilet bowls, wash basins and ornamentals like ceramic statues. forming produces a " green " part, ready for drying. green parts are soft, pliable, and over time will lose shape. handling the green product will change its shape. for example, a green brick can be " squeezed ", and after squeezing it will stay that way. drying is removing the water or binder from the formed material. spray drying is widely used to prepare powder for pressing operations. other dryers are tunnel dryers and periodic dryers. controlled heat is applied in this two - stage process. first, heat removes water. this step needs careful control, as rapid heating causes cracks and surface defects. the dried part is smaller than the green part, and is brittle, necessitating careful handling, since a small impact will cause crumbling and breaking. sintering is where the dried parts pass through a controlled heating process, and the oxides are chemically changed to cause bonding and densification. the fired part will be smaller than the dried part. = = forming methods = = ceramic forming techniques include throwing, slipcasting, tape casting, freeze - casting, injection molding, dry pressing, isostatic pressing, hot isostatic pressing ( hip ), 3d printing and others. methods for forming ceramic powders into complex shapes are desirable in many areas of technology. such methods are required for producing advanced, high - temperature structural parts such as heat engine components and turbines. materials other than ceramics which are used in these processes may include : wood, metal,
10 kgy most food, which is ( with regard to warming ) physically equivalent to water, would warm by only about 2. 5 Β°c ( 4. 5 Β°f ). the specialty of processing food by ionizing radiation is the fact, that the energy density per atomic transition is very high, it can cleave molecules and induce ionization ( hence the name ) which cannot be achieved by mere heating. this is the reason for new beneficial effects, however at the same time, for new concerns. the treatment of solid food by ionizing radiation can provide an effect similar to heat pasteurization of liquids, such as milk. however, the use of the term, cold pasteurization, to describe irradiated foods is controversial, because pasteurization and irradiation are fundamentally different processes, although the intended end results can in some cases be similar. detractors of food irradiation have concerns about the health hazards of induced radioactivity. a report for the industry advocacy group american council on science and health entitled " irradiated foods " states : " the types of radiation sources approved for the treatment of foods have specific energy levels well below that which would cause any element in food to become radioactive. food undergoing irradiation does not become any more radioactive than luggage passing through an airport x - ray scanner or teeth that have been x - rayed. " food irradiation is currently permitted by over 40 countries and volumes are estimated to exceed 500, 000 metric tons ( 490, 000 long tons ; 550, 000 short tons ) annually worldwide. food irradiation is essentially a non - nuclear technology ; it relies on the use of ionizing radiation which may be generated by accelerators for electrons and conversion into bremsstrahlung, but which may use also gamma - rays from nuclear decay. there is a worldwide industry for processing by ionizing radiation, the majority by number and by processing power using accelerators. food irradiation is only a niche application compared to medical supplies, plastic materials, raw materials, gemstones, cables and wires, etc. = = accidents = = nuclear accidents, because of the powerful forces involved, are often very dangerous. historically, the first incidents involved fatal radiation exposure. marie curie died from aplastic anemia which resulted from her high levels of exposure. two scientists, an american and canadian respectively, harry daghlian and louis slotin, died after mishandling the same plutonium mass. unlike conventional weapons, the intense light, heat, and explosive force is
which constitutes anywhere from 30 % [ m / m ] to 90 % [ m / m ] of its composition by volume, yielding an array of materials with interesting thermomechanical properties. in the processing of glass - ceramics, molten glass is cooled down gradually before reheating and annealing. in this heat treatment the glass partly crystallizes. in many cases, so - called ' nucleation agents ' are added in order to regulate and control the crystallization process. because there is usually no pressing and sintering, glass - ceramics do not contain the volume fraction of porosity typically present in sintered ceramics. the term mainly refers to a mix of lithium and aluminosilicates which yields an array of materials with interesting thermomechanical properties. the most commercially important of these have the distinction of being impervious to thermal shock. thus, glass - ceramics have become extremely useful for countertop cooking. the negative thermal expansion coefficient ( tec ) of the crystalline ceramic phase can be balanced with the positive tec of the glassy phase. at a certain point ( ~ 70 % crystalline ) the glass - ceramic has a net tec near zero. this type of glass - ceramic exhibits excellent mechanical properties and can sustain repeated and quick temperature changes up to 1000 Β°c. = = processing steps = = the traditional ceramic process generally follows this sequence : milling β batching β mixing β forming β drying β firing β assembly. milling is the process by which materials are reduced from a large size to a smaller size. milling may involve breaking up cemented material ( in which case individual particles retain their shape ) or pulverization ( which involves grinding the particles themselves to a smaller size ). milling is generally done by mechanical means, including attrition ( which is particle - to - particle collision that results in agglomerate break up or particle shearing ), compression ( which applies a forces that results in fracturing ), and impact ( which employs a milling medium or the particles themselves to cause fracturing ). attrition milling equipment includes the wet scrubber ( also called the planetary mill or wet attrition mill ), which has paddles in water creating vortexes in which the material collides and break up. compression mills include the jaw crusher, roller crusher and cone crusher. impact mills include the ball mill, which has media that tumble and fracture the material, or the resonantacoustic mixer. shaft impactors cause particle - to particle attrition and compression
thermal expansion coefficient ( tec ) of the crystalline ceramic phase can be balanced with the positive tec of the glassy phase. at a certain point ( ~ 70 % crystalline ) the glass - ceramic has a net tec near zero. this type of glass - ceramic exhibits excellent mechanical properties and can sustain repeated and quick temperature changes up to 1000 Β°c. = = processing steps = = the traditional ceramic process generally follows this sequence : milling β batching β mixing β forming β drying β firing β assembly. milling is the process by which materials are reduced from a large size to a smaller size. milling may involve breaking up cemented material ( in which case individual particles retain their shape ) or pulverization ( which involves grinding the particles themselves to a smaller size ). milling is generally done by mechanical means, including attrition ( which is particle - to - particle collision that results in agglomerate break up or particle shearing ), compression ( which applies a forces that results in fracturing ), and impact ( which employs a milling medium or the particles themselves to cause fracturing ). attrition milling equipment includes the wet scrubber ( also called the planetary mill or wet attrition mill ), which has paddles in water creating vortexes in which the material collides and break up. compression mills include the jaw crusher, roller crusher and cone crusher. impact mills include the ball mill, which has media that tumble and fracture the material, or the resonantacoustic mixer. shaft impactors cause particle - to particle attrition and compression. batching is the process of weighing the oxides according to recipes, and preparing them for mixing and drying. mixing occurs after batching and is performed with various machines, such as dry mixing ribbon mixers ( a type of cement mixer ), resonantacoustic mixers, mueller mixers, and pug mills. wet mixing generally involves the same equipment. forming is making the mixed material into shapes, ranging from toilet bowls to spark plug insulators. forming can involve : ( 1 ) extrusion, such as extruding " slugs " to make bricks, ( 2 ) pressing to make shaped parts, ( 3 ) slip casting, as in making toilet bowls, wash basins and ornamentals like ceramic statues. forming produces a " green " part, ready for drying. green parts are soft, pliable, and over time will lose shape. handling the green product will change its shape. for example, a green brick can
wrought, which itself is the original past passive participle of the word work, now superseded by the weak verb forms worker and worked respectively. ) blacksmithing and the various related smithing and metal - crafts. folk music played on acoustic instruments. mathematics ( particularly, pure mathematics ) organic farming and animal husbandry ( i. e. ; agriculture as practiced by all american farmers prior to world war ii ). milling in the sense of operating hand - constructed equipment with the intent to either grind grain, or the reduction of timber to lumber as practiced in a saw - mill. fulling, felting, drop spindle spinning, hand knitting, crochet, & similar textile preparation. the production of charcoal by the collier, for use in home heating, foundry operations, smelting, the various smithing trades, and for brushing ones teeth as in colonial america. glass - blowing. various subskills of food preservation : smoking salting pickling drying note : home canning is a counter example of a low technology since some of the supplies needed to pursue this skill rely on a global trade network and an existing manufacturing infrastructure. the production of various alcoholic beverages : wine : poorly preserved fruit juice. beer : a way to preserve the calories of grain products from decay. whiskey : an improved ( distilled ) form of beer. flint - knapping masonry as used in castles, cathedrals, and root cellars. = = = domestic or consumer = = = ( non exhaustive ) list of low - tech in a westerner ' s everyday life : getting around by bike, and repairing it with second - hand materials using a cargo bike to carry loads ( rather than a gasoline vehicle ) drying clothes on a clothesline or on a drying rack washing clothes by hand, or in a human - powered washing machine cooling one ' s home with a fan or an air expander ( rather than electrical appliances such as air conditioners ) using a bell as door bell a cellar, " desert fridge ", or icebox ( rather than a fridge or freezer ) long - distance travel by sailing boat ( rather than by plane ) a wicker bag or a tote bag ( rather than a plastic bag ) to carry things swedish lighter ( rather than disposable lighter or matches ) a hand drill, instead of an electric one lighting with sunlight or candles hemp textiles to water plants with drip irrigation paper sheets for note - taking to clean with a broom ( rather than a vacuum cleaner ) to find one ' s way with map
; however, a successful large - scale industrial application of the process was the development of continuous freeze drying of coffee. high - temperature short time processing β these processes, for the most part, are characterized by rapid heating and cooling, holding for a short time at a relatively high temperature and filling aseptically into sterile containers. decaffeination of coffee and tea β decaffeinated coffee and tea was first developed on a commercial basis in europe around 1900. the process is described in u. s. patent 897, 763. green coffee beans are treated with water, heat and solvents to remove the caffeine from the beans. process optimization β food technology now allows production of foods to be more efficient, oil saving technologies are now available on different forms. production methods and methodology have also become increasingly sophisticated. aseptic packaging β the process of filling a commercially sterile product into a sterile container and hermetically sealing the containers so that re - infection is prevented. thus, this results into a shelf stable product at ambient conditions. food irradiation β the process of exposing food and food packaging to ionizing radiation can effectively destroy organisms responsible for spoilage and foodborne illness and inhibit sprouting, extending shelf life. commercial fruit ripening rooms using ethylene as a plant hormone. food delivery β an order is typically made either through a restaurant or grocer ' s website or mobile app, or through a food ordering company. the ordered food is typically delivered in boxes or bags to the customer ' s doorsteps. = = categories = = technology has innovated these categories from the food industry : agricultural technology β or agtech, it is the use of technology in agriculture, horticulture, and aquaculture with the aim of improving yield, efficiency, and profitability. agricultural technology can be products, services or applications derived from agriculture that improve various input / output processes. food science β technology in this sector focuses on the development of new functional ingredients and alternative proteins. foodservice β technology innovated the way establishments prepare, supply, and serve food outside the home. there ' s a tendency to create the conditions for the restaurant of the future with robotics and cloudkitchens. consumer tech β technology allows what we call consumer electronics, which is the equipment of consumers with devices that facilitates the cooking process. food delivery β as the food delivery market is growing, companies and startups are rapidly revolutionizing the communication process between consumers and food establishments, with platform - to - consumer delivery as the
Question: A student combined rice cereal with melted marshmallows. Which best describes what the student made?
A) an atom
B) a mixture
C) a molecule
D) an element
|
B) a mixture
|
Context:
; austrian experts have established that the wheel is between 5, 100 and 5, 350 years old. the invention of the wheel revolutionized trade and war. it did not take long to discover that wheeled wagons could be used to carry heavy loads. the ancient sumerians used a potter ' s wheel and may have invented it. a stone pottery wheel found in the city - state of ur dates to around 3, 429 bce, and even older fragments of wheel - thrown pottery have been found in the same area. fast ( rotary ) potters ' wheels enabled early mass production of pottery, but it was the use of the wheel as a transformer of energy ( through water wheels, windmills, and even treadmills ) that revolutionized the application of nonhuman power sources. the first two - wheeled carts were derived from travois and were first used in mesopotamia and iran in around 3, 000 bce. the oldest known constructed roadways are the stone - paved streets of the city - state of ur, dating to c. 4, 000 bce, and timber roads leading through the swamps of glastonbury, england, dating to around the same period. the first long - distance road, which came into use around 3, 500 bce, spanned 2, 400 km from the persian gulf to the mediterranean sea, but was not paved and was only partially maintained. in around 2, 000 bce, the minoans on the greek island of crete built a 50 km road leading from the palace of gortyn on the south side of the island, through the mountains, to the palace of knossos on the north side of the island. unlike the earlier road, the minoan road was completely paved. ancient minoan private homes had running water. a bathtub virtually identical to modern ones was unearthed at the palace of knossos. several minoan private homes also had toilets, which could be flushed by pouring water down the drain. the ancient romans had many public flush toilets, which emptied into an extensive sewage system. the primary sewer in rome was the cloaca maxima ; construction began on it in the sixth century bce and it is still in use today. the ancient romans also had a complex system of aqueducts, which were used to transport water across long distances. the first roman aqueduct was built in 312 bce. the eleventh and final ancient roman aqueduct was built in 226 ce. put together, the roman aqueducts extended over 450 km, but less than 70 km of this was above ground
final version. to appear in discrete and continuous dynamical systems - a.
it. a stone pottery wheel found in the city - state of ur dates to around 3, 429 bce, and even older fragments of wheel - thrown pottery have been found in the same area. fast ( rotary ) potters ' wheels enabled early mass production of pottery, but it was the use of the wheel as a transformer of energy ( through water wheels, windmills, and even treadmills ) that revolutionized the application of nonhuman power sources. the first two - wheeled carts were derived from travois and were first used in mesopotamia and iran in around 3, 000 bce. the oldest known constructed roadways are the stone - paved streets of the city - state of ur, dating to c. 4, 000 bce, and timber roads leading through the swamps of glastonbury, england, dating to around the same period. the first long - distance road, which came into use around 3, 500 bce, spanned 2, 400 km from the persian gulf to the mediterranean sea, but was not paved and was only partially maintained. in around 2, 000 bce, the minoans on the greek island of crete built a 50 km road leading from the palace of gortyn on the south side of the island, through the mountains, to the palace of knossos on the north side of the island. unlike the earlier road, the minoan road was completely paved. ancient minoan private homes had running water. a bathtub virtually identical to modern ones was unearthed at the palace of knossos. several minoan private homes also had toilets, which could be flushed by pouring water down the drain. the ancient romans had many public flush toilets, which emptied into an extensive sewage system. the primary sewer in rome was the cloaca maxima ; construction began on it in the sixth century bce and it is still in use today. the ancient romans also had a complex system of aqueducts, which were used to transport water across long distances. the first roman aqueduct was built in 312 bce. the eleventh and final ancient roman aqueduct was built in 226 ce. put together, the roman aqueducts extended over 450 km, but less than 70 km of this was above ground and supported by arches. = = = pre - modern = = = innovations continued through the middle ages with the introduction of silk production ( in asia and later europe ), the horse collar, and horseshoes. simple machines ( such as the lever, the screw, and the pulley ) were combined into more complicated tools
any two generating systems of the fundamental group of a closed surface are nielsen equivalent.
a transformer of energy ( through water wheels, windmills, and even treadmills ) that revolutionized the application of nonhuman power sources. the first two - wheeled carts were derived from travois and were first used in mesopotamia and iran in around 3, 000 bce. the oldest known constructed roadways are the stone - paved streets of the city - state of ur, dating to c. 4, 000 bce, and timber roads leading through the swamps of glastonbury, england, dating to around the same period. the first long - distance road, which came into use around 3, 500 bce, spanned 2, 400 km from the persian gulf to the mediterranean sea, but was not paved and was only partially maintained. in around 2, 000 bce, the minoans on the greek island of crete built a 50 km road leading from the palace of gortyn on the south side of the island, through the mountains, to the palace of knossos on the north side of the island. unlike the earlier road, the minoan road was completely paved. ancient minoan private homes had running water. a bathtub virtually identical to modern ones was unearthed at the palace of knossos. several minoan private homes also had toilets, which could be flushed by pouring water down the drain. the ancient romans had many public flush toilets, which emptied into an extensive sewage system. the primary sewer in rome was the cloaca maxima ; construction began on it in the sixth century bce and it is still in use today. the ancient romans also had a complex system of aqueducts, which were used to transport water across long distances. the first roman aqueduct was built in 312 bce. the eleventh and final ancient roman aqueduct was built in 226 ce. put together, the roman aqueducts extended over 450 km, but less than 70 km of this was above ground and supported by arches. = = = pre - modern = = = innovations continued through the middle ages with the introduction of silk production ( in asia and later europe ), the horse collar, and horseshoes. simple machines ( such as the lever, the screw, and the pulley ) were combined into more complicated tools, such as the wheelbarrow, windmills, and clocks. a system of universities developed and spread scientific ideas and practices, including oxford and cambridge. the renaissance era produced many innovations, including the introduction of the movable type printing press to europe, which facilitated the communication of knowledge. technology became increasingly influenced by science
the paper is withdrawn by the author because it is superseded by cond - mat / 0303357.
wrought, which itself is the original past passive participle of the word work, now superseded by the weak verb forms worker and worked respectively. ) blacksmithing and the various related smithing and metal - crafts. folk music played on acoustic instruments. mathematics ( particularly, pure mathematics ) organic farming and animal husbandry ( i. e. ; agriculture as practiced by all american farmers prior to world war ii ). milling in the sense of operating hand - constructed equipment with the intent to either grind grain, or the reduction of timber to lumber as practiced in a saw - mill. fulling, felting, drop spindle spinning, hand knitting, crochet, & similar textile preparation. the production of charcoal by the collier, for use in home heating, foundry operations, smelting, the various smithing trades, and for brushing ones teeth as in colonial america. glass - blowing. various subskills of food preservation : smoking salting pickling drying note : home canning is a counter example of a low technology since some of the supplies needed to pursue this skill rely on a global trade network and an existing manufacturing infrastructure. the production of various alcoholic beverages : wine : poorly preserved fruit juice. beer : a way to preserve the calories of grain products from decay. whiskey : an improved ( distilled ) form of beer. flint - knapping masonry as used in castles, cathedrals, and root cellars. = = = domestic or consumer = = = ( non exhaustive ) list of low - tech in a westerner ' s everyday life : getting around by bike, and repairing it with second - hand materials using a cargo bike to carry loads ( rather than a gasoline vehicle ) drying clothes on a clothesline or on a drying rack washing clothes by hand, or in a human - powered washing machine cooling one ' s home with a fan or an air expander ( rather than electrical appliances such as air conditioners ) using a bell as door bell a cellar, " desert fridge ", or icebox ( rather than a fridge or freezer ) long - distance travel by sailing boat ( rather than by plane ) a wicker bag or a tote bag ( rather than a plastic bag ) to carry things swedish lighter ( rather than disposable lighter or matches ) a hand drill, instead of an electric one lighting with sunlight or candles hemp textiles to water plants with drip irrigation paper sheets for note - taking to clean with a broom ( rather than a vacuum cleaner ) to find one ' s way with map
comment on " liquids on topologically nanopatterned surfaces " by o. gang et al, phys. rev. lett. 95, 217801 ( 2005 ). see also an erratum published by o. gang et al ( phys rev lett, to appear )
two possible interpretations of frw cosmologies ( perfect fluid or dissipative fluid ) are considered as consecutive phases of the system. necessary conditions are found, for the transition from perfect fluid to dissipative regime to occur, bringing out the conspicuous role played by a particular state of the system ( the ' ' critical point ' ' ).
fluid dynamics video demonstrating the evolution of dynamic stall on a wind turbine blade.
Question: Brittany pedaled her bicycle quickly on a level surface. When she stopped pedaling and coasted, the bicycle slowed and came to a stop. What happened to the energy of the moving bicycle as it coasted to a stop?
A) It was transferred to heat due to friction.
B) It was reabsorbed into Brittany's muscles.
C) It was stored in the bicycle wheels as potential energy.
D) It was transferred to the air in the form of visible light.
|
A) It was transferred to heat due to friction.
|
Context:
industry is making composite materials. these are structured materials composed of two or more macroscopic phases. applications range from structural elements such as steel - reinforced concrete, to the thermal insulating tiles, which play a key and integral role in nasa ' s space shuttle thermal protection system, which is used to protect the surface of the shuttle from the heat of re - entry into the earth ' s atmosphere. one example is reinforced carbon - carbon ( rcc ), the light gray material, which withstands re - entry temperatures up to 1, 510 Β°c ( 2, 750 Β°f ) and protects the space shuttle ' s wing leading edges and nose cap. rcc is a laminated composite material made from graphite rayon cloth and impregnated with a phenolic resin. after curing at high temperature in an autoclave, the laminate is pyrolized to convert the resin to carbon, impregnated with furfuryl alcohol in a vacuum chamber, and cured - pyrolized to convert the furfuryl alcohol to carbon. to provide oxidation resistance for reusability, the outer layers of the rcc are converted to silicon carbide. other examples can be seen in the " plastic " casings of television sets, cell - phones and so on. these plastic casings are usually a composite material made up of a thermoplastic matrix such as acrylonitrile butadiene styrene ( abs ) in which calcium carbonate chalk, talc, glass fibers or carbon fibers have been added for added strength, bulk, or electrostatic dispersion. these additions may be termed reinforcing fibers, or dispersants, depending on their purpose. = = = polymers = = = polymers are chemical compounds made up of a large number of identical components linked together like chains. polymers are the raw materials ( the resins ) used to make what are commonly called plastics and rubber. plastics and rubber are the final product, created after one or more polymers or additives have been added to a resin during processing, which is then shaped into a final form. plastics in former and in current widespread use include polyethylene, polypropylene, polyvinyl chloride ( pvc ), polystyrene, nylons, polyesters, acrylics, polyurethanes, and polycarbonates. rubbers include natural rubber, styrene - butadiene rubber, chloroprene, and butadiene rubber. plastics are generally classified as commodity
shuttle from the heat of re - entry into the earth ' s atmosphere. one example is reinforced carbon - carbon ( rcc ), the light gray material, which withstands re - entry temperatures up to 1, 510 Β°c ( 2, 750 Β°f ) and protects the space shuttle ' s wing leading edges and nose cap. rcc is a laminated composite material made from graphite rayon cloth and impregnated with a phenolic resin. after curing at high temperature in an autoclave, the laminate is pyrolized to convert the resin to carbon, impregnated with furfuryl alcohol in a vacuum chamber, and cured - pyrolized to convert the furfuryl alcohol to carbon. to provide oxidation resistance for reusability, the outer layers of the rcc are converted to silicon carbide. other examples can be seen in the " plastic " casings of television sets, cell - phones and so on. these plastic casings are usually a composite material made up of a thermoplastic matrix such as acrylonitrile butadiene styrene ( abs ) in which calcium carbonate chalk, talc, glass fibers or carbon fibers have been added for added strength, bulk, or electrostatic dispersion. these additions may be termed reinforcing fibers, or dispersants, depending on their purpose. = = = polymers = = = polymers are chemical compounds made up of a large number of identical components linked together like chains. polymers are the raw materials ( the resins ) used to make what are commonly called plastics and rubber. plastics and rubber are the final product, created after one or more polymers or additives have been added to a resin during processing, which is then shaped into a final form. plastics in former and in current widespread use include polyethylene, polypropylene, polyvinyl chloride ( pvc ), polystyrene, nylons, polyesters, acrylics, polyurethanes, and polycarbonates. rubbers include natural rubber, styrene - butadiene rubber, chloroprene, and butadiene rubber. plastics are generally classified as commodity, specialty and engineering plastics. polyvinyl chloride ( pvc ) is widely used, inexpensive, and annual production quantities are large. it lends itself to a vast array of applications, from artificial leather to electrical insulation and cabling, packaging, and containers. its fabrication and processing are simple and well - established.
reflect radar waves back to the emitting radar is with orthogonal metal plates, forming a corner reflector consisting of either a dihedral ( two plates ) or a trihedral ( three orthogonal plates ). this configuration occurs in the tail of a conventional aircraft, where the vertical and horizontal components of the tail are set at right angles. stealth aircraft such as the f - 117 use a different arrangement, tilting the tail surfaces to reduce corner reflections formed between them. a more radical method is to omit the tail, as in the b - 2 spirit. the b - 2 ' s clean, low - drag flying wing configuration gives it exceptional range and reduces its radar profile. the flying wing design most closely resembles a so - called infinite flat plate ( as vertical control surfaces dramatically increase rcs ), the perfect stealth shape, as it would have no angles to reflect back radar waves. in addition to altering the tail, stealth design must bury the engines within the wing or fuselage, or in some cases where stealth is applied to an extant aircraft, install baffles in the air intakes, so that the compressor blades are not visible to radar. a stealthy shape must be devoid of complex bumps or protrusions of any kind, meaning that weapons, fuel tanks, and other stores must not be carried externally. any stealthy vehicle becomes un - stealthy when a door or hatch opens. parallel alignment of edges or even surfaces is also often used in stealth designs. the technique involves using a small number of edge orientations in the shape of the structure. for example, on the f - 22a raptor, the leading edges of the wing and the tail planes are set at the same angle. other smaller structures, such as the air intake bypass doors and the air refueling aperture, also use the same angles. the effect of this is to return a narrow radar signal in a very specific direction away from the radar emitter rather than returning a diffuse signal detectable at many angles. the effect is sometimes called " glitter " after the very brief signal seen when the reflected beam passes across a detector. it can be difficult for the radar operator to distinguish between a glitter event and a digital glitch in the processing system. stealth airframes sometimes display distinctive serrations on some exposed edges, such as the engine ports. the yf - 23 has such serrations on the exhaust ports. this is another example in the parallel alignment of features, this time on the external airframe. the shaping requirements detracted greatly from the f - 117 '
building block. ceramics β not to be confused with raw, unfired clay β are usually seen in crystalline form. the vast majority of commercial glasses contain a metal oxide fused with silica. at the high temperatures used to prepare glass, the material is a viscous liquid which solidifies into a disordered state upon cooling. windowpanes and eyeglasses are important examples. fibers of glass are also used for long - range telecommunication and optical transmission. scratch resistant corning gorilla glass is a well - known example of the application of materials science to drastically improve the properties of common components. engineering ceramics are known for their stiffness and stability under high temperatures, compression and electrical stress. alumina, silicon carbide, and tungsten carbide are made from a fine powder of their constituents in a process of sintering with a binder. hot pressing provides higher density material. chemical vapor deposition can place a film of a ceramic on another material. cermets are ceramic particles containing some metals. the wear resistance of tools is derived from cemented carbides with the metal phase of cobalt and nickel typically added to modify properties. ceramics can be significantly strengthened for engineering applications using the principle of crack deflection. this process involves the strategic addition of second - phase particles within a ceramic matrix, optimizing their shape, size, and distribution to direct and control crack propagation. this approach enhances fracture toughness, paving the way for the creation of advanced, high - performance ceramics in various industries. = = = composites = = = another application of materials science in industry is making composite materials. these are structured materials composed of two or more macroscopic phases. applications range from structural elements such as steel - reinforced concrete, to the thermal insulating tiles, which play a key and integral role in nasa ' s space shuttle thermal protection system, which is used to protect the surface of the shuttle from the heat of re - entry into the earth ' s atmosphere. one example is reinforced carbon - carbon ( rcc ), the light gray material, which withstands re - entry temperatures up to 1, 510 Β°c ( 2, 750 Β°f ) and protects the space shuttle ' s wing leading edges and nose cap. rcc is a laminated composite material made from graphite rayon cloth and impregnated with a phenolic resin. after curing at high temperature in an autoclave, the laminate is pyrolized to convert the resin to carbon, impregnated with furfuryl alcohol in a
passage of carbon dioxide as aluminum and glass. = = = ceramics and glasses = = = another application of materials science is the study of ceramics and glasses, typically the most brittle materials with industrial relevance. many ceramics and glasses exhibit covalent or ionic - covalent bonding with sio2 ( silica ) as a fundamental building block. ceramics β not to be confused with raw, unfired clay β are usually seen in crystalline form. the vast majority of commercial glasses contain a metal oxide fused with silica. at the high temperatures used to prepare glass, the material is a viscous liquid which solidifies into a disordered state upon cooling. windowpanes and eyeglasses are important examples. fibers of glass are also used for long - range telecommunication and optical transmission. scratch resistant corning gorilla glass is a well - known example of the application of materials science to drastically improve the properties of common components. engineering ceramics are known for their stiffness and stability under high temperatures, compression and electrical stress. alumina, silicon carbide, and tungsten carbide are made from a fine powder of their constituents in a process of sintering with a binder. hot pressing provides higher density material. chemical vapor deposition can place a film of a ceramic on another material. cermets are ceramic particles containing some metals. the wear resistance of tools is derived from cemented carbides with the metal phase of cobalt and nickel typically added to modify properties. ceramics can be significantly strengthened for engineering applications using the principle of crack deflection. this process involves the strategic addition of second - phase particles within a ceramic matrix, optimizing their shape, size, and distribution to direct and control crack propagation. this approach enhances fracture toughness, paving the way for the creation of advanced, high - performance ceramics in various industries. = = = composites = = = another application of materials science in industry is making composite materials. these are structured materials composed of two or more macroscopic phases. applications range from structural elements such as steel - reinforced concrete, to the thermal insulating tiles, which play a key and integral role in nasa ' s space shuttle thermal protection system, which is used to protect the surface of the shuttle from the heat of re - entry into the earth ' s atmosphere. one example is reinforced carbon - carbon ( rcc ), the light gray material, which withstands re - entry temperatures up to 1, 510 Β°c ( 2, 750 Β°f ) and protects the space shuttle ' s wing leading edges and nose cap
, behind which are structures termed reentrant triangles. radar waves penetrating the skin get trapped in these structures, reflecting off the internal faces and losing energy. this method was first used on the blackbird series : a - 12, yf - 12a, lockheed sr - 71 blackbird. the most efficient way to reflect radar waves back to the emitting radar is with orthogonal metal plates, forming a corner reflector consisting of either a dihedral ( two plates ) or a trihedral ( three orthogonal plates ). this configuration occurs in the tail of a conventional aircraft, where the vertical and horizontal components of the tail are set at right angles. stealth aircraft such as the f - 117 use a different arrangement, tilting the tail surfaces to reduce corner reflections formed between them. a more radical method is to omit the tail, as in the b - 2 spirit. the b - 2 ' s clean, low - drag flying wing configuration gives it exceptional range and reduces its radar profile. the flying wing design most closely resembles a so - called infinite flat plate ( as vertical control surfaces dramatically increase rcs ), the perfect stealth shape, as it would have no angles to reflect back radar waves. in addition to altering the tail, stealth design must bury the engines within the wing or fuselage, or in some cases where stealth is applied to an extant aircraft, install baffles in the air intakes, so that the compressor blades are not visible to radar. a stealthy shape must be devoid of complex bumps or protrusions of any kind, meaning that weapons, fuel tanks, and other stores must not be carried externally. any stealthy vehicle becomes un - stealthy when a door or hatch opens. parallel alignment of edges or even surfaces is also often used in stealth designs. the technique involves using a small number of edge orientations in the shape of the structure. for example, on the f - 22a raptor, the leading edges of the wing and the tail planes are set at the same angle. other smaller structures, such as the air intake bypass doors and the air refueling aperture, also use the same angles. the effect of this is to return a narrow radar signal in a very specific direction away from the radar emitter rather than returning a diffuse signal detectable at many angles. the effect is sometimes called " glitter " after the very brief signal seen when the reflected beam passes across a detector. it can be difficult for the radar operator to distinguish between a glitter event and a digital glitch in the processing system. stealth air
, specialty and engineering plastics. polyvinyl chloride ( pvc ) is widely used, inexpensive, and annual production quantities are large. it lends itself to a vast array of applications, from artificial leather to electrical insulation and cabling, packaging, and containers. its fabrication and processing are simple and well - established. the versatility of pvc is due to the wide range of plasticisers and other additives that it accepts. the term " additives " in polymer science refers to the chemicals and compounds added to the polymer base to modify its material properties. polycarbonate would be normally considered an engineering plastic ( other examples include peek, abs ). such plastics are valued for their superior strengths and other special material properties. they are usually not used for disposable applications, unlike commodity plastics. specialty plastics are materials with unique characteristics, such as ultra - high strength, electrical conductivity, electro - fluorescence, high thermal stability, etc. the dividing lines between the various types of plastics is not based on material but rather on their properties and applications. for example, polyethylene ( pe ) is a cheap, low friction polymer commonly used to make disposable bags for shopping and trash, and is considered a commodity plastic, whereas medium - density polyethylene ( mdpe ) is used for underground gas and water pipes, and another variety called ultra - high - molecular - weight polyethylene ( uhmwpe ) is an engineering plastic which is used extensively as the glide rails for industrial equipment and the low - friction socket in implanted hip joints. = = = metal alloys = = = the alloys of iron ( steel, stainless steel, cast iron, tool steel, alloy steels ) make up the largest proportion of metals today both by quantity and commercial value. iron alloyed with various proportions of carbon gives low, mid and high carbon steels. an iron - carbon alloy is only considered steel if the carbon level is between 0. 01 % and 2. 00 % by weight. for steels, the hardness and tensile strength of the steel is related to the amount of carbon present, with increasing carbon levels also leading to lower ductility and toughness. heat treatment processes such as quenching and tempering can significantly change these properties, however. in contrast, certain metal alloys exhibit unique properties where their size and density remain unchanged across a range of temperatures. cast iron is defined as an iron β carbon alloy with more than 2. 00 %, but less than 6. 67 %
emitter rather than returning a diffuse signal detectable at many angles. the effect is sometimes called " glitter " after the very brief signal seen when the reflected beam passes across a detector. it can be difficult for the radar operator to distinguish between a glitter event and a digital glitch in the processing system. stealth airframes sometimes display distinctive serrations on some exposed edges, such as the engine ports. the yf - 23 has such serrations on the exhaust ports. this is another example in the parallel alignment of features, this time on the external airframe. the shaping requirements detracted greatly from the f - 117 ' s aerodynamic properties. it is inherently unstable, and cannot be flown without a fly - by - wire control system. similarly, coating the cockpit canopy with a thin film transparent conductor ( vapor - deposited gold or indium tin oxide ) helps to reduce the aircraft ' s radar profile, because radar waves would normally enter the cockpit, reflect off objects ( the inside of a cockpit has a complex shape, with a pilot helmet alone forming a sizeable return ), and possibly return to the radar, but the conductive coating creates a controlled shape that deflects the incoming radar waves away from the radar. the coating is thin enough that it has no adverse effect on pilot vision. = = = = ships = = = = ships have also adopted similar methods. though the earlier american arleigh burke - class destroyers incorporated some signature - reduction features. the norwegian skjold - class corvettes was the first coastal defence and the french la fayette - class frigates the first ocean - going stealth ships to enter service. other examples are the dutch de zeven provincien - class frigates, the taiwanese tuo chiang - class corvettes, german sachsen - class frigates, the swedish visby - class corvette, the american san antonio - class amphibious transport docks, and most modern warship designs. = = = materials = = = = = = = non - metallic airframe = = = = dielectric composite materials are more transparent to radar, whereas electrically conductive materials such as metals and carbon fibers reflect electromagnetic energy incident on the material ' s surface. composites may also contain ferrites to optimize the dielectric and magnetic properties of a material for its application. = = = = radar - absorbent material = = = = radiation - absorbent material ( ram ), often as paints, are used especially on the edges of metal surfaces. while the material and thickness of ram coatings can
bear ' ) was conspicuous on radar. it is now known that propellers and jet turbine blades produce a bright radar image ; the bear has four pairs of large 18 - foot ( 5. 6 m ) diameter contra - rotating propellers. another important factor is internal construction. some stealth aircraft have skin that is radar transparent or absorbing, behind which are structures termed reentrant triangles. radar waves penetrating the skin get trapped in these structures, reflecting off the internal faces and losing energy. this method was first used on the blackbird series : a - 12, yf - 12a, lockheed sr - 71 blackbird. the most efficient way to reflect radar waves back to the emitting radar is with orthogonal metal plates, forming a corner reflector consisting of either a dihedral ( two plates ) or a trihedral ( three orthogonal plates ). this configuration occurs in the tail of a conventional aircraft, where the vertical and horizontal components of the tail are set at right angles. stealth aircraft such as the f - 117 use a different arrangement, tilting the tail surfaces to reduce corner reflections formed between them. a more radical method is to omit the tail, as in the b - 2 spirit. the b - 2 ' s clean, low - drag flying wing configuration gives it exceptional range and reduces its radar profile. the flying wing design most closely resembles a so - called infinite flat plate ( as vertical control surfaces dramatically increase rcs ), the perfect stealth shape, as it would have no angles to reflect back radar waves. in addition to altering the tail, stealth design must bury the engines within the wing or fuselage, or in some cases where stealth is applied to an extant aircraft, install baffles in the air intakes, so that the compressor blades are not visible to radar. a stealthy shape must be devoid of complex bumps or protrusions of any kind, meaning that weapons, fuel tanks, and other stores must not be carried externally. any stealthy vehicle becomes un - stealthy when a door or hatch opens. parallel alignment of edges or even surfaces is also often used in stealth designs. the technique involves using a small number of edge orientations in the shape of the structure. for example, on the f - 22a raptor, the leading edges of the wing and the tail planes are set at the same angle. other smaller structures, such as the air intake bypass doors and the air refueling aperture, also use the same angles. the effect of this is to return a narrow radar signal in a very specific direction away from the radar
chloroplasts make are used for many things, such as providing material to build cell membranes out of and making the polymer cutin which is found in the plant cuticle that protects land plants from drying out. plants synthesise a number of unique polymers like the polysaccharide molecules cellulose, pectin and xyloglucan from which the land plant cell wall is constructed. vascular land plants make lignin, a polymer used to strengthen the secondary cell walls of xylem tracheids and vessels to keep them from collapsing when a plant sucks water through them under water stress. lignin is also used in other cell types like sclerenchyma fibres that provide structural support for a plant and is a major constituent of wood. sporopollenin is a chemically resistant polymer found in the outer cell walls of spores and pollen of land plants responsible for the survival of early land plant spores and the pollen of seed plants in the fossil record. it is widely regarded as a marker for the start of land plant evolution during the ordovician period. the concentration of carbon dioxide in the atmosphere today is much lower than it was when plants emerged onto land during the ordovician and silurian periods. many monocots like maize and the pineapple and some dicots like the asteraceae have since independently evolved pathways like crassulacean acid metabolism and the c4 carbon fixation pathway for photosynthesis which avoid the losses resulting from photorespiration in the more common c3 carbon fixation pathway. these biochemical strategies are unique to land plants. = = = medicine and materials = = = phytochemistry is a branch of plant biochemistry primarily concerned with the chemical substances produced by plants during secondary metabolism. some of these compounds are toxins such as the alkaloid coniine from hemlock. others, such as the essential oils peppermint oil and lemon oil are useful for their aroma, as flavourings and spices ( e. g., capsaicin ), and in medicine as pharmaceuticals as in opium from opium poppies. many medicinal and recreational drugs, such as tetrahydrocannabinol ( active ingredient in cannabis ), caffeine, morphine and nicotine come directly from plants. others are simple derivatives of botanical natural products. for example, the pain killer aspirin is the acetyl ester of salicylic acid, originally isolated from the bark of willow trees, and
Question: Homes that are built to be environmentally friendly because they use energy more efficiently than other homes are called "green" homes. "Green" homes often have reflective roofs and walls made of recycled materials. The windows in these energy-saving homes are double-paned, meaning each window has two pieces of glass. Double-paned windows have a layer of air between the window panes. This layer is a barrier against extreme temperatures and saves energy. A solar panel on a "green" home uses
A) chemical energy
B) mechanical energy
C) a renewable energy source
D) a nonrenewable energy source
|
C) a renewable energy source
|
Context:
##ctonics, mountain ranges, volcanoes, and earthquakes are geological phenomena that can be explained in terms of physical and chemical processes in the earth ' s crust. beneath the earth ' s crust lies the mantle which is heated by the radioactive decay of heavy elements. the mantle is not quite solid and consists of magma which is in a state of semi - perpetual convection. this convection process causes the lithospheric plates to move, albeit slowly. the resulting process is known as plate tectonics. areas of the crust where new crust is created are called divergent boundaries, those where it is brought back into the earth are convergent boundaries and those where plates slide past each other, but no new lithospheric material is created or destroyed, are referred to as transform ( or conservative ) boundaries. earthquakes result from the movement of the lithospheric plates, and they often occur near convergent boundaries where parts of the crust are forced into the earth as part of subduction. plate tectonics might be thought of as the process by which the earth is resurfaced. as the result of seafloor spreading, new crust and lithosphere is created by the flow of magma from the mantle to the near surface, through fissures, where it cools and solidifies. through subduction, oceanic crust and lithosphere vehemently returns to the convecting mantle. volcanoes result primarily from the melting of subducted crust material. crust material that is forced into the asthenosphere melts, and some portion of the melted material becomes light enough to rise to the surface β giving birth to volcanoes. = = atmospheric science = = atmospheric science initially developed in the late - 19th century as a means to forecast the weather through meteorology, the study of weather. atmospheric chemistry was developed in the 20th century to measure air pollution and expanded in the 1970s in response to acid rain. climatology studies the climate and climate change. the troposphere, stratosphere, mesosphere, thermosphere, and exosphere are the five layers which make up earth ' s atmosphere. 75 % of the mass in the atmosphere is located within the troposphere, the lowest layer. in all, the atmosphere is made up of about 78. 0 % nitrogen, 20. 9 % oxygen, and 0. 92 % argon, and small amounts of other gases including co2 and water vapor. water vapor and co2 cause the earth ' s atmosphere to catch and hold the sun ' s
##morphology studies the origin of landscapes. structural geology studies the deformation of rocks to produce mountains and lowlands. resource geology studies how energy resources can be obtained from minerals. environmental geology studies how pollution and contaminants affect soil and rock. mineralogy is the study of minerals and includes the study of mineral formation, crystal structure, hazards associated with minerals, and the physical and chemical properties of minerals. petrology is the study of rocks, including the formation and composition of rocks. petrography is a branch of petrology that studies the typology and classification of rocks. = = earth ' s interior = = plate tectonics, mountain ranges, volcanoes, and earthquakes are geological phenomena that can be explained in terms of physical and chemical processes in the earth ' s crust. beneath the earth ' s crust lies the mantle which is heated by the radioactive decay of heavy elements. the mantle is not quite solid and consists of magma which is in a state of semi - perpetual convection. this convection process causes the lithospheric plates to move, albeit slowly. the resulting process is known as plate tectonics. areas of the crust where new crust is created are called divergent boundaries, those where it is brought back into the earth are convergent boundaries and those where plates slide past each other, but no new lithospheric material is created or destroyed, are referred to as transform ( or conservative ) boundaries. earthquakes result from the movement of the lithospheric plates, and they often occur near convergent boundaries where parts of the crust are forced into the earth as part of subduction. plate tectonics might be thought of as the process by which the earth is resurfaced. as the result of seafloor spreading, new crust and lithosphere is created by the flow of magma from the mantle to the near surface, through fissures, where it cools and solidifies. through subduction, oceanic crust and lithosphere vehemently returns to the convecting mantle. volcanoes result primarily from the melting of subducted crust material. crust material that is forced into the asthenosphere melts, and some portion of the melted material becomes light enough to rise to the surface β giving birth to volcanoes. = = atmospheric science = = atmospheric science initially developed in the late - 19th century as a means to forecast the weather through meteorology, the study of weather. atmospheric chemistry was developed in the 20th century to measure air pollution and expanded in the 1970s in response to
, crystal structure, hazards associated with minerals, and the physical and chemical properties of minerals. petrology is the study of rocks, including the formation and composition of rocks. petrography is a branch of petrology that studies the typology and classification of rocks. = = earth ' s interior = = plate tectonics, mountain ranges, volcanoes, and earthquakes are geological phenomena that can be explained in terms of physical and chemical processes in the earth ' s crust. beneath the earth ' s crust lies the mantle which is heated by the radioactive decay of heavy elements. the mantle is not quite solid and consists of magma which is in a state of semi - perpetual convection. this convection process causes the lithospheric plates to move, albeit slowly. the resulting process is known as plate tectonics. areas of the crust where new crust is created are called divergent boundaries, those where it is brought back into the earth are convergent boundaries and those where plates slide past each other, but no new lithospheric material is created or destroyed, are referred to as transform ( or conservative ) boundaries. earthquakes result from the movement of the lithospheric plates, and they often occur near convergent boundaries where parts of the crust are forced into the earth as part of subduction. plate tectonics might be thought of as the process by which the earth is resurfaced. as the result of seafloor spreading, new crust and lithosphere is created by the flow of magma from the mantle to the near surface, through fissures, where it cools and solidifies. through subduction, oceanic crust and lithosphere vehemently returns to the convecting mantle. volcanoes result primarily from the melting of subducted crust material. crust material that is forced into the asthenosphere melts, and some portion of the melted material becomes light enough to rise to the surface β giving birth to volcanoes. = = atmospheric science = = atmospheric science initially developed in the late - 19th century as a means to forecast the weather through meteorology, the study of weather. atmospheric chemistry was developed in the 20th century to measure air pollution and expanded in the 1970s in response to acid rain. climatology studies the climate and climate change. the troposphere, stratosphere, mesosphere, thermosphere, and exosphere are the five layers which make up earth ' s atmosphere. 75 % of the mass in the atmosphere is located within the troposphere, the lowest
made of steel. the shoe is generally wider than the caisson to reduce friction, and the leading edge may be supplied with pressurised bentonite slurry, which swells in water, stabilizing settlement by filling depressions and voids. an open caisson may fill with water during sinking. the material is excavated by clamshell excavator bucket on crane. the formation level subsoil may still not be suitable for excavation or bearing capacity. the water in the caisson ( due to a high water table ) balances the upthrust forces of the soft soils underneath. if dewatered, the base may " pipe " or " boil ", causing the caisson to sink. to combat this problem, piles may be driven from the surface to act as : load - bearing walls, in that they transmit loads to deeper soils. anchors, in that they resist flotation because of the friction at the interface between their surfaces and the surrounding earth into which they have been driven. h - beam sections ( typical column sections, due to resistance to bending in all axis ) may be driven at angles " raked " to rock or other firmer soils ; the h - beams are left extended above the base. a reinforced concrete plug may be placed under the water, a process known as tremie concrete placement. when the caisson is dewatered, this plug acts as a pile cap, resisting the upward forces of the subsoil. = = = monolithic = = = a monolithic caisson ( or simply a monolith ) is larger than the other types of caisson, but similar to open caissons. such caissons are often found in quay walls, where resistance to impact from ships is required. = = = pneumatic = = = shallow caissons may be open to the air, whereas pneumatic caissons ( sometimes called pressurized caissons ), which penetrate soft mud, are bottomless boxes sealed at the top and filled with compressed air to keep water and mud out at depth. an airlock allows access to the chamber. workers, called sandhogs in american english, move mud and rock debris ( called muck ) from the edge of the workspace to a water - filled pit, connected by a tube ( called the muck tube ) to the surface. a crane at the surface removes the soil with a clamshell bucket. the water pressure in the tube balances the air pressure, with excess air escaping up
to be separated conceptually from geology and crop production and treated as a whole. as a founding father of soil science, fallou has primacy in time. fallou was working on the origins of soil before dokuchaev was born ; however dokuchaev ' s work was more extensive and is considered to be the more significant to modern soil theory than fallou ' s. previously, soil had been considered a product of chemical transformations of rocks, a dead substrate from which plants derive nutritious elements. soil and bedrock were in fact equated. dokuchaev considers the soil as a natural body having its own genesis and its own history of development, a body with complex and multiform processes taking place within it. the soil is considered as different from bedrock. the latter becomes soil under the influence of a series of soil - formation factors ( climate, vegetation, country, relief and age ). according to him, soil should be called the " daily " or outward horizons of rocks regardless of the type ; they are changed naturally by the common effect of water, air and various kinds of living and dead organisms. a 1914 encyclopedic definition : " the different forms of earth on the surface of the rocks, formed by the breaking down or weathering of rocks ". serves to illustrate the historic view of soil which persisted from the 19th century. dokuchaev ' s late 19th century soil concept developed in the 20th century to one of soil as earthy material that has been altered by living processes. a corollary concept is that soil without a living component is simply a part of earth ' s outer layer. further refinement of the soil concept is occurring in view of an appreciation of energy transport and transformation within soil. the term is popularly applied to the material on the surface of the earth ' s moon and mars, a usage acceptable within a portion of the scientific community. accurate to this modern understanding of soil is nikiforoff ' s 1959 definition of soil as the " excited skin of the sub aerial part of the earth ' s crust ". = = areas of practice = = academically, soil scientists tend to be drawn to one of five areas of specialization : microbiology, pedology, edaphology, physics, or chemistry. yet the work specifics are very much dictated by the challenges facing our civilization ' s desire to sustain the land that supports it, and the distinctions between the sub - disciplines of soil science often blur in the process. soil science professionals commonly stay current
of tool usage was found in ethiopia within the great rift valley, dating back to 2. 5 million years ago. the earliest methods of stone tool making, known as the oldowan " industry ", date back to at least 2. 3 million years ago. this era of stone tool use is called the paleolithic, or " old stone age ", and spans all of human history up to the development of agriculture approximately 12, 000 years ago. to make a stone tool, a " core " of hard stone with specific flaking properties ( such as flint ) was struck with a hammerstone. this flaking produced sharp edges which could be used as tools, primarily in the form of choppers or scrapers. these tools greatly aided the early humans in their hunter - gatherer lifestyle to perform a variety of tasks including butchering carcasses ( and breaking bones to get at the marrow ) ; chopping wood ; cracking open nuts ; skinning an animal for its hide, and even forming other tools out of softer materials such as bone and wood. the earliest stone tools were irrelevant, being little more than a fractured rock. in the acheulian era, beginning approximately 1. 65 million years ago, methods of working these stones into specific shapes, such as hand axes emerged. this early stone age is described as the lower paleolithic. the middle paleolithic, approximately 300, 000 years ago, saw the introduction of the prepared - core technique, where multiple blades could be rapidly formed from a single core stone. the upper paleolithic, beginning approximately 40, 000 years ago, saw the introduction of pressure flaking, where a wood, bone, or antler punch could be used to shape a stone very finely. the end of the last ice age about 10, 000 years ago is taken as the end point of the upper paleolithic and the beginning of the epipaleolithic / mesolithic. the mesolithic technology included the use of microliths as composite stone tools, along with wood, bone, and antler tools. the later stone age, during which the rudiments of agricultural technology were developed, is called the neolithic period. during this period, polished stone tools were made from a variety of hard rocks such as flint, jade, jadeite, and greenstone, largely by working exposures as quarries, but later the valuable rocks were pursued by tunneling underground, the first steps in mining technology. the polished axes were used for forest clearance and the establishment of crop
be the more significant to modern soil theory than fallou ' s. previously, soil had been considered a product of chemical transformations of rocks, a dead substrate from which plants derive nutritious elements. soil and bedrock were in fact equated. dokuchaev considers the soil as a natural body having its own genesis and its own history of development, a body with complex and multiform processes taking place within it. the soil is considered as different from bedrock. the latter becomes soil under the influence of a series of soil - formation factors ( climate, vegetation, country, relief and age ). according to him, soil should be called the " daily " or outward horizons of rocks regardless of the type ; they are changed naturally by the common effect of water, air and various kinds of living and dead organisms. a 1914 encyclopedic definition : " the different forms of earth on the surface of the rocks, formed by the breaking down or weathering of rocks ". serves to illustrate the historic view of soil which persisted from the 19th century. dokuchaev ' s late 19th century soil concept developed in the 20th century to one of soil as earthy material that has been altered by living processes. a corollary concept is that soil without a living component is simply a part of earth ' s outer layer. further refinement of the soil concept is occurring in view of an appreciation of energy transport and transformation within soil. the term is popularly applied to the material on the surface of the earth ' s moon and mars, a usage acceptable within a portion of the scientific community. accurate to this modern understanding of soil is nikiforoff ' s 1959 definition of soil as the " excited skin of the sub aerial part of the earth ' s crust ". = = areas of practice = = academically, soil scientists tend to be drawn to one of five areas of specialization : microbiology, pedology, edaphology, physics, or chemistry. yet the work specifics are very much dictated by the challenges facing our civilization ' s desire to sustain the land that supports it, and the distinctions between the sub - disciplines of soil science often blur in the process. soil science professionals commonly stay current in soil chemistry, soil physics, soil microbiology, pedology, and applied soil science in related disciplines. one exciting effort drawing in soil scientists in the u. s. as of 2004 is the soil quality initiative. central to the soil quality initiative is developing indices of soil health and then monitoring them in a way
a state of semi - perpetual convection. this convection process causes the lithospheric plates to move, albeit slowly. the resulting process is known as plate tectonics. areas of the crust where new crust is created are called divergent boundaries, those where it is brought back into the earth are convergent boundaries and those where plates slide past each other, but no new lithospheric material is created or destroyed, are referred to as transform ( or conservative ) boundaries. earthquakes result from the movement of the lithospheric plates, and they often occur near convergent boundaries where parts of the crust are forced into the earth as part of subduction. plate tectonics might be thought of as the process by which the earth is resurfaced. as the result of seafloor spreading, new crust and lithosphere is created by the flow of magma from the mantle to the near surface, through fissures, where it cools and solidifies. through subduction, oceanic crust and lithosphere vehemently returns to the convecting mantle. volcanoes result primarily from the melting of subducted crust material. crust material that is forced into the asthenosphere melts, and some portion of the melted material becomes light enough to rise to the surface β giving birth to volcanoes. = = atmospheric science = = atmospheric science initially developed in the late - 19th century as a means to forecast the weather through meteorology, the study of weather. atmospheric chemistry was developed in the 20th century to measure air pollution and expanded in the 1970s in response to acid rain. climatology studies the climate and climate change. the troposphere, stratosphere, mesosphere, thermosphere, and exosphere are the five layers which make up earth ' s atmosphere. 75 % of the mass in the atmosphere is located within the troposphere, the lowest layer. in all, the atmosphere is made up of about 78. 0 % nitrogen, 20. 9 % oxygen, and 0. 92 % argon, and small amounts of other gases including co2 and water vapor. water vapor and co2 cause the earth ' s atmosphere to catch and hold the sun ' s energy through the greenhouse effect. this makes earth ' s surface warm enough for liquid water and life. in addition to trapping heat, the atmosphere also protects living organisms by shielding the earth ' s surface from cosmic rays. the magnetic field β created by the internal motions of the core β produces the magnetosphere which protects earth '
also known as the gradient or slope. when two rivers of different sizes have the same fall, the larger river has the quicker flow, as its retardation by friction against its bed and banks is less in proportion to its volume than is the case with the smaller river. the fall available in a section of a river approximately corresponds to the slope of the country it traverses ; as rivers rise close to the highest part of their basins, generally in hilly regions, their fall is rapid near their source and gradually diminishes, with occasional irregularities, until, in traversing plains along the latter part of their course, their fall usually becomes quite gentle. accordingly, in large basins, rivers in most cases begin as torrents with a variable flow, and end as gently flowing rivers with a comparatively regular discharge. the irregular flow of rivers throughout their course forms one of the main difficulties in devising works for mitigating inundations or for increasing the navigable capabilities of rivers. in tropical countries subject to periodical rains, the rivers are in flood during the rainy season and have hardly any flow during the rest of the year, while in temperate regions, where the rainfall is more evenly distributed throughout the year, evaporation causes the available rainfall to be much less in hot summer weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in
a watershed ( called a " divide " in north america ) over which rainfall flows down towards the river traversing the lowest part of the valley, whereas the rain falling on the far slope of the watershed flows away to another river draining an adjacent basin. river basins vary in extent according to the configuration of the country, ranging from the insignificant drainage areas of streams rising on high ground near the coast and flowing straight down into the sea, up to immense tracts of continents, where rivers rising on the slopes of mountain ranges far inland have to traverse vast stretches of valleys and plains before reaching the ocean. the size of the largest river basin of any country depends on the extent of the continent in which it is situated, its position in relation to the hilly regions in which rivers generally arise and the sea into which they flow, and the distance between the source and the outlet into the sea of the river draining it. the rate of flow of rivers depends mainly upon their fall, also known as the gradient or slope. when two rivers of different sizes have the same fall, the larger river has the quicker flow, as its retardation by friction against its bed and banks is less in proportion to its volume than is the case with the smaller river. the fall available in a section of a river approximately corresponds to the slope of the country it traverses ; as rivers rise close to the highest part of their basins, generally in hilly regions, their fall is rapid near their source and gradually diminishes, with occasional irregularities, until, in traversing plains along the latter part of their course, their fall usually becomes quite gentle. accordingly, in large basins, rivers in most cases begin as torrents with a variable flow, and end as gently flowing rivers with a comparatively regular discharge. the irregular flow of rivers throughout their course forms one of the main difficulties in devising works for mitigating inundations or for increasing the navigable capabilities of rivers. in tropical countries subject to periodical rains, the rivers are in flood during the rainy season and have hardly any flow during the rest of the year, while in temperate regions, where the rainfall is more evenly distributed throughout the year, evaporation causes the available rainfall to be much less in hot summer weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern
Question: Which process occurs when rocks and soil slide down a mountain?
A) erosion
B) migration
C) recycling
D) maintenance
|
A) erosion
|
Context:
. species boundaries in plants may be weaker than in animals, and cross species hybrids are often possible. a familiar example is peppermint, mentha Γ piperita, a sterile hybrid between mentha aquatica and spearmint, mentha spicata. the many cultivated varieties of wheat are the result of multiple inter - and intra - specific crosses between wild species and their hybrids. angiosperms with monoecious flowers often have self - incompatibility mechanisms that operate between the pollen and stigma so that the pollen either fails to reach the stigma or fails to germinate and produce male gametes. this is one of several methods used by plants to promote outcrossing. in many land plants the male and female gametes are produced by separate individuals. these species are said to be dioecious when referring to vascular plant sporophytes and dioicous when referring to bryophyte gametophytes. charles darwin in his 1878 book the effects of cross and self - fertilization in the vegetable kingdom at the start of chapter xii noted " the first and most important of the conclusions which may be drawn from the observations given in this volume, is that generally cross - fertilisation is beneficial and self - fertilisation often injurious, at least with the plants on which i experimented. " an important adaptive benefit of outcrossing is that it allows the masking of deleterious mutations in the genome of progeny. this beneficial effect is also known as hybrid vigor or heterosis. once outcrossing is established, subsequent switching to inbreeding becomes disadvantageous since it allows expression of the previously masked deleterious recessive mutations, commonly referred to as inbreeding depression. unlike in higher animals, where parthenogenesis is rare, asexual reproduction may occur in plants by several different mechanisms. the formation of stem tubers in potato is one example. particularly in arctic or alpine habitats, where opportunities for fertilisation of flowers by animals are rare, plantlets or bulbs, may develop instead of flowers, replacing sexual reproduction with asexual reproduction and giving rise to clonal populations genetically identical to the parent. this is one of several types of apomixis that occur in plants. apomixis can also happen in a seed, producing a seed that contains an embryo genetically identical to the parent. most sexually reproducing organisms are diploid, with paired chromosomes, but doubling of their chromosome number may occur due to errors in
##tase, human chorionic gonadotrophin, Ξ± - fetoprotein and others are organ - associated antigens and the production of monoclonal antibodies against these antigens helps in determining the nature of a primary tumor. monoclonal antibodies are especially useful in distinguishing morphologically similar lesions, like pleural and peritoneal mesothelioma, adenocarcinoma, and in the determination of the organ or tissue origin of undifferentiated metastases. selected monoclonal antibodies help in the detection of occult metastases ( cancer of unknown primary origin ) by immuno - cytological analysis of bone marrow, other tissue aspirates, as well as lymph nodes and other tissues and can have increased sensitivity over normal histopathological staining. one study performed a sensitive immuno - histochemical assay on bone marrow aspirates of 20 patients with localized prostate cancer. three monoclonal antibodies ( t16, c26, and ae - 1 ), capable of recognizing membrane and cytoskeletal antigens expressed by epithelial cells to detect tumour cells, were used in the assay. bone marrow aspirates of 22 % of patients with localized prostate cancer ( stage b, 0 / 5 ; stage c, 2 / 4 ), and 36 % patients with metastatic prostate cancer ( stage d1, 0 / 7 patients ; stage d2, 4 / 4 patients ) had antigen - positive cells in their bone marrow. it was concluded that immuno - histochemical staining of bone marrow aspirates are very useful to detect occult bone marrow metastases in patients with apparently localized prostate cancer. although immuno - cytochemistry using tumor - associated monoclonal antibodies has led to an improved ability to detect occult breast cancer cells in bone marrow aspirates and peripheral blood, further development of this method is necessary before it can be used routinely. one major drawback of immuno - cytochemistry is that only tumor - associated and not tumor - specific monoclonal antibodies are used, and as a result, some cross - reaction with normal cells can occur. in order to effectively stage breast cancer and assess the efficacy of purging regimens prior to autologous stem cell infusion, it is important to detect even small quantities of breast cancer cells. immuno - histochemical methods are ideal for this purpose because they are simple, sensitive, and quite specific
inter - and intra - specific crosses between wild species and their hybrids. angiosperms with monoecious flowers often have self - incompatibility mechanisms that operate between the pollen and stigma so that the pollen either fails to reach the stigma or fails to germinate and produce male gametes. this is one of several methods used by plants to promote outcrossing. in many land plants the male and female gametes are produced by separate individuals. these species are said to be dioecious when referring to vascular plant sporophytes and dioicous when referring to bryophyte gametophytes. charles darwin in his 1878 book the effects of cross and self - fertilization in the vegetable kingdom at the start of chapter xii noted " the first and most important of the conclusions which may be drawn from the observations given in this volume, is that generally cross - fertilisation is beneficial and self - fertilisation often injurious, at least with the plants on which i experimented. " an important adaptive benefit of outcrossing is that it allows the masking of deleterious mutations in the genome of progeny. this beneficial effect is also known as hybrid vigor or heterosis. once outcrossing is established, subsequent switching to inbreeding becomes disadvantageous since it allows expression of the previously masked deleterious recessive mutations, commonly referred to as inbreeding depression. unlike in higher animals, where parthenogenesis is rare, asexual reproduction may occur in plants by several different mechanisms. the formation of stem tubers in potato is one example. particularly in arctic or alpine habitats, where opportunities for fertilisation of flowers by animals are rare, plantlets or bulbs, may develop instead of flowers, replacing sexual reproduction with asexual reproduction and giving rise to clonal populations genetically identical to the parent. this is one of several types of apomixis that occur in plants. apomixis can also happen in a seed, producing a seed that contains an embryo genetically identical to the parent. most sexually reproducing organisms are diploid, with paired chromosomes, but doubling of their chromosome number may occur due to errors in cytokinesis. this can occur early in development to produce an autopolyploid or partly autopolyploid organism, or during normal processes of cellular differentiation to produce some cell types that are polyploid ( endopolyploidy ), or during gamete formation. an allopolyploid
##angulation from bearings taken by two rdf stations separated geographically, as the point where the two bearing lines cross, this is called a " fix ". military forces use rdf to locate enemy forces by their tactical radio transmissions, counterintelligence services use it to locate clandestine transmitters used by espionage agents, and governments use it to locate unlicensed transmitters or interference sources. older rdf receivers used rotatable loop antennas, the antenna is rotated until the radio signal strength is weakest, indicating the transmitter is in one of the antenna ' s two nulls. the nulls are used since they are sharper than the antenna ' s lobes ( maxima ). more modern receivers use phased array antennas which have a much greater angular resolution. animal migration tracking β a widely used technique in wildlife biology, conservation biology, and wildlife management in which small battery - powered radio transmitters are attached to wild animals so their movements can be tracked with a directional rdf receiver. sometimes the transmitter is implanted in the animal. the vhf band is typically used since antennas in this band are fairly compact. the receiver has a directional antenna ( typically a small yagi ) which is rotated until the received signal is strongest ; at this point the antenna is pointing in the direction of the animal. sophisticated systems used in recent years use satellites to track the animal, or geolocation tags with gps receivers which record and transmit a log of the animal ' s location. = = = = remote control = = = = radio remote control is the use of electronic control signals sent by radio waves from a transmitter to control the actions of a device at a remote location. remote control systems may also include telemetry channels in the other direction, used to transmit real - time information on the state of the device back to the control station. uncrewed spacecraft are an example of remote - controlled machines, controlled by commands transmitted by satellite ground stations. most handheld remote controls used to control consumer electronics products like televisions or dvd players actually operate by infrared light rather than radio waves, so are not examples of radio remote control. a security concern with remote control systems is spoofing, in which an unauthorized person transmits an imitation of the control signal to take control of the device. examples of radio remote control : unmanned aerial vehicle ( uav, drone ) β a drone is an aircraft without an onboard pilot, flown by remote control by a pilot in another location, usually in a piloting station on the ground. they are used by the military for reconnaissance and ground attack, and
inherited traits such as shape in pisum sativum ( peas ). what mendel learned from studying plants has had far - reaching benefits outside of botany. similarly, " jumping genes " were discovered by barbara mcclintock while she was studying maize. nevertheless, there are some distinctive genetic differences between plants and other organisms. species boundaries in plants may be weaker than in animals, and cross species hybrids are often possible. a familiar example is peppermint, mentha Γ piperita, a sterile hybrid between mentha aquatica and spearmint, mentha spicata. the many cultivated varieties of wheat are the result of multiple inter - and intra - specific crosses between wild species and their hybrids. angiosperms with monoecious flowers often have self - incompatibility mechanisms that operate between the pollen and stigma so that the pollen either fails to reach the stigma or fails to germinate and produce male gametes. this is one of several methods used by plants to promote outcrossing. in many land plants the male and female gametes are produced by separate individuals. these species are said to be dioecious when referring to vascular plant sporophytes and dioicous when referring to bryophyte gametophytes. charles darwin in his 1878 book the effects of cross and self - fertilization in the vegetable kingdom at the start of chapter xii noted " the first and most important of the conclusions which may be drawn from the observations given in this volume, is that generally cross - fertilisation is beneficial and self - fertilisation often injurious, at least with the plants on which i experimented. " an important adaptive benefit of outcrossing is that it allows the masking of deleterious mutations in the genome of progeny. this beneficial effect is also known as hybrid vigor or heterosis. once outcrossing is established, subsequent switching to inbreeding becomes disadvantageous since it allows expression of the previously masked deleterious recessive mutations, commonly referred to as inbreeding depression. unlike in higher animals, where parthenogenesis is rare, asexual reproduction may occur in plants by several different mechanisms. the formation of stem tubers in potato is one example. particularly in arctic or alpine habitats, where opportunities for fertilisation of flowers by animals are rare, plantlets or bulbs, may develop instead of flowers, replacing sexual reproduction with asexual reproduction and giving rise to clonal populations genetically identical to the parent. this is one
planetary nebulae retain the signature of the nucleosynthesis and mixing events that occurred during the previous agb phase. observational signatures complement observations of agb and post - agb stars and their binary companions. the abundances of the elements heavier than iron such as kr and xe in planetary nebulae can be used to complement abundances of sr / y / zr and ba / la / ce in agb stars, respectively, to determine the operation of the slow neutron - capture process ( the s process ) in agb stars. additionally, observations of the rb abundance in type i planetary nebulae may allow us to infer the initial mass of the central star. several noble gas components present in meteoritic stardust silicon carbide ( sic ) grains are associated with implantation into the dust grains in the high - energy environment connected to the fast winds from the central stars during the planetary nebulae phase.
##rates and peripheral blood, further development of this method is necessary before it can be used routinely. one major drawback of immuno - cytochemistry is that only tumor - associated and not tumor - specific monoclonal antibodies are used, and as a result, some cross - reaction with normal cells can occur. in order to effectively stage breast cancer and assess the efficacy of purging regimens prior to autologous stem cell infusion, it is important to detect even small quantities of breast cancer cells. immuno - histochemical methods are ideal for this purpose because they are simple, sensitive, and quite specific. franklin et al. performed a sensitive immuno - cytochemical assay by using a combination of four monoclonal antibodies ( 260f9, 520c9, 317g5 and bre - 3 ) against tumor cell surface glycoproteins to identify breast tumour cells in bone marrow and peripheral blood. they concluded from the results that immuno - cytochemical staining of bone marrow and peripheral blood is a sensitive and simple way to detect and quantify breast cancer cells. one of the main reasons for metastatic relapse in patients with solid tumours is the early dissemination of malignant cells. the use of monoclonal antibodies ( mabs ) specific for cytokeratins can identify disseminated individual epithelial tumor cells in the bone marrow. one study reports on having developed an immuno - cytochemical procedure for simultaneous labeling of cytokeratin component no. 18 ( ck18 ) and prostate specific antigen ( psa ). this would help in the further characterization of disseminated individual epithelial tumor cells in patients with prostate cancer. the twelve control aspirates from patients with benign prostatic hyperplasia showed negative staining, which further supports the specificity of ck18 in detecting epithelial tumour cells in bone marrow. in most cases of malignant disease complicated by effusion, neoplastic cells can be easily recognized. however, in some cases, malignant cells are not so easily seen or their presence is too doubtful to call it a positive report. the use of immuno - cytochemical techniques increases diagnostic accuracy in these cases. ghosh, mason and spriggs analysed 53 samples of pleural or peritoneal fluid from 41 patients with malignant disease. conventional cytological examination had not revealed any neoplastic cells. three monocl
the galactic microquasar ss 433 is a member of a binary system but there is a lack of data on the orbital velocities of the components. the emission lines of the c ii doublet at 7231 and 7236 angstrom have been tracked nightly over two orbital cycles. the spectra are adequate to establish that these lines are eclipsed by the companion and hence to extract a measure of the orbital velocity of the compact object ; the lines are formed in the disk photosphere. this velocity is 176 plus / minus 13 km / s. could xshooter do better?
histochemical staining of bone marrow aspirates are very useful to detect occult bone marrow metastases in patients with apparently localized prostate cancer. although immuno - cytochemistry using tumor - associated monoclonal antibodies has led to an improved ability to detect occult breast cancer cells in bone marrow aspirates and peripheral blood, further development of this method is necessary before it can be used routinely. one major drawback of immuno - cytochemistry is that only tumor - associated and not tumor - specific monoclonal antibodies are used, and as a result, some cross - reaction with normal cells can occur. in order to effectively stage breast cancer and assess the efficacy of purging regimens prior to autologous stem cell infusion, it is important to detect even small quantities of breast cancer cells. immuno - histochemical methods are ideal for this purpose because they are simple, sensitive, and quite specific. franklin et al. performed a sensitive immuno - cytochemical assay by using a combination of four monoclonal antibodies ( 260f9, 520c9, 317g5 and bre - 3 ) against tumor cell surface glycoproteins to identify breast tumour cells in bone marrow and peripheral blood. they concluded from the results that immuno - cytochemical staining of bone marrow and peripheral blood is a sensitive and simple way to detect and quantify breast cancer cells. one of the main reasons for metastatic relapse in patients with solid tumours is the early dissemination of malignant cells. the use of monoclonal antibodies ( mabs ) specific for cytokeratins can identify disseminated individual epithelial tumor cells in the bone marrow. one study reports on having developed an immuno - cytochemical procedure for simultaneous labeling of cytokeratin component no. 18 ( ck18 ) and prostate specific antigen ( psa ). this would help in the further characterization of disseminated individual epithelial tumor cells in patients with prostate cancer. the twelve control aspirates from patients with benign prostatic hyperplasia showed negative staining, which further supports the specificity of ck18 in detecting epithelial tumour cells in bone marrow. in most cases of malignant disease complicated by effusion, neoplastic cells can be easily recognized. however, in some cases, malignant cells are not so easily seen or their presence is too doubtful to call it a positive report. the use
. in order to effectively stage breast cancer and assess the efficacy of purging regimens prior to autologous stem cell infusion, it is important to detect even small quantities of breast cancer cells. immuno - histochemical methods are ideal for this purpose because they are simple, sensitive, and quite specific. franklin et al. performed a sensitive immuno - cytochemical assay by using a combination of four monoclonal antibodies ( 260f9, 520c9, 317g5 and bre - 3 ) against tumor cell surface glycoproteins to identify breast tumour cells in bone marrow and peripheral blood. they concluded from the results that immuno - cytochemical staining of bone marrow and peripheral blood is a sensitive and simple way to detect and quantify breast cancer cells. one of the main reasons for metastatic relapse in patients with solid tumours is the early dissemination of malignant cells. the use of monoclonal antibodies ( mabs ) specific for cytokeratins can identify disseminated individual epithelial tumor cells in the bone marrow. one study reports on having developed an immuno - cytochemical procedure for simultaneous labeling of cytokeratin component no. 18 ( ck18 ) and prostate specific antigen ( psa ). this would help in the further characterization of disseminated individual epithelial tumor cells in patients with prostate cancer. the twelve control aspirates from patients with benign prostatic hyperplasia showed negative staining, which further supports the specificity of ck18 in detecting epithelial tumour cells in bone marrow. in most cases of malignant disease complicated by effusion, neoplastic cells can be easily recognized. however, in some cases, malignant cells are not so easily seen or their presence is too doubtful to call it a positive report. the use of immuno - cytochemical techniques increases diagnostic accuracy in these cases. ghosh, mason and spriggs analysed 53 samples of pleural or peritoneal fluid from 41 patients with malignant disease. conventional cytological examination had not revealed any neoplastic cells. three monoclonal antibodies ( anti - cea, ca 1 and hmfg - 2 ) were used to search for malignant cells. immunocytochemical labelling was performed on unstained smears, which had been stored at - 20 Β°c up to 18 months. twelve of the forty - one cases
Question: Tony is studying ladybugs in a garden. Which of these would BEST help Tony to count the spots on a ladybug?
A) Stopwatch
B) Thermometer
C) Measuring cup
D) Magnifying glass
|
D) Magnifying glass
|
Context:
intense research in the materials science community due to the unique properties that they exhibit. nanostructure deals with objects and structures that are in the 1 β 100 nm range. in many materials, atoms or molecules agglomerate to form objects at the nanoscale. this causes many interesting electrical, magnetic, optical, and mechanical properties. in describing nanostructures, it is necessary to differentiate between the number of dimensions on the nanoscale. nanotextured surfaces have one dimension on the nanoscale, i. e., only the thickness of the surface of an object is between 0. 1 and 100 nm. nanotubes have two dimensions on the nanoscale, i. e., the diameter of the tube is between 0. 1 and 100 nm ; its length could be much greater. finally, spherical nanoparticles have three dimensions on the nanoscale, i. e., the particle is between 0. 1 and 100 nm in each spatial dimension. the terms nanoparticles and ultrafine particles ( ufp ) often are used synonymously although ufp can reach into the micrometre range. the term ' nanostructure ' is often used, when referring to magnetic technology. nanoscale structure in biology is often called ultrastructure. = = = = microstructure = = = = microstructure is defined as the structure of a prepared surface or thin foil of material as revealed by a microscope above 25Γ magnification. it deals with objects from 100 nm to a few cm. the microstructure of a material ( which can be broadly classified into metallic, polymeric, ceramic and composite ) can strongly influence physical properties such as strength, toughness, ductility, hardness, corrosion resistance, high / low temperature behavior, wear resistance, and so on. most of the traditional materials ( such as metals and ceramics ) are microstructured. the manufacture of a perfect crystal of a material is physically impossible. for example, any crystalline material will contain defects such as precipitates, grain boundaries ( hall β petch relationship ), vacancies, interstitial atoms or substitutional atoms. the microstructure of materials reveals these larger defects and advances in simulation have allowed an increased understanding of how defects can be used to enhance material properties. = = = = macrostructure = = = = macrostructure is the appearance of a material in the scale millimeters to meters, it is the structure of
no quantitative theory describing all physical phenomena can be made if any arbitrary standard spacetime structure is assumed. this statement is a consequence of transforming the peano arithmetic axioms into sentences with a physical content.
the motion of celestial bodies through a higher power such as god. aristotle did not have the technological advancements that would have explained the motion of celestial bodies. in addition, aristotle had many views on the elements. he believed that everything was derived of the elements earth, water, air, fire, and lastly the aether. the aether was a celestial element, and therefore made up the matter of the celestial bodies. the elements of earth, water, air and fire were derived of a combination of two of the characteristics of hot, wet, cold, and dry, and all had their inevitable place and motion. the motion of these elements begins with earth being the closest to " the earth, " then water, air, fire, and finally aether. in addition to the makeup of all things, aristotle came up with theories as to why things did not return to their natural motion. he understood that water sits above earth, air above water, and fire above air in their natural state. he explained that although all elements must return to their natural state, the human body and other living things have a constraint on the elements β thus not allowing the elements making one who they are to return to their natural state. the important legacy of this period included substantial advances in factual knowledge, especially in anatomy, zoology, botany, mineralogy, geography, mathematics and astronomy ; an awareness of the importance of certain scientific problems, especially those related to the problem of change and its causes ; and a recognition of the methodological importance of applying mathematics to natural phenomena and of undertaking empirical research. in the hellenistic age scholars frequently employed the principles developed in earlier greek thought : the application of mathematics and deliberate empirical research, in their scientific investigations. thus, clear unbroken lines of influence lead from ancient greek and hellenistic philosophers, to medieval muslim philosophers and scientists, to the european renaissance and enlightenment, to the secular sciences of the modern day. neither reason nor inquiry began with the ancient greeks, but the socratic method did, along with the idea of forms, give great advances in geometry, logic, and the natural sciences. according to benjamin farrington, former professor of classics at swansea university : " men were weighing for thousands of years before archimedes worked out the laws of equilibrium ; they must have had practical and intuitional knowledge of the principals involved. what archimedes did was to sort out the theoretical implications of this practical knowledge and present the resulting body of knowledge as a logically coherent system. " and again : " with astonishment we find ourselves on the threshold of modern science
0. 1 and 100 nm in each spatial dimension. the terms nanoparticles and ultrafine particles ( ufp ) often are used synonymously although ufp can reach into the micrometre range. the term ' nanostructure ' is often used, when referring to magnetic technology. nanoscale structure in biology is often called ultrastructure. = = = = microstructure = = = = microstructure is defined as the structure of a prepared surface or thin foil of material as revealed by a microscope above 25Γ magnification. it deals with objects from 100 nm to a few cm. the microstructure of a material ( which can be broadly classified into metallic, polymeric, ceramic and composite ) can strongly influence physical properties such as strength, toughness, ductility, hardness, corrosion resistance, high / low temperature behavior, wear resistance, and so on. most of the traditional materials ( such as metals and ceramics ) are microstructured. the manufacture of a perfect crystal of a material is physically impossible. for example, any crystalline material will contain defects such as precipitates, grain boundaries ( hall β petch relationship ), vacancies, interstitial atoms or substitutional atoms. the microstructure of materials reveals these larger defects and advances in simulation have allowed an increased understanding of how defects can be used to enhance material properties. = = = = macrostructure = = = = macrostructure is the appearance of a material in the scale millimeters to meters, it is the structure of the material as seen with the naked eye. = = = properties = = = materials exhibit myriad properties, including the following. mechanical properties, see strength of materials chemical properties, see chemistry electrical properties, see electricity thermal properties, see thermodynamics optical properties, see optics and photonics magnetic properties, see magnetism the properties of a material determine its usability and hence its engineering application. = = = processing = = = synthesis and processing involves the creation of a material with the desired micro - nanostructure. a material cannot be used in industry if no economically viable production method for it has been developed. therefore, developing processing methods for materials that are reasonably effective and cost - efficient is vital to the field of materials science. different materials require different processing or synthesis methods. for example, the processing of metals has historically defined eras such as the bronze age and iron age and is studied under the branch of materials science named physical metallurgy.
##odynamic and mechanical descriptions of physical properties. = = = = nanostructure = = = = materials, which atoms and molecules form constituents in the nanoscale ( i. e., they form nanostructures ) are called nanomaterials. nanomaterials are the subject of intense research in the materials science community due to the unique properties that they exhibit. nanostructure deals with objects and structures that are in the 1 β 100 nm range. in many materials, atoms or molecules agglomerate to form objects at the nanoscale. this causes many interesting electrical, magnetic, optical, and mechanical properties. in describing nanostructures, it is necessary to differentiate between the number of dimensions on the nanoscale. nanotextured surfaces have one dimension on the nanoscale, i. e., only the thickness of the surface of an object is between 0. 1 and 100 nm. nanotubes have two dimensions on the nanoscale, i. e., the diameter of the tube is between 0. 1 and 100 nm ; its length could be much greater. finally, spherical nanoparticles have three dimensions on the nanoscale, i. e., the particle is between 0. 1 and 100 nm in each spatial dimension. the terms nanoparticles and ultrafine particles ( ufp ) often are used synonymously although ufp can reach into the micrometre range. the term ' nanostructure ' is often used, when referring to magnetic technology. nanoscale structure in biology is often called ultrastructure. = = = = microstructure = = = = microstructure is defined as the structure of a prepared surface or thin foil of material as revealed by a microscope above 25Γ magnification. it deals with objects from 100 nm to a few cm. the microstructure of a material ( which can be broadly classified into metallic, polymeric, ceramic and composite ) can strongly influence physical properties such as strength, toughness, ductility, hardness, corrosion resistance, high / low temperature behavior, wear resistance, and so on. most of the traditional materials ( such as metals and ceramics ) are microstructured. the manufacture of a perfect crystal of a material is physically impossible. for example, any crystalline material will contain defects such as precipitates, grain boundaries ( hall β petch relationship ), vacancies, interstitial atoms or substitutional atoms. the micro
ultramagnetized neutron stars or magnetars are magnetically powered neutron stars. their strong magnetic fields dominate the physical processes in their crusts and their surroundings. the past few years have seen several advances in our theoretical and observational understanding of these objects. in spite of a surfeit of observations, their spectra are still poorly understood. i will discuss the emission from strongly magnetized condensed matter surfaces of neutron stars, recent advances in our expectations of the surface composition of magnetars and a model for the non - thermal emission from these objects.
we have written a java applet to illustrate the meaning of curved geometry. the applet provides a mapping interface similar to mapquest or google maps ; features include the ability to navigate through a space and place permanent point objects and / or shapes at arbitrary positions. the underlying two - dimensional space has a constant, positive curvature, which causes the apparent paths and shapes of the objects in the map to appear distorted in ways that change as you view them from different relative angles and distances.
joints. = = = metal alloys = = = the alloys of iron ( steel, stainless steel, cast iron, tool steel, alloy steels ) make up the largest proportion of metals today both by quantity and commercial value. iron alloyed with various proportions of carbon gives low, mid and high carbon steels. an iron - carbon alloy is only considered steel if the carbon level is between 0. 01 % and 2. 00 % by weight. for steels, the hardness and tensile strength of the steel is related to the amount of carbon present, with increasing carbon levels also leading to lower ductility and toughness. heat treatment processes such as quenching and tempering can significantly change these properties, however. in contrast, certain metal alloys exhibit unique properties where their size and density remain unchanged across a range of temperatures. cast iron is defined as an iron β carbon alloy with more than 2. 00 %, but less than 6. 67 % carbon. stainless steel is defined as a regular steel alloy with greater than 10 % by weight alloying content of chromium. nickel and molybdenum are typically also added in stainless steels. other significant metallic alloys are those of aluminium, titanium, copper and magnesium. copper alloys have been known for a long time ( since the bronze age ), while the alloys of the other three metals have been relatively recently developed. due to the chemical reactivity of these metals, the electrolytic extraction processes required were only developed relatively recently. the alloys of aluminium, titanium and magnesium are also known and valued for their high strength to weight ratios and, in the case of magnesium, their ability to provide electromagnetic shielding. these materials are ideal for situations where high strength to weight ratios are more important than bulk cost, such as in the aerospace industry and certain automotive engineering applications. = = = semiconductors = = = a semiconductor is a material that has a resistivity between a conductor and insulator. modern day electronics run on semiconductors, and the industry had an estimated us $ 530 billion market in 2021. its electronic properties can be greatly altered through intentionally introducing impurities in a process referred to as doping. semiconductor materials are used to build diodes, transistors, light - emitting diodes ( leds ), and analog and digital electric circuits, among their many uses. semiconductor devices have replaced thermionic devices like vacuum tubes in most applications. semiconductor devices are manufactured both as single discrete devices and as integrated circuits ( ics ), which consist of a number β from a
some properties of the nuclear matter as revealed by cherenkov gluons are discussed.
new crop traits as well as a far greater control over a food ' s genetic structure than previously afforded by methods such as selective breeding and mutation breeding. commercial sale of genetically modified foods began in 1994, when calgene first marketed its flavr savr delayed ripening tomato. to date most genetic modification of foods have primarily focused on cash crops in high demand by farmers such as soybean, corn, canola, and cotton seed oil. these have been engineered for resistance to pathogens and herbicides and better nutrient profiles. gm livestock have also been experimentally developed ; in november 2013 none were available on the market, but in 2015 the fda approved the first gm salmon for commercial production and consumption. there is a scientific consensus that currently available food derived from gm crops poses no greater risk to human health than conventional food, but that each gm food needs to be tested on a case - by - case basis before introduction. nonetheless, members of the public are much less likely than scientists to perceive gm foods as safe. the legal and regulatory status of gm foods varies by country, with some nations banning or restricting them, and others permitting them with widely differing degrees of regulation. gm crops also provide a number of ecological benefits, if not used in excess. insect - resistant crops have proven to lower pesticide usage, therefore reducing the environmental impact of pesticides as a whole. however, opponents have objected to gm crops per se on several grounds, including environmental concerns, whether food produced from gm crops is safe, whether gm crops are needed to address the world ' s food needs, and economic concerns raised by the fact these organisms are subject to intellectual property law. biotechnology has several applications in the realm of food security. crops like golden rice are engineered to have higher nutritional content, and there is potential for food products with longer shelf lives. though not a form of agricultural biotechnology, vaccines can help prevent diseases found in animal agriculture. additionally, agricultural biotechnology can expedite breeding processes in order to yield faster results and provide greater quantities of food. transgenic biofortification in cereals has been considered as a promising method to combat malnutrition in india and other countries. = = = industrial = = = industrial biotechnology ( known mainly in europe as white biotechnology ) is the application of biotechnology for industrial purposes, including industrial fermentation. it includes the practice of using cells such as microorganisms, or components of cells like enzymes, to generate industrially useful products in sectors such as chemicals, food and feed, detergents, paper
Question: Which is a physical property of an apple?
A) what color it is
B) how pretty it is
C) how much it costs
D) when it was picked
|
A) what color it is
|
Context:
on a large scale provided protection from insect pests or tolerance to herbicides. fungal and virus resistant crops have also been developed or are in development. this makes the insect and weed management of crops easier and can indirectly increase crop yield. gm crops that directly improve yield by accelerating growth or making the plant more hardy ( by improving salt, cold or drought tolerance ) are also under development. in 2016 salmon have been genetically modified with growth hormones to reach normal adult size much faster. gmos have been developed that modify the quality of produce by increasing the nutritional value or providing more industrially useful qualities or quantities. the amflora potato produces a more industrially useful blend of starches. soybeans and canola have been genetically modified to produce more healthy oils. the first commercialised gm food was a tomato that had delayed ripening, increasing its shelf life. plants and animals have been engineered to produce materials they do not normally make. pharming uses crops and animals as bioreactors to produce vaccines, drug intermediates, or the drugs themselves ; the useful product is purified from the harvest and then used in the standard pharmaceutical production process. cows and goats have been engineered to express drugs and other proteins in their milk, and in 2009 the fda approved a drug produced in goat milk. = = = other applications = = = genetic engineering has potential applications in conservation and natural area management. gene transfer through viral vectors has been proposed as a means of controlling invasive species as well as vaccinating threatened fauna from disease. transgenic trees have been suggested as a way to confer resistance to pathogens in wild populations. with the increasing risks of maladaptation in organisms as a result of climate change and other perturbations, facilitated adaptation through gene tweaking could be one solution to reducing extinction risks. applications of genetic engineering in conservation are thus far mostly theoretical and have yet to be put into practice. genetic engineering is also being used to create microbial art. some bacteria have been genetically engineered to create black and white photographs. novelty items such as lavender - colored carnations, blue roses, and glowing fish, have also been produced through genetic engineering. = = regulation = = the regulation of genetic engineering concerns the approaches taken by governments to assess and manage the risks associated with the development and release of gmos. the development of a regulatory framework began in 1975, at asilomar, california. the asilomar meeting recommended a set of voluntary guidelines regarding the use of recombinant technology. as the technology improved
best - known and controversial applications of genetic engineering is the creation and use of genetically modified crops or genetically modified livestock to produce genetically modified food. crops have been developed to increase production, increase tolerance to abiotic stresses, alter the composition of the food, or to produce novel products. the first crops to be released commercially on a large scale provided protection from insect pests or tolerance to herbicides. fungal and virus resistant crops have also been developed or are in development. this makes the insect and weed management of crops easier and can indirectly increase crop yield. gm crops that directly improve yield by accelerating growth or making the plant more hardy ( by improving salt, cold or drought tolerance ) are also under development. in 2016 salmon have been genetically modified with growth hormones to reach normal adult size much faster. gmos have been developed that modify the quality of produce by increasing the nutritional value or providing more industrially useful qualities or quantities. the amflora potato produces a more industrially useful blend of starches. soybeans and canola have been genetically modified to produce more healthy oils. the first commercialised gm food was a tomato that had delayed ripening, increasing its shelf life. plants and animals have been engineered to produce materials they do not normally make. pharming uses crops and animals as bioreactors to produce vaccines, drug intermediates, or the drugs themselves ; the useful product is purified from the harvest and then used in the standard pharmaceutical production process. cows and goats have been engineered to express drugs and other proteins in their milk, and in 2009 the fda approved a drug produced in goat milk. = = = other applications = = = genetic engineering has potential applications in conservation and natural area management. gene transfer through viral vectors has been proposed as a means of controlling invasive species as well as vaccinating threatened fauna from disease. transgenic trees have been suggested as a way to confer resistance to pathogens in wild populations. with the increasing risks of maladaptation in organisms as a result of climate change and other perturbations, facilitated adaptation through gene tweaking could be one solution to reducing extinction risks. applications of genetic engineering in conservation are thus far mostly theoretical and have yet to be put into practice. genetic engineering is also being used to create microbial art. some bacteria have been genetically engineered to create black and white photographs. novelty items such as lavender - colored carnations, blue roses, and glowing fish, have also been produced through genetic engineering. = = regulation = = the regulation of genetic engineering
life, but most current gm crops are modified to increase resistance to insects and herbicides. glofish, the first gmo designed as a pet, was sold in the united states in december 2003. in 2016 salmon modified with a growth hormone were sold. genetic engineering has been applied in numerous fields including research, medicine, industrial biotechnology and agriculture. in research, gmos are used to study gene function and expression through loss of function, gain of function, tracking and expression experiments. by knocking out genes responsible for certain conditions it is possible to create animal model organisms of human diseases. as well as producing hormones, vaccines and other drugs, genetic engineering has the potential to cure genetic diseases through gene therapy. chinese hamster ovary ( cho ) cells are used in industrial genetic engineering. additionally mrna vaccines are made through genetic engineering to prevent infections by viruses such as covid - 19. the same techniques that are used to produce drugs can also have industrial applications such as producing enzymes for laundry detergent, cheeses and other products. the rise of commercialised genetically modified crops has provided economic benefit to farmers in many different countries, but has also been the source of most of the controversy surrounding the technology. this has been present since its early use ; the first field trials were destroyed by anti - gm activists. although there is a scientific consensus that currently available food derived from gm crops poses no greater risk to human health than conventional food, critics consider gm food safety a leading concern. gene flow, impact on non - target organisms, control of the food supply and intellectual property rights have also been raised as potential issues. these concerns have led to the development of a regulatory framework, which started in 1975. it has led to an international treaty, the cartagena protocol on biosafety, that was adopted in 2000. individual countries have developed their own regulatory systems regarding gmos, with the most marked differences occurring between the united states and europe. = = overview = = genetic engineering is a process that alters the genetic structure of an organism by either removing or introducing dna, or modifying existing genetic material in situ. unlike traditional animal and plant breeding, which involves doing multiple crosses and then selecting for the organism with the desired phenotype, genetic engineering takes the gene directly from one organism and delivers it to the other. this is much faster, can be used to insert any genes from any organism ( even ones from different domains ) and prevents other undesirable genes from also being added. genetic engineering could potentially fix severe genetic disorders in humans by replacing the
no offspring, to reduce the population. in industrial and food applications, radiation is used for sterilization of tools and equipment. an advantage is that the object may be sealed in plastic before sterilization. an emerging use in food production is the sterilization of food using food irradiation. food irradiation is the process of exposing food to ionizing radiation in order to destroy microorganisms, bacteria, viruses, or insects that might be present in the food. the radiation sources used include radioisotope gamma ray sources, x - ray generators and electron accelerators. further applications include sprout inhibition, delay of ripening, increase of juice yield, and improvement of re - hydration. irradiation is a more general term of deliberate exposure of materials to radiation to achieve a technical goal ( in this context ' ionizing radiation ' is implied ). as such it is also used on non - food items, such as medical hardware, plastics, tubes for gas - pipelines, hoses for floor - heating, shrink - foils for food packaging, automobile parts, wires and cables ( isolation ), tires, and even gemstones. compared to the amount of food irradiated, the volume of those every - day applications is huge but not noticed by the consumer. the genuine effect of processing food by ionizing radiation relates to damages to the dna, the basic genetic information for life. microorganisms can no longer proliferate and continue their malignant or pathogenic activities. spoilage causing micro - organisms cannot continue their activities. insects do not survive or become incapable of procreation. plants cannot continue the natural ripening or aging process. all these effects are beneficial to the consumer and the food industry, likewise. the amount of energy imparted for effective food irradiation is low compared to cooking the same ; even at a typical dose of 10 kgy most food, which is ( with regard to warming ) physically equivalent to water, would warm by only about 2. 5 Β°c ( 4. 5 Β°f ). the specialty of processing food by ionizing radiation is the fact, that the energy density per atomic transition is very high, it can cleave molecules and induce ionization ( hence the name ) which cannot be achieved by mere heating. this is the reason for new beneficial effects, however at the same time, for new concerns. the treatment of solid food by ionizing radiation can provide an effect similar to heat pasteurization of liquids, such as milk. however
industrial applications. this branch of biotechnology is the most used for the industries of refining and combustion principally on the production of bio - oils with photosynthetic micro - algae. green biotechnology is biotechnology applied to agricultural processes. an example would be the selection and domestication of plants via micropropagation. another example is the designing of transgenic plants to grow under specific environments in the presence ( or absence ) of chemicals. one hope is that green biotechnology might produce more environmentally friendly solutions than traditional industrial agriculture. an example of this is the engineering of a plant to express a pesticide, thereby ending the need of external application of pesticides. an example of this would be bt corn. whether or not green biotechnology products such as this are ultimately more environmentally friendly is a topic of considerable debate. it is commonly considered as the next phase of green revolution, which can be seen as a platform to eradicate world hunger by using technologies which enable the production of more fertile and resistant, towards biotic and abiotic stress, plants and ensures application of environmentally friendly fertilizers and the use of biopesticides, it is mainly focused on the development of agriculture. on the other hand, some of the uses of green biotechnology involve microorganisms to clean and reduce waste. red biotechnology is the use of biotechnology in the medical and pharmaceutical industries, and health preservation. this branch involves the production of vaccines and antibiotics, regenerative therapies, creation of artificial organs and new diagnostics of diseases. as well as the development of hormones, stem cells, antibodies, sirna and diagnostic tests. white biotechnology, also known as industrial biotechnology, is biotechnology applied to industrial processes. an example is the designing of an organism to produce a useful chemical. another example is the using of enzymes as industrial catalysts to either produce valuable chemicals or destroy hazardous / polluting chemicals. white biotechnology tends to consume less in resources than traditional processes used to produce industrial goods. yellow biotechnology refers to the use of biotechnology in food production ( food industry ), for example in making wine ( winemaking ), cheese ( cheesemaking ), and beer ( brewing ) by fermentation. it has also been used to refer to biotechnology applied to insects. this includes biotechnology - based approaches for the control of harmful insects, the characterisation and utilisation of active ingredients or genes of insects for research, or application in agriculture and medicine and various other approaches. gray biotechnology is dedicated to environmental applications, and focused on the maintenance of biodiversity and the remotion of poll
in 2015 the fda approved the first gm salmon for commercial production and consumption. there is a scientific consensus that currently available food derived from gm crops poses no greater risk to human health than conventional food, but that each gm food needs to be tested on a case - by - case basis before introduction. nonetheless, members of the public are much less likely than scientists to perceive gm foods as safe. the legal and regulatory status of gm foods varies by country, with some nations banning or restricting them, and others permitting them with widely differing degrees of regulation. gm crops also provide a number of ecological benefits, if not used in excess. insect - resistant crops have proven to lower pesticide usage, therefore reducing the environmental impact of pesticides as a whole. however, opponents have objected to gm crops per se on several grounds, including environmental concerns, whether food produced from gm crops is safe, whether gm crops are needed to address the world ' s food needs, and economic concerns raised by the fact these organisms are subject to intellectual property law. biotechnology has several applications in the realm of food security. crops like golden rice are engineered to have higher nutritional content, and there is potential for food products with longer shelf lives. though not a form of agricultural biotechnology, vaccines can help prevent diseases found in animal agriculture. additionally, agricultural biotechnology can expedite breeding processes in order to yield faster results and provide greater quantities of food. transgenic biofortification in cereals has been considered as a promising method to combat malnutrition in india and other countries. = = = industrial = = = industrial biotechnology ( known mainly in europe as white biotechnology ) is the application of biotechnology for industrial purposes, including industrial fermentation. it includes the practice of using cells such as microorganisms, or components of cells like enzymes, to generate industrially useful products in sectors such as chemicals, food and feed, detergents, paper and pulp, textiles and biofuels. in the current decades, significant progress has been done in creating genetically modified organisms ( gmos ) that enhance the diversity of applications and economical viability of industrial biotechnology. by using renewable raw materials to produce a variety of chemicals and fuels, industrial biotechnology is actively advancing towards lowering greenhouse gas emissions and moving away from a petrochemical - based economy. synthetic biology is considered one of the essential cornerstones in industrial biotechnology due to its financial and sustainable contribution to the manufacturing sector. jointly biotechnology and synthetic biology play a crucial role in generating cost - effective products with nature - friendly features by using bio - based
the designing of transgenic plants to grow under specific environments in the presence ( or absence ) of chemicals. one hope is that green biotechnology might produce more environmentally friendly solutions than traditional industrial agriculture. an example of this is the engineering of a plant to express a pesticide, thereby ending the need of external application of pesticides. an example of this would be bt corn. whether or not green biotechnology products such as this are ultimately more environmentally friendly is a topic of considerable debate. it is commonly considered as the next phase of green revolution, which can be seen as a platform to eradicate world hunger by using technologies which enable the production of more fertile and resistant, towards biotic and abiotic stress, plants and ensures application of environmentally friendly fertilizers and the use of biopesticides, it is mainly focused on the development of agriculture. on the other hand, some of the uses of green biotechnology involve microorganisms to clean and reduce waste. red biotechnology is the use of biotechnology in the medical and pharmaceutical industries, and health preservation. this branch involves the production of vaccines and antibiotics, regenerative therapies, creation of artificial organs and new diagnostics of diseases. as well as the development of hormones, stem cells, antibodies, sirna and diagnostic tests. white biotechnology, also known as industrial biotechnology, is biotechnology applied to industrial processes. an example is the designing of an organism to produce a useful chemical. another example is the using of enzymes as industrial catalysts to either produce valuable chemicals or destroy hazardous / polluting chemicals. white biotechnology tends to consume less in resources than traditional processes used to produce industrial goods. yellow biotechnology refers to the use of biotechnology in food production ( food industry ), for example in making wine ( winemaking ), cheese ( cheesemaking ), and beer ( brewing ) by fermentation. it has also been used to refer to biotechnology applied to insects. this includes biotechnology - based approaches for the control of harmful insects, the characterisation and utilisation of active ingredients or genes of insects for research, or application in agriculture and medicine and various other approaches. gray biotechnology is dedicated to environmental applications, and focused on the maintenance of biodiversity and the remotion of pollutants. brown biotechnology is related to the management of arid lands and deserts. one application is the creation of enhanced seeds that resist extreme environmental conditions of arid regions, which is related to the innovation, creation of agriculture techniques and management of resources. violet biotechnology is related to law, ethical and philosophical issues around biotechnology. micro
tubers in potato is one example. particularly in arctic or alpine habitats, where opportunities for fertilisation of flowers by animals are rare, plantlets or bulbs, may develop instead of flowers, replacing sexual reproduction with asexual reproduction and giving rise to clonal populations genetically identical to the parent. this is one of several types of apomixis that occur in plants. apomixis can also happen in a seed, producing a seed that contains an embryo genetically identical to the parent. most sexually reproducing organisms are diploid, with paired chromosomes, but doubling of their chromosome number may occur due to errors in cytokinesis. this can occur early in development to produce an autopolyploid or partly autopolyploid organism, or during normal processes of cellular differentiation to produce some cell types that are polyploid ( endopolyploidy ), or during gamete formation. an allopolyploid plant may result from a hybridisation event between two different species. both autopolyploid and allopolyploid plants can often reproduce normally, but may be unable to cross - breed successfully with the parent population because there is a mismatch in chromosome numbers. these plants that are reproductively isolated from the parent species but live within the same geographical area, may be sufficiently successful to form a new species. some otherwise sterile plant polyploids can still reproduce vegetatively or by seed apomixis, forming clonal populations of identical individuals. durum wheat is a fertile tetraploid allopolyploid, while bread wheat is a fertile hexaploid. the commercial banana is an example of a sterile, seedless triploid hybrid. common dandelion is a triploid that produces viable seeds by apomictic seed. as in other eukaryotes, the inheritance of endosymbiotic organelles like mitochondria and chloroplasts in plants is non - mendelian. chloroplasts are inherited through the male parent in gymnosperms but often through the female parent in flowering plants. = = = molecular genetics = = = a considerable amount of new knowledge about plant function comes from studies of the molecular genetics of model plants such as the thale cress, arabidopsis thaliana, a weedy species in the mustard family ( brassicaceae ). the genome or hereditary information contained in the genes of this species is encoded by about 135 million base pairs of dna, forming one of the
herbicides. the people ' s republic of china was the first country to commercialise transgenic plants, introducing a virus - resistant tobacco in 1992. in 1994 calgene attained approval to commercially release the first genetically modified food, the flavr savr, a tomato engineered to have a longer shelf life. in 1994, the european union approved tobacco engineered to be resistant to the herbicide bromoxynil, making it the first genetically engineered crop commercialised in europe. in 1995, bt potato was approved safe by the environmental protection agency, after having been approved by the fda, making it the first pesticide producing crop to be approved in the us. in 2009 11 transgenic crops were grown commercially in 25 countries, the largest of which by area grown were the us, brazil, argentina, india, canada, china, paraguay and south africa. in 2010, scientists at the j. craig venter institute created the first synthetic genome and inserted it into an empty bacterial cell. the resulting bacterium, named mycoplasma laboratorium, could replicate and produce proteins. four years later this was taken a step further when a bacterium was developed that replicated a plasmid containing a unique base pair, creating the first organism engineered to use an expanded genetic alphabet. in 2012, jennifer doudna and emmanuelle charpentier collaborated to develop the crispr / cas9 system, a technique which can be used to easily and specifically alter the genome of almost any organism. = = process = = creating a gmo is a multi - step process. genetic engineers must first choose what gene they wish to insert into the organism. this is driven by what the aim is for the resultant organism and is built on earlier research. genetic screens can be carried out to determine potential genes and further tests then used to identify the best candidates. the development of microarrays, transcriptomics and genome sequencing has made it much easier to find suitable genes. luck also plays its part ; the roundup ready gene was discovered after scientists noticed a bacterium thriving in the presence of the herbicide. = = = gene isolation and cloning = = = the next step is to isolate the candidate gene. the cell containing the gene is opened and the dna is purified. the gene is separated by using restriction enzymes to cut the dna into fragments or polymerase chain reaction ( pcr ) to amplify up the gene segment. these segments can then be extracted through gel electrophoresis. if the chosen gene or the donor organism ' s
= = = = = = environmental remediation = = = environmental remediation is the process through which contaminants or pollutants in soil, water and other media are removed to improve environmental quality. the main focus is the reduction of hazardous substances within the environment. some of the areas involved in environmental remediation include ; soil contamination, hazardous waste, groundwater contamination, oil, gas and chemical spills. there are three most common types of environmental remediation. these include soil, water, and sediment remediation. soil remediation consists of removing contaminants in soil, as these pose great risks to humans and the ecosystem. some examples of this are heavy metals, pesticides, and radioactive materials. depending on the contaminant the remedial processes can be physical, chemical, thermal, or biological. water remediation is one of the most important considering water is an essential natural resource. depending on the source of water there will be different contaminants. surface water contamination mainly consists of agricultural, animal, and industrial waste, as well as acid mine drainage. there has been a rise in the need for water remediation due to the increased discharge of industrial waste, leading to a demand for sustainable water solutions. the market for water remediation is expected to consistently increase to $ 19. 6 billion by 2030. sediment remediation consists of removing contaminated sediments. is it almost similar to soil remediation except it is often more sophisticated as it involves additional contaminants. to reduce the contaminants it is likely to use physical, chemical, and biological processes that help with source control, but if these processes are executed correctly, there ' s a risk of contamination resurfacing. = = = solid waste management = = = solid waste management is the purification, consumption, reuse, disposal, and treatment of solid waste that is undertaken by the government or the ruling bodies of a city / town. it refers to the collection, treatment, and disposal of non - soluble, solid waste material. solid waste is associated with both industrial, institutional, commercial and residential activities. hazardous solid waste, when improperly disposed can encourage the infestation of insects and rodents, contributing to the spread of diseases. some of the most common types of solid waste management include ; landfills, vermicomposting, composting, recycling, and incineration. however, a major barrier for solid waste management practices is the high costs associated with recycling
Question: Repeated pesticide use on an insect population can result in some of the insects becoming resistant to that pesticide over time. Which process is responsible for the pesticides no longer being effective on the insects?
A) divergent evolution
B) natural selection
C) bottleneck effect
D) selective breeding
|
B) natural selection
|
Context:
the transition of our energy system to renewable energies is necessary in order not to heat up the climate any further and to achieve climate neutrality. the use of wind energy plays an important role in this transition in germany. but how much wind energy can be used and what are the possible consequences for the atmosphere if more and more wind energy is used?
proteins in the thylakoid membranes. the absorbed light energy is used to remove electrons from a donor ( water ) to a primary electron acceptor, a quinone designated as q. in the second stage, electrons move from the quinone primary electron acceptor through a series of electron carriers until they reach a final electron acceptor, which is usually the oxidized form of nadp +, which is reduced to nadph, a process that takes place in a protein complex called photosystem i ( psi ). the transport of electrons is coupled to the movement of protons ( or hydrogen ) from the stroma to the thylakoid membrane, which forms a ph gradient across the membrane as hydrogen becomes more concentrated in the lumen than in the stroma. this is analogous to the proton - motive force generated across the inner mitochondrial membrane in aerobic respiration. during the third stage of photosynthesis, the movement of protons down their concentration gradients from the thylakoid lumen to the stroma through the atp synthase is coupled to the synthesis of atp by that same atp synthase. the nadph and atps generated by the light - dependent reactions in the second and third stages, respectively, provide the energy and electrons to drive the synthesis of glucose by fixing atmospheric carbon dioxide into existing organic carbon compounds, such as ribulose bisphosphate ( rubp ) in a sequence of light - independent ( or dark ) reactions called the calvin cycle. = = = cell signaling = = = cell signaling ( or communication ) is the ability of cells to receive, process, and transmit signals with its environment and with itself. signals can be non - chemical such as light, electrical impulses, and heat, or chemical signals ( or ligands ) that interact with receptors, which can be found embedded in the cell membrane of another cell or located deep inside a cell. there are generally four types of chemical signals : autocrine, paracrine, juxtacrine, and hormones. in autocrine signaling, the ligand affects the same cell that releases it. tumor cells, for example, can reproduce uncontrollably because they release signals that initiate their own self - division. in paracrine signaling, the ligand diffuses to nearby cells and affects them. for example, brain cells called neurons release ligands called neurotransmitters that diffuse across a synaptic cleft to bind with a receptor on an adjacent cell such as another neuron or muscle
this process may release or absorb energy. when the resulting nucleus is lighter than that of iron, energy is normally released ; when the nucleus is heavier than that of iron, energy is generally absorbed. this process of fusion occurs in stars, which derive their energy from hydrogen and helium. they form, through stellar nucleosynthesis, the light elements ( lithium to calcium ) as well as some of the heavy elements ( beyond iron and nickel, via the s - process ). the remaining abundance of heavy elements, from nickel to uranium and beyond, is due to supernova nucleosynthesis, the r - process. of course, these natural processes of astrophysics are not examples of nuclear " technology ". because of the very strong repulsion of nuclei, fusion is difficult to achieve in a controlled fashion. hydrogen bombs, formally known as thermonuclear weapons, obtain their enormous destructive power from fusion, but their energy cannot be controlled. controlled fusion is achieved in particle accelerators ; this is how many synthetic elements are produced. a fusor can also produce controlled fusion and is a useful neutron source. however, both of these devices operate at a net energy loss. controlled, viable fusion power has proven elusive, despite the occasional hoax. technical and theoretical difficulties have hindered the development of working civilian fusion technology, though research continues to this day around the world. nuclear fusion was initially pursued only in theoretical stages during world war ii, when scientists on the manhattan project ( led by edward teller ) investigated it as a method to build a bomb. the project abandoned fusion after concluding that it would require a fission reaction to detonate. it took until 1952 for the first full hydrogen bomb to be detonated, so - called because it used reactions between deuterium and tritium. fusion reactions are much more energetic per unit mass of fuel than fission reactions, but starting the fusion chain reaction is much more difficult. = = nuclear weapons = = a nuclear weapon is an explosive device that derives its destructive force from nuclear reactions, either fission or a combination of fission and fusion. both reactions release vast quantities of energy from relatively small amounts of matter. even small nuclear devices can devastate a city by blast, fire and radiation. nuclear weapons are considered weapons of mass destruction, and their use and control has been a major aspect of international policy since their debut. the design of a nuclear weapon is more complicated than it might seem. such a weapon must hold one or more subcritical fissile masses stable for deployment, then induce criticality
is collected and processed to extract valuable metals. ore bodies often contain more than one valuable metal. tailings of a previous process may be used as a feed in another process to extract a secondary product from the original ore. additionally, a concentrate may contain more than one valuable metal. that concentrate would then be processed to separate the valuable metals into individual constituents. = = metal and its alloys = = much effort has been placed on understanding iron β carbon alloy system, which includes steels and cast irons. plain carbon steels ( those that contain essentially only carbon as an alloying element ) are used in low - cost, high - strength applications, where neither weight nor corrosion are a major concern. cast irons, including ductile iron, are also part of the iron - carbon system. iron - manganese - chromium alloys ( hadfield - type steels ) are also used in non - magnetic applications such as directional drilling. other engineering metals include aluminium, chromium, copper, magnesium, nickel, titanium, zinc, and silicon. these metals are most often used as alloys with the noted exception of silicon, which is not a metal. other forms include : stainless steel, particularly austenitic stainless steels, galvanized steel, nickel alloys, titanium alloys, or occasionally copper alloys are used, where resistance to corrosion is important. aluminium alloys and magnesium alloys are commonly used, when a lightweight strong part is required such as in automotive and aerospace applications. copper - nickel alloys ( such as monel ) are used in highly corrosive environments and for non - magnetic applications. nickel - based superalloys like inconel are used in high - temperature applications such as gas turbines, turbochargers, pressure vessels, and heat exchangers. for extremely high temperatures, single crystal alloys are used to minimize creep. in modern electronics, high purity single crystal silicon is essential for metal - oxide - silicon transistors ( mos ) and integrated circuits. = = production = = in production engineering, metallurgy is concerned with the production of metallic components for use in consumer or engineering products. this involves production of alloys, shaping, heat treatment and surface treatment of product. the task of the metallurgist is to achieve balance between material properties, such as cost, weight, strength, toughness, hardness, corrosion, fatigue resistance and performance in temperature extremes. to achieve this goal, the operating environment must be carefully considered. determining the hardness of the metal using the rockwell, vickers, and brinell hardness scales
if wood has been with us since time immemorial, being part of our environment, housing and tools, now wood has gain momentum, as it is clear that wood improves our life style. because of the healthiness, resistance, ecology and comfort, wood is important for all of us, no matter what our life style is. woodtouch project aims to open a completely new market for furniture and interior design companies, enabling touch interaction between the user and wooden furniture surfaces. why not switch on or dim the lights touching a wooden table? why not turn on the heating system? why not use wood as a touch sensitive surface for domotic control? the furniture designed with this novel technology, offers a wooden outer image and has different touch sensitive areas over the ones the user is able to control all sorts of electric appliances touching over a wooden surface.
nuclear jets containing relativistic ` ` hot ' ' particles close to the central engine cool dramatically by producing high energy radiation. the radiative dissipation is similar to the famous compton drag acting upon ` ` cold ' ' thermal particles in a relativistic bulk flow. highly relativistic protons induce anisotropic showers raining electromagnetic power down onto the putative accretion disk. thus, the radiative signature of hot hadronic jets is x - ray irradiation of cold thermal matter. the synchrotron radio emission of the accelerated electrons is self - absorbed due to the strong magnetic fields close to the magnetic nozzle.
in building up composite structures using carbon nanotube ( cnt ) fibers, the fiber - to - matrix interfacial shear strength ( ifss ) is one of the most important issues. originating from the assembly characteristics of cnt fiber, the ifss strongly depends on the fiber ' s twisting level and densification level. furthermore, there are rich ways to modify the fiber surface and thus enhance the ifss, including the physical and chemical modification on fiber surface, the infiltration of matrix resin into cnt fiber, and the introduction of silane coupling agent. a new feature differing from carbon fibers is that all these treatments either change the fiber ' s surface structure or form a several - hundreds - nm - thick interphase inside rather than around the fiber. these " generalized " treatments obviously extend the common concept of surface sizing and can be used for various forms of cnt assembly structures.
the gravitational poynting vector provides a mechanism for the transfer of gravitational energy to a system of falling objects. in the following we will show that the gravitational poynting vector together with the gravitational larmor theorem also provides a mechanism to explain how massive bodies acquire rotational kinetic energy when external mechanical forces are applied on them.
to dye denim and the artist ' s pigments gamboge and rose madder. sugar, starch, cotton, linen, hemp, some types of rope, wood and particle boards, papyrus and paper, vegetable oils, wax, and natural rubber are examples of commercially important materials made from plant tissues or their secondary products. charcoal, a pure form of carbon made by pyrolysis of wood, has a long history as a metal - smelting fuel, as a filter material and adsorbent and as an artist ' s material and is one of the three ingredients of gunpowder. cellulose, the world ' s most abundant organic polymer, can be converted into energy, fuels, materials and chemical feedstock. products made from cellulose include rayon and cellophane, wallpaper paste, biobutanol and gun cotton. sugarcane, rapeseed and soy are some of the plants with a highly fermentable sugar or oil content that are used as sources of biofuels, important alternatives to fossil fuels, such as biodiesel. sweetgrass was used by native americans to ward off bugs like mosquitoes. these bug repelling properties of sweetgrass were later found by the american chemical society in the molecules phytol and coumarin. = = plant ecology = = plant ecology is the science of the functional relationships between plants and their habitats β the environments where they complete their life cycles. plant ecologists study the composition of local and regional floras, their biodiversity, genetic diversity and fitness, the adaptation of plants to their environment, and their competitive or mutualistic interactions with other species. some ecologists even rely on empirical data from indigenous people that is gathered by ethnobotanists. this information can relay a great deal of information on how the land once was thousands of years ago and how it has changed over that time. the goals of plant ecology are to understand the causes of their distribution patterns, productivity, environmental impact, evolution, and responses to environmental change. plants depend on certain edaphic ( soil ) and climatic factors in their environment but can modify these factors too. for example, they can change their environment ' s albedo, increase runoff interception, stabilise mineral soils and develop their organic content, and affect local temperature. plants compete with other organisms in their ecosystem for resources. they interact with their neighbours at a variety of spatial scales in groups, populations and communities that collectively constitute vegetation. regions with characteristic vegetation types and dominant plants as well as similar abiot
##s can be tested against these mouse models. gene therapy is the genetic engineering of humans, generally by replacing defective genes with effective ones. clinical research using somatic gene therapy has been conducted with several diseases, including x - linked scid, chronic lymphocytic leukemia ( cll ), and parkinson ' s disease. in 2012, alipogene tiparvovec became the first gene therapy treatment to be approved for clinical use. in 2015 a virus was used to insert a healthy gene into the skin cells of a boy suffering from a rare skin disease, epidermolysis bullosa, in order to grow, and then graft healthy skin onto 80 percent of the boy ' s body which was affected by the illness. germline gene therapy would result in any change being inheritable, which has raised concerns within the scientific community. in 2015, crispr was used to edit the dna of non - viable human embryos, leading scientists of major world academies to call for a moratorium on inheritable human genome edits. there are also concerns that the technology could be used not just for treatment, but for enhancement, modification or alteration of a human beings ' appearance, adaptability, intelligence, character or behavior. the distinction between cure and enhancement can also be difficult to establish. in november 2018, he jiankui announced that he had edited the genomes of two human embryos, to attempt to disable the ccr5 gene, which codes for a receptor that hiv uses to enter cells. the work was widely condemned as unethical, dangerous, and premature. currently, germline modification is banned in 40 countries. scientists that do this type of research will often let embryos grow for a few days without allowing it to develop into a baby. researchers are altering the genome of pigs to induce the growth of human organs, with the aim of increasing the success of pig to human organ transplantation. scientists are creating " gene drives ", changing the genomes of mosquitoes to make them immune to malaria, and then looking to spread the genetically altered mosquitoes throughout the mosquito population in the hopes of eliminating the disease. = = = research = = = genetic engineering is an important tool for natural scientists, with the creation of transgenic organisms one of the most important tools for analysis of gene function. genes and other genetic information from a wide range of organisms can be inserted into bacteria for storage and modification, creating genetically modified bacteria in the process. bacteria are cheap, easy to grow, clonal, multi
Question: A boy pulls a sweater off over his head. His hair is attracted to the sweater due to
A) a magnetic force
B) a heat transfer
C) a chemical change
D) an electrical charge
|
D) an electrical charge
|
Context:
= = = = = = environmental remediation = = = environmental remediation is the process through which contaminants or pollutants in soil, water and other media are removed to improve environmental quality. the main focus is the reduction of hazardous substances within the environment. some of the areas involved in environmental remediation include ; soil contamination, hazardous waste, groundwater contamination, oil, gas and chemical spills. there are three most common types of environmental remediation. these include soil, water, and sediment remediation. soil remediation consists of removing contaminants in soil, as these pose great risks to humans and the ecosystem. some examples of this are heavy metals, pesticides, and radioactive materials. depending on the contaminant the remedial processes can be physical, chemical, thermal, or biological. water remediation is one of the most important considering water is an essential natural resource. depending on the source of water there will be different contaminants. surface water contamination mainly consists of agricultural, animal, and industrial waste, as well as acid mine drainage. there has been a rise in the need for water remediation due to the increased discharge of industrial waste, leading to a demand for sustainable water solutions. the market for water remediation is expected to consistently increase to $ 19. 6 billion by 2030. sediment remediation consists of removing contaminated sediments. is it almost similar to soil remediation except it is often more sophisticated as it involves additional contaminants. to reduce the contaminants it is likely to use physical, chemical, and biological processes that help with source control, but if these processes are executed correctly, there ' s a risk of contamination resurfacing. = = = solid waste management = = = solid waste management is the purification, consumption, reuse, disposal, and treatment of solid waste that is undertaken by the government or the ruling bodies of a city / town. it refers to the collection, treatment, and disposal of non - soluble, solid waste material. solid waste is associated with both industrial, institutional, commercial and residential activities. hazardous solid waste, when improperly disposed can encourage the infestation of insects and rodents, contributing to the spread of diseases. some of the most common types of solid waste management include ; landfills, vermicomposting, composting, recycling, and incineration. however, a major barrier for solid waste management practices is the high costs associated with recycling
, lightning strikes, tornadoes, building fires, wildfires, and mass shootings disabling most of the system if not the entirety of it. geographic redundancy locations can be more than 621 miles ( 999 km ) continental, more than 62 miles apart and less than 93 miles ( 150 km ) apart, less than 62 miles apart, but not on the same campus, or different buildings that are more than 300 feet ( 91 m ) apart on the same campus. the following methods can reduce the risks of damage by a fire conflagration : large buildings at least 80 feet ( 24 m ) to 110 feet ( 34 m ) apart, but sometimes a minimum of 210 feet ( 64 m ) apart. : 9 high - rise buildings at least 82 feet ( 25 m ) apart : 12 open spaces clear of flammable vegetation within 200 feet ( 61 m ) on each side of objects different wings on the same building, in rooms that are separated by more than 300 feet ( 91 m ) different floors on the same wing of a building in rooms that are horizontally offset by a minimum of 70 feet ( 21 m ) with fire walls between the rooms that are on different floors two rooms separated by another room, leaving at least a 70 - foot gap between the two rooms there should be a minimum of two separated fire walls and on opposite sides of a corridor geographic redundancy is used by amazon web services ( aws ), google cloud platform ( gcp ), microsoft azure, netflix, dropbox, salesforce, linkedin, paypal, twitter, facebook, apple icloud, cisco meraki, and many others to provide geographic redundancy, high availability, fault tolerance and to ensure availability and reliability for their cloud services. as another example, to minimize risk of damage from severe windstorms or water damage, buildings can be located at least 2 miles ( 3. 2 km ) away from the shore, with an elevation of at least 5 feet ( 1. 5 m ) above sea level. for additional protection, they can be located at least 100 feet ( 30 m ) away from flood plain areas. = = functions of redundancy = = the two functions of redundancy are passive redundancy and active redundancy. both functions prevent performance decline from exceeding specification limits without human intervention using extra capacity. passive redundancy uses excess capacity to reduce the impact of component failures. one common form of passive redundancy is the extra strength of cabling and struts used in bridges.
industrial applications. this branch of biotechnology is the most used for the industries of refining and combustion principally on the production of bio - oils with photosynthetic micro - algae. green biotechnology is biotechnology applied to agricultural processes. an example would be the selection and domestication of plants via micropropagation. another example is the designing of transgenic plants to grow under specific environments in the presence ( or absence ) of chemicals. one hope is that green biotechnology might produce more environmentally friendly solutions than traditional industrial agriculture. an example of this is the engineering of a plant to express a pesticide, thereby ending the need of external application of pesticides. an example of this would be bt corn. whether or not green biotechnology products such as this are ultimately more environmentally friendly is a topic of considerable debate. it is commonly considered as the next phase of green revolution, which can be seen as a platform to eradicate world hunger by using technologies which enable the production of more fertile and resistant, towards biotic and abiotic stress, plants and ensures application of environmentally friendly fertilizers and the use of biopesticides, it is mainly focused on the development of agriculture. on the other hand, some of the uses of green biotechnology involve microorganisms to clean and reduce waste. red biotechnology is the use of biotechnology in the medical and pharmaceutical industries, and health preservation. this branch involves the production of vaccines and antibiotics, regenerative therapies, creation of artificial organs and new diagnostics of diseases. as well as the development of hormones, stem cells, antibodies, sirna and diagnostic tests. white biotechnology, also known as industrial biotechnology, is biotechnology applied to industrial processes. an example is the designing of an organism to produce a useful chemical. another example is the using of enzymes as industrial catalysts to either produce valuable chemicals or destroy hazardous / polluting chemicals. white biotechnology tends to consume less in resources than traditional processes used to produce industrial goods. yellow biotechnology refers to the use of biotechnology in food production ( food industry ), for example in making wine ( winemaking ), cheese ( cheesemaking ), and beer ( brewing ) by fermentation. it has also been used to refer to biotechnology applied to insects. this includes biotechnology - based approaches for the control of harmful insects, the characterisation and utilisation of active ingredients or genes of insects for research, or application in agriculture and medicine and various other approaches. gray biotechnology is dedicated to environmental applications, and focused on the maintenance of biodiversity and the remotion of poll
masculinity and warmth. the five phases β fire, earth, metal, wood, and water β described a cycle of transformations in nature. the water turned into wood, which turned into the fire when it burned. the ashes left by fire were earth. using these principles, chinese philosophers and doctors explored human anatomy, characterizing organs as predominantly yin or yang, and understood the relationship between the pulse, the heart, and the flow of blood in the body centuries before it became accepted in the west. little evidence survives of how ancient indian cultures around the indus river understood nature, but some of their perspectives may be reflected in the vedas, a set of sacred hindu texts. they reveal a conception of the universe as ever - expanding and constantly being recycled and reformed. surgeons in the ayurvedic tradition saw health and illness as a combination of three humors : wind, bile and phlegm. a healthy life resulted from a balance among these humors. in ayurvedic thought, the body consisted of five elements : earth, water, fire, wind, and space. ayurvedic surgeons performed complex surgeries and developed a detailed understanding of human anatomy. pre - socratic philosophers in ancient greek culture brought natural philosophy a step closer to direct inquiry about cause and effect in nature between 600 and 400 bc. however, an element of magic and mythology remained. natural phenomena such as earthquakes and eclipses were explained increasingly in the context of nature itself instead of being attributed to angry gods. thales of miletus, an early philosopher who lived from 625 to 546 bc, explained earthquakes by theorizing that the world floated on water and that water was the fundamental element in nature. in the 5th century bc, leucippus was an early exponent of atomism, the idea that the world is made up of fundamental indivisible particles. pythagoras applied greek innovations in mathematics to astronomy and suggested that the earth was spherical. = = = aristotelian natural philosophy ( 400 bc β 1100 ad ) = = = later socratic and platonic thought focused on ethics, morals, and art and did not attempt an investigation of the physical world ; plato criticized pre - socratic thinkers as materialists and anti - religionists. aristotle, however, a student of plato who lived from 384 to 322 bc, paid closer attention to the natural world in his philosophy. in his history of animals, he described the inner workings of 110 species, including the stingray, catfish and
remediation include ; soil contamination, hazardous waste, groundwater contamination, oil, gas and chemical spills. there are three most common types of environmental remediation. these include soil, water, and sediment remediation. soil remediation consists of removing contaminants in soil, as these pose great risks to humans and the ecosystem. some examples of this are heavy metals, pesticides, and radioactive materials. depending on the contaminant the remedial processes can be physical, chemical, thermal, or biological. water remediation is one of the most important considering water is an essential natural resource. depending on the source of water there will be different contaminants. surface water contamination mainly consists of agricultural, animal, and industrial waste, as well as acid mine drainage. there has been a rise in the need for water remediation due to the increased discharge of industrial waste, leading to a demand for sustainable water solutions. the market for water remediation is expected to consistently increase to $ 19. 6 billion by 2030. sediment remediation consists of removing contaminated sediments. is it almost similar to soil remediation except it is often more sophisticated as it involves additional contaminants. to reduce the contaminants it is likely to use physical, chemical, and biological processes that help with source control, but if these processes are executed correctly, there ' s a risk of contamination resurfacing. = = = solid waste management = = = solid waste management is the purification, consumption, reuse, disposal, and treatment of solid waste that is undertaken by the government or the ruling bodies of a city / town. it refers to the collection, treatment, and disposal of non - soluble, solid waste material. solid waste is associated with both industrial, institutional, commercial and residential activities. hazardous solid waste, when improperly disposed can encourage the infestation of insects and rodents, contributing to the spread of diseases. some of the most common types of solid waste management include ; landfills, vermicomposting, composting, recycling, and incineration. however, a major barrier for solid waste management practices is the high costs associated with recycling and the risks of creating more pollution. = = = e - waste recycling = = = the recycling of electronic waste ( e - waste ) has seen significant technological advancements due to increasing environmental concerns and the growing volume of electronic product disposals. traditional e - waste recycling methods, which often involve manual disassemb
##ructing the channel depends on the nature of the shoals. a soft shoal in the bed of a river is due to deposit from a diminution in velocity of flow, produced by a reduction in fall and by a widening of the channel, or to a loss in concentration of the scour of the main current in passing over from one concave bank to the next on the opposite side. the lowering of such a shoal by dredging merely effects a temporary deepening, for it soon forms again from the causes which produced it. the removal, moreover, of the rocky obstructions at rapids, though increasing the depth and equalizing the flow at these places, produces a lowering of the river above the rapids by facilitating the efflux, which may result in the appearance of fresh shoals at the low stage of the river. where, however, narrow rocky reefs or other hard shoals stretch across the bottom of a river and present obstacles to the erosion by the current of the soft materials forming the bed of the river above and below, their removal may result in permanent improvement by enabling the river to deepen its bed by natural scour. the capability of a river to provide a waterway for navigation during the summer or throughout the dry season depends on the depth that can be secured in the channel at the lowest stage. the problem in the dry season is the small discharge and deficiency in scour during this period. a typical solution is to restrict the width of the low - water channel, concentrate all of the flow in it, and also to fix its position so that it is scoured out every year by the floods which follow the deepest part of the bed along the line of the strongest current. this can be effected by closing subsidiary low - water channels with dikes across them, and narrowing the channel at the low stage by low - dipping cross dikes extending from the river banks down the slope and pointing slightly up - stream so as to direct the water flowing over them into a central channel. = = estuarine works = = the needs of navigation may also require that a stable, continuous, navigable channel is prolonged from the navigable river to deep water at the mouth of the estuary. the interaction of river flow and tide needs to be modeled by computer or using scale models, moulded to the configuration of the estuary under consideration and reproducing in miniature the tidal ebb and flow and fresh - water discharge over a bed of fine sand, in which various lines of training walls can be successively inserted. the models
the designing of transgenic plants to grow under specific environments in the presence ( or absence ) of chemicals. one hope is that green biotechnology might produce more environmentally friendly solutions than traditional industrial agriculture. an example of this is the engineering of a plant to express a pesticide, thereby ending the need of external application of pesticides. an example of this would be bt corn. whether or not green biotechnology products such as this are ultimately more environmentally friendly is a topic of considerable debate. it is commonly considered as the next phase of green revolution, which can be seen as a platform to eradicate world hunger by using technologies which enable the production of more fertile and resistant, towards biotic and abiotic stress, plants and ensures application of environmentally friendly fertilizers and the use of biopesticides, it is mainly focused on the development of agriculture. on the other hand, some of the uses of green biotechnology involve microorganisms to clean and reduce waste. red biotechnology is the use of biotechnology in the medical and pharmaceutical industries, and health preservation. this branch involves the production of vaccines and antibiotics, regenerative therapies, creation of artificial organs and new diagnostics of diseases. as well as the development of hormones, stem cells, antibodies, sirna and diagnostic tests. white biotechnology, also known as industrial biotechnology, is biotechnology applied to industrial processes. an example is the designing of an organism to produce a useful chemical. another example is the using of enzymes as industrial catalysts to either produce valuable chemicals or destroy hazardous / polluting chemicals. white biotechnology tends to consume less in resources than traditional processes used to produce industrial goods. yellow biotechnology refers to the use of biotechnology in food production ( food industry ), for example in making wine ( winemaking ), cheese ( cheesemaking ), and beer ( brewing ) by fermentation. it has also been used to refer to biotechnology applied to insects. this includes biotechnology - based approaches for the control of harmful insects, the characterisation and utilisation of active ingredients or genes of insects for research, or application in agriculture and medicine and various other approaches. gray biotechnology is dedicated to environmental applications, and focused on the maintenance of biodiversity and the remotion of pollutants. brown biotechnology is related to the management of arid lands and deserts. one application is the creation of enhanced seeds that resist extreme environmental conditions of arid regions, which is related to the innovation, creation of agriculture techniques and management of resources. violet biotechnology is related to law, ethical and philosophical issues around biotechnology. micro
current in passing over from one concave bank to the next on the opposite side. the lowering of such a shoal by dredging merely effects a temporary deepening, for it soon forms again from the causes which produced it. the removal, moreover, of the rocky obstructions at rapids, though increasing the depth and equalizing the flow at these places, produces a lowering of the river above the rapids by facilitating the efflux, which may result in the appearance of fresh shoals at the low stage of the river. where, however, narrow rocky reefs or other hard shoals stretch across the bottom of a river and present obstacles to the erosion by the current of the soft materials forming the bed of the river above and below, their removal may result in permanent improvement by enabling the river to deepen its bed by natural scour. the capability of a river to provide a waterway for navigation during the summer or throughout the dry season depends on the depth that can be secured in the channel at the lowest stage. the problem in the dry season is the small discharge and deficiency in scour during this period. a typical solution is to restrict the width of the low - water channel, concentrate all of the flow in it, and also to fix its position so that it is scoured out every year by the floods which follow the deepest part of the bed along the line of the strongest current. this can be effected by closing subsidiary low - water channels with dikes across them, and narrowing the channel at the low stage by low - dipping cross dikes extending from the river banks down the slope and pointing slightly up - stream so as to direct the water flowing over them into a central channel. = = estuarine works = = the needs of navigation may also require that a stable, continuous, navigable channel is prolonged from the navigable river to deep water at the mouth of the estuary. the interaction of river flow and tide needs to be modeled by computer or using scale models, moulded to the configuration of the estuary under consideration and reproducing in miniature the tidal ebb and flow and fresh - water discharge over a bed of fine sand, in which various lines of training walls can be successively inserted. the models should be capable of furnishing valuable indications of the respective effects and comparative merits of the different schemes proposed for works. = = see also = = bridge scour flood control = = references = = = = external links = = u. s. army corps of engineers β civil works program river morphology and stream restoration references
##nts from the air to reduce the potential adverse effects on humans and the environment. the process of air purification may be performed using methods such as mechanical filtration, ionization, activated carbon adsorption, photocatalytic oxidation, and ultraviolet light germicidal irradiation. = = = sewage treatment = = = = = = environmental remediation = = = environmental remediation is the process through which contaminants or pollutants in soil, water and other media are removed to improve environmental quality. the main focus is the reduction of hazardous substances within the environment. some of the areas involved in environmental remediation include ; soil contamination, hazardous waste, groundwater contamination, oil, gas and chemical spills. there are three most common types of environmental remediation. these include soil, water, and sediment remediation. soil remediation consists of removing contaminants in soil, as these pose great risks to humans and the ecosystem. some examples of this are heavy metals, pesticides, and radioactive materials. depending on the contaminant the remedial processes can be physical, chemical, thermal, or biological. water remediation is one of the most important considering water is an essential natural resource. depending on the source of water there will be different contaminants. surface water contamination mainly consists of agricultural, animal, and industrial waste, as well as acid mine drainage. there has been a rise in the need for water remediation due to the increased discharge of industrial waste, leading to a demand for sustainable water solutions. the market for water remediation is expected to consistently increase to $ 19. 6 billion by 2030. sediment remediation consists of removing contaminated sediments. is it almost similar to soil remediation except it is often more sophisticated as it involves additional contaminants. to reduce the contaminants it is likely to use physical, chemical, and biological processes that help with source control, but if these processes are executed correctly, there ' s a risk of contamination resurfacing. = = = solid waste management = = = solid waste management is the purification, consumption, reuse, disposal, and treatment of solid waste that is undertaken by the government or the ruling bodies of a city / town. it refers to the collection, treatment, and disposal of non - soluble, solid waste material. solid waste is associated with both industrial, institutional, commercial and residential activities. hazardous solid waste, when improperly disposed can encourage the
the injuries of the inundations they have been designed to prevent, as the escape of floods from the raised river must occur sooner or later. inadequate planning controls which have permitted development on floodplains have been blamed for the flooding of domestic properties. channelization was done under the auspices or overall direction of engineers employed by the local authority or the national government. one of the most heavily channelized areas in the united states is west tennessee, where every major stream with one exception ( the hatchie river ) has been partially or completely channelized. channelization of a stream may be undertaken for several reasons. one is to make a stream more suitable for navigation or for navigation by larger vessels with deep draughts. another is to restrict water to a certain area of a stream ' s natural bottom lands so that the bulk of such lands can be made available for agriculture. a third reason is flood control, with the idea of giving a stream a sufficiently large and deep channel so that flooding beyond those limits will be minimal or nonexistent, at least on a routine basis. one major reason is to reduce natural erosion ; as a natural waterway curves back and forth, it usually deposits sand and gravel on the inside of the corners where the water flows slowly, and cuts sand, gravel, subsoil, and precious topsoil from the outside corners where it flows rapidly due to a change in direction. unlike sand and gravel, the topsoil that is eroded does not get deposited on the inside of the next corner of the river. it simply washes away. = = loss of wetlands = = channelization has several predictable and negative effects. one of them is loss of wetlands. wetlands are an excellent habitat for multiple forms of wildlife, and additionally serve as a " filter " for much of the world ' s surface fresh water. another is the fact that channelized streams are almost invariably straightened. for example, the channelization of florida ' s kissimmee river has been cited as a cause contributing to the loss of wetlands. this straightening causes the streams to flow more rapidly, which can, in some instances, vastly increase soil erosion. it can also increase flooding downstream from the channelized area, as larger volumes of water traveling more rapidly than normal can reach choke points over a shorter period of time than they otherwise would, with a net effect of flood control in one area coming at the expense of aggravated flooding in another. in addition, studies have shown that stream channelization results in declines of river fish populations. : 3 - 1ff a
Question: In some communities, forests are cleared through burning, and the deforested land is used for farming. Which of these is a negative consequence of this practice?
A) decreased plant and animal diversity
B) increased availability of soil nutrients
C) greater access to water for plants and animals
D) less sunlight reaching the ground
|
A) decreased plant and animal diversity
|
Context:
from the oil of jasminum grandiflorum which regulates wound responses in plants by unblocking the expression of genes required in the systemic acquired resistance response to pathogen attack. in addition to being the primary energy source for plants, light functions as a signalling device, providing information to the plant, such as how much sunlight the plant receives each day. this can result in adaptive changes in a process known as photomorphogenesis. phytochromes are the photoreceptors in a plant that are sensitive to light. = = plant anatomy and morphology = = plant anatomy is the study of the structure of plant cells and tissues, whereas plant morphology is the study of their external form. all plants are multicellular eukaryotes, their dna stored in nuclei. the characteristic features of plant cells that distinguish them from those of animals and fungi include a primary cell wall composed of the polysaccharides cellulose, hemicellulose and pectin, larger vacuoles than in animal cells and the presence of plastids with unique photosynthetic and biosynthetic functions as in the chloroplasts. other plastids contain storage products such as starch ( amyloplasts ) or lipids ( elaioplasts ). uniquely, streptophyte cells and those of the green algal order trentepohliales divide by construction of a phragmoplast as a template for building a cell plate late in cell division. the bodies of vascular plants including clubmosses, ferns and seed plants ( gymnosperms and angiosperms ) generally have aerial and subterranean subsystems. the shoots consist of stems bearing green photosynthesising leaves and reproductive structures. the underground vascularised roots bear root hairs at their tips and generally lack chlorophyll. non - vascular plants, the liverworts, hornworts and mosses do not produce ground - penetrating vascular roots and most of the plant participates in photosynthesis. the sporophyte generation is nonphotosynthetic in liverworts but may be able to contribute part of its energy needs by photosynthesis in mosses and hornworts. the root system and the shoot system are interdependent β the usually nonphotosynthetic root system depends on the shoot system for food, and the usually photosynthetic shoot system depends on water and minerals from the root system. cells in each system are capable
chloroplasts make are used for many things, such as providing material to build cell membranes out of and making the polymer cutin which is found in the plant cuticle that protects land plants from drying out. plants synthesise a number of unique polymers like the polysaccharide molecules cellulose, pectin and xyloglucan from which the land plant cell wall is constructed. vascular land plants make lignin, a polymer used to strengthen the secondary cell walls of xylem tracheids and vessels to keep them from collapsing when a plant sucks water through them under water stress. lignin is also used in other cell types like sclerenchyma fibres that provide structural support for a plant and is a major constituent of wood. sporopollenin is a chemically resistant polymer found in the outer cell walls of spores and pollen of land plants responsible for the survival of early land plant spores and the pollen of seed plants in the fossil record. it is widely regarded as a marker for the start of land plant evolution during the ordovician period. the concentration of carbon dioxide in the atmosphere today is much lower than it was when plants emerged onto land during the ordovician and silurian periods. many monocots like maize and the pineapple and some dicots like the asteraceae have since independently evolved pathways like crassulacean acid metabolism and the c4 carbon fixation pathway for photosynthesis which avoid the losses resulting from photorespiration in the more common c3 carbon fixation pathway. these biochemical strategies are unique to land plants. = = = medicine and materials = = = phytochemistry is a branch of plant biochemistry primarily concerned with the chemical substances produced by plants during secondary metabolism. some of these compounds are toxins such as the alkaloid coniine from hemlock. others, such as the essential oils peppermint oil and lemon oil are useful for their aroma, as flavourings and spices ( e. g., capsaicin ), and in medicine as pharmaceuticals as in opium from opium poppies. many medicinal and recreational drugs, such as tetrahydrocannabinol ( active ingredient in cannabis ), caffeine, morphine and nicotine come directly from plants. others are simple derivatives of botanical natural products. for example, the pain killer aspirin is the acetyl ester of salicylic acid, originally isolated from the bark of willow trees, and
much sunlight the plant receives each day. this can result in adaptive changes in a process known as photomorphogenesis. phytochromes are the photoreceptors in a plant that are sensitive to light. = = plant anatomy and morphology = = plant anatomy is the study of the structure of plant cells and tissues, whereas plant morphology is the study of their external form. all plants are multicellular eukaryotes, their dna stored in nuclei. the characteristic features of plant cells that distinguish them from those of animals and fungi include a primary cell wall composed of the polysaccharides cellulose, hemicellulose and pectin, larger vacuoles than in animal cells and the presence of plastids with unique photosynthetic and biosynthetic functions as in the chloroplasts. other plastids contain storage products such as starch ( amyloplasts ) or lipids ( elaioplasts ). uniquely, streptophyte cells and those of the green algal order trentepohliales divide by construction of a phragmoplast as a template for building a cell plate late in cell division. the bodies of vascular plants including clubmosses, ferns and seed plants ( gymnosperms and angiosperms ) generally have aerial and subterranean subsystems. the shoots consist of stems bearing green photosynthesising leaves and reproductive structures. the underground vascularised roots bear root hairs at their tips and generally lack chlorophyll. non - vascular plants, the liverworts, hornworts and mosses do not produce ground - penetrating vascular roots and most of the plant participates in photosynthesis. the sporophyte generation is nonphotosynthetic in liverworts but may be able to contribute part of its energy needs by photosynthesis in mosses and hornworts. the root system and the shoot system are interdependent β the usually nonphotosynthetic root system depends on the shoot system for food, and the usually photosynthetic shoot system depends on water and minerals from the root system. cells in each system are capable of creating cells of the other and producing adventitious shoots or roots. stolons and tubers are examples of shoots that can grow roots. roots that spread out close to the surface, such as those of willows, can produce shoots and ultimately new plants. in the event that one of the systems is lost
sugar ) for export to the rest of the plant. unlike in animals ( which lack chloroplasts ), plants and their eukaryote relatives have delegated many biochemical roles to their chloroplasts, including synthesising all their fatty acids, and most amino acids. the fatty acids that chloroplasts make are used for many things, such as providing material to build cell membranes out of and making the polymer cutin which is found in the plant cuticle that protects land plants from drying out. plants synthesise a number of unique polymers like the polysaccharide molecules cellulose, pectin and xyloglucan from which the land plant cell wall is constructed. vascular land plants make lignin, a polymer used to strengthen the secondary cell walls of xylem tracheids and vessels to keep them from collapsing when a plant sucks water through them under water stress. lignin is also used in other cell types like sclerenchyma fibres that provide structural support for a plant and is a major constituent of wood. sporopollenin is a chemically resistant polymer found in the outer cell walls of spores and pollen of land plants responsible for the survival of early land plant spores and the pollen of seed plants in the fossil record. it is widely regarded as a marker for the start of land plant evolution during the ordovician period. the concentration of carbon dioxide in the atmosphere today is much lower than it was when plants emerged onto land during the ordovician and silurian periods. many monocots like maize and the pineapple and some dicots like the asteraceae have since independently evolved pathways like crassulacean acid metabolism and the c4 carbon fixation pathway for photosynthesis which avoid the losses resulting from photorespiration in the more common c3 carbon fixation pathway. these biochemical strategies are unique to land plants. = = = medicine and materials = = = phytochemistry is a branch of plant biochemistry primarily concerned with the chemical substances produced by plants during secondary metabolism. some of these compounds are toxins such as the alkaloid coniine from hemlock. others, such as the essential oils peppermint oil and lemon oil are useful for their aroma, as flavourings and spices ( e. g., capsaicin ), and in medicine as pharmaceuticals as in opium from opium poppies. many medicinal and recreational drugs, such as tetrahydrocannabino
pectin and xyloglucan from which the land plant cell wall is constructed. vascular land plants make lignin, a polymer used to strengthen the secondary cell walls of xylem tracheids and vessels to keep them from collapsing when a plant sucks water through them under water stress. lignin is also used in other cell types like sclerenchyma fibres that provide structural support for a plant and is a major constituent of wood. sporopollenin is a chemically resistant polymer found in the outer cell walls of spores and pollen of land plants responsible for the survival of early land plant spores and the pollen of seed plants in the fossil record. it is widely regarded as a marker for the start of land plant evolution during the ordovician period. the concentration of carbon dioxide in the atmosphere today is much lower than it was when plants emerged onto land during the ordovician and silurian periods. many monocots like maize and the pineapple and some dicots like the asteraceae have since independently evolved pathways like crassulacean acid metabolism and the c4 carbon fixation pathway for photosynthesis which avoid the losses resulting from photorespiration in the more common c3 carbon fixation pathway. these biochemical strategies are unique to land plants. = = = medicine and materials = = = phytochemistry is a branch of plant biochemistry primarily concerned with the chemical substances produced by plants during secondary metabolism. some of these compounds are toxins such as the alkaloid coniine from hemlock. others, such as the essential oils peppermint oil and lemon oil are useful for their aroma, as flavourings and spices ( e. g., capsaicin ), and in medicine as pharmaceuticals as in opium from opium poppies. many medicinal and recreational drugs, such as tetrahydrocannabinol ( active ingredient in cannabis ), caffeine, morphine and nicotine come directly from plants. others are simple derivatives of botanical natural products. for example, the pain killer aspirin is the acetyl ester of salicylic acid, originally isolated from the bark of willow trees, and a wide range of opiate painkillers like heroin are obtained by chemical modification of morphine obtained from the opium poppy. popular stimulants come from plants, such as caffeine from coffee, tea and chocolate, and nicotine from tobacco. most alcoholic beverages come from fermentation of carbohy
frits went. the first known auxin, indole - 3 - acetic acid ( iaa ), which promotes cell growth, was only isolated from plants about 50 years later. this compound mediates the tropic responses of shoots and roots towards light and gravity. the finding in 1939 that plant callus could be maintained in culture containing iaa, followed by the observation in 1947 that it could be induced to form roots and shoots by controlling the concentration of growth hormones were key steps in the development of plant biotechnology and genetic modification. cytokinins are a class of plant hormones named for their control of cell division ( especially cytokinesis ). the natural cytokinin zeatin was discovered in corn, zea mays, and is a derivative of the purine adenine. zeatin is produced in roots and transported to shoots in the xylem where it promotes cell division, bud development, and the greening of chloroplasts. the gibberelins, such as gibberelic acid are diterpenes synthesised from acetyl coa via the mevalonate pathway. they are involved in the promotion of germination and dormancy - breaking in seeds, in regulation of plant height by controlling stem elongation and the control of flowering. abscisic acid ( aba ) occurs in all land plants except liverworts, and is synthesised from carotenoids in the chloroplasts and other plastids. it inhibits cell division, promotes seed maturation, and dormancy, and promotes stomatal closure. it was so named because it was originally thought to control abscission. ethylene is a gaseous hormone that is produced in all higher plant tissues from methionine. it is now known to be the hormone that stimulates or regulates fruit ripening and abscission, and it, or the synthetic growth regulator ethephon which is rapidly metabolised to produce ethylene, are used on industrial scale to promote ripening of cotton, pineapples and other climacteric crops. another class of phytohormones is the jasmonates, first isolated from the oil of jasminum grandiflorum which regulates wound responses in plants by unblocking the expression of genes required in the systemic acquired resistance response to pathogen attack. in addition to being the primary energy source for plants, light functions as a signalling device, providing information to the plant, such as how
could be maintained in culture containing iaa, followed by the observation in 1947 that it could be induced to form roots and shoots by controlling the concentration of growth hormones were key steps in the development of plant biotechnology and genetic modification. cytokinins are a class of plant hormones named for their control of cell division ( especially cytokinesis ). the natural cytokinin zeatin was discovered in corn, zea mays, and is a derivative of the purine adenine. zeatin is produced in roots and transported to shoots in the xylem where it promotes cell division, bud development, and the greening of chloroplasts. the gibberelins, such as gibberelic acid are diterpenes synthesised from acetyl coa via the mevalonate pathway. they are involved in the promotion of germination and dormancy - breaking in seeds, in regulation of plant height by controlling stem elongation and the control of flowering. abscisic acid ( aba ) occurs in all land plants except liverworts, and is synthesised from carotenoids in the chloroplasts and other plastids. it inhibits cell division, promotes seed maturation, and dormancy, and promotes stomatal closure. it was so named because it was originally thought to control abscission. ethylene is a gaseous hormone that is produced in all higher plant tissues from methionine. it is now known to be the hormone that stimulates or regulates fruit ripening and abscission, and it, or the synthetic growth regulator ethephon which is rapidly metabolised to produce ethylene, are used on industrial scale to promote ripening of cotton, pineapples and other climacteric crops. another class of phytohormones is the jasmonates, first isolated from the oil of jasminum grandiflorum which regulates wound responses in plants by unblocking the expression of genes required in the systemic acquired resistance response to pathogen attack. in addition to being the primary energy source for plants, light functions as a signalling device, providing information to the plant, such as how much sunlight the plant receives each day. this can result in adaptive changes in a process known as photomorphogenesis. phytochromes are the photoreceptors in a plant that are sensitive to light. = = plant anatomy and morphology = = plant anatomy is the study of the structure of
shaping the cell. cell membranes are involved in various cellular processes such as cell adhesion, storing electrical energy, and cell signalling and serve as the attachment surface for several extracellular structures such as a cell wall, glycocalyx, and cytoskeleton. within the cytoplasm of a cell, there are many biomolecules such as proteins and nucleic acids. in addition to biomolecules, eukaryotic cells have specialized structures called organelles that have their own lipid bilayers or are spatially units. these organelles include the cell nucleus, which contains most of the cell ' s dna, or mitochondria, which generate adenosine triphosphate ( atp ) to power cellular processes. other organelles such as endoplasmic reticulum and golgi apparatus play a role in the synthesis and packaging of proteins, respectively. biomolecules such as proteins can be engulfed by lysosomes, another specialized organelle. plant cells have additional organelles that distinguish them from animal cells such as a cell wall that provides support for the plant cell, chloroplasts that harvest sunlight energy to produce sugar, and vacuoles that provide storage and structural support as well as being involved in reproduction and breakdown of plant seeds. eukaryotic cells also have cytoskeleton that is made up of microtubules, intermediate filaments, and microfilaments, all of which provide support for the cell and are involved in the movement of the cell and its organelles. in terms of their structural composition, the microtubules are made up of tubulin ( e. g., Ξ± - tubulin and Ξ² - tubulin ) whereas intermediate filaments are made up of fibrous proteins. microfilaments are made up of actin molecules that interact with other strands of proteins. = = = metabolism = = = all cells require energy to sustain cellular processes. metabolism is the set of chemical reactions in an organism. the three main purposes of metabolism are : the conversion of food to energy to run cellular processes ; the conversion of food / fuel to monomer building blocks ; and the elimination of metabolic wastes. these enzyme - catalyzed reactions allow organisms to grow and reproduce, maintain their structures, and respond to their environments. metabolic reactions may be categorized as catabolic β the breaking down of compounds ( for example, the breaking down of glucose to pyruvate by cellular respiration
paper withdrawn due to a crucial algebraic error in section 3.
soil erosion. plants are crucial to the future of human society as they provide food, oxygen, biochemicals, and products for people, as well as creating and preserving soil. historically, all living things were classified as either animals or plants and botany covered the study of all organisms not considered animals. botanists examine both the internal functions and processes within plant organelles, cells, tissues, whole plants, plant populations and plant communities. at each of these levels, a botanist may be concerned with the classification ( taxonomy ), phylogeny and evolution, structure ( anatomy and morphology ), or function ( physiology ) of plant life. the strictest definition of " plant " includes only the " land plants " or embryophytes, which include seed plants ( gymnosperms, including the pines, and flowering plants ) and the free - sporing cryptogams including ferns, clubmosses, liverworts, hornworts and mosses. embryophytes are multicellular eukaryotes descended from an ancestor that obtained its energy from sunlight by photosynthesis. they have life cycles with alternating haploid and diploid phases. the sexual haploid phase of embryophytes, known as the gametophyte, nurtures the developing diploid embryo sporophyte within its tissues for at least part of its life, even in the seed plants, where the gametophyte itself is nurtured by its parent sporophyte. other groups of organisms that were previously studied by botanists include bacteria ( now studied in bacteriology ), fungi ( mycology ) β including lichen - forming fungi ( lichenology ), non - chlorophyte algae ( phycology ), and viruses ( virology ). however, attention is still given to these groups by botanists, and fungi ( including lichens ) and photosynthetic protists are usually covered in introductory botany courses. palaeobotanists study ancient plants in the fossil record to provide information about the evolutionary history of plants. cyanobacteria, the first oxygen - releasing photosynthetic organisms on earth, are thought to have given rise to the ancestor of plants by entering into an endosymbiotic relationship with an early eukaryote, ultimately becoming the chloroplasts in plant cells. the new photosynthetic plants ( along with their algal relatives ) accelerated the rise in atmospheric oxygen started by the cyanobacteria, changing the
Question: Some plant cells are coated in a waxy layer that prevents water loss. This specialization is found in which plant cells?
A) internal stem cells
B) internal root cells
C) cells on the plant surface
D) cells in the plant reproductive system
|
C) cells on the plant surface
|
Context:
variation in total solar irradiance is thought to have little effect on the earth ' s surface temperature because of the thermal time constant - - the characteristic response time of the earth ' s global surface temperature to changes in forcing. this time constant is large enough to smooth annual variations but not necessarily variations having a longer period such as those due to solar inertial motion ; the magnitude of these surface temperature variations is estimated.
due to its location and climate, antarctica offers unique conditions for long - period observations across a broad wavelength regime, where important diagnostic lines for molecules and ions can be found, that are essential to understand the chemical properties of the interstellar medium. in addition to the natural benefits of the site, new technologies, resulting from astrophotonics, may allow miniaturised instruments, that are easier to winterise and advanced filters to further reduce the background in the infrared.
weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under ordinary conditions, most of the materials brought down from the high lands by torrential water courses are carried forward by the main river to the sea, or partially strewn over flat alluvial plains during floods ; the size of the materials forming the bed of the river or borne along by the stream is gradually reduced on proceeding seawards, so that in the po river in italy, for instance, pebbles and gravel are found for about 140 miles below turin, sand along the next 100 miles, and silt and mud in the last 110 miles ( 176 km ). = = channelization = = the removal of obstructions, natural or artificial
navigable capabilities of rivers. in tropical countries subject to periodical rains, the rivers are in flood during the rainy season and have hardly any flow during the rest of the year, while in temperate regions, where the rainfall is more evenly distributed throughout the year, evaporation causes the available rainfall to be much less in hot summer weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under ordinary conditions, most of the materials brought down from the high lands by torrential water courses are carried forward by the main river to the sea, or partially strewn over flat alluvial plains during floods ; the size of the materials forming the bed of the river or borne along by the stream is gradually reduced on proceeding sea
a comparison of the sensitivities of methods which allow us to determine the coordinates of a moving hot body is made.
river valley during ancient times. the papyrus was harvested by field workers and brought to processing centers where it was cut into thin strips. the strips were then laid - out side by side and covered in plant resin. the second layer of strips was laid on perpendicularly, then both pressed together until the sheet was dry. the sheets were then joined to form a roll and later used for writing. egyptian society made several significant advances during dynastic periods in many areas of technology. according to hossam elanzeery, they were the first civilization to use timekeeping devices such as sundials, shadow clocks, and obelisks and successfully leveraged their knowledge of astronomy to create a calendar model that society still uses today. they developed shipbuilding technology that saw them progress from papyrus reed vessels to cedar wood ships while also pioneering the use of rope trusses and stem - mounted rudders. the egyptians also used their knowledge of anatomy to lay the foundation for many modern medical techniques and practiced the earliest known version of neuroscience. elanzeery also states that they used and furthered mathematical science, as evidenced in the building of the pyramids. ancient egyptians also invented and pioneered many food technologies that have become the basis of modern food technology processes. based on paintings and reliefs found in tombs, as well as archaeological artifacts, scholars like paul t nicholson believe that the ancient egyptians established systematic farming practices, engaged in cereal processing, brewed beer and baked bread, processed meat, practiced viticulture and created the basis for modern wine production, and created condiments to complement, preserve and mask the flavors of their food. = = = = indus valley = = = = the indus valley civilization, situated in a resource - rich area ( in modern pakistan and northwestern india ), is notable for its early application of city planning, sanitation technologies, and plumbing. indus valley construction and architecture, called ' vaastu shastra ', suggests a thorough understanding of materials engineering, hydrology, and sanitation. = = = = china = = = = the chinese made many first - known discoveries and developments. major technological contributions from china include the earliest known form of the binary code and epigenetic sequencing, early seismological detectors, matches, paper, helicopter rotor, raised - relief map, the double - action piston pump, cast iron, water powered blast furnace bellows, the iron plough, the multi - tube seed drill, the wheelbarrow, the parachute, the compass, the rudder, the crossbow, the south pointing chariot and gunpowder
approximately corresponds to the slope of the country it traverses ; as rivers rise close to the highest part of their basins, generally in hilly regions, their fall is rapid near their source and gradually diminishes, with occasional irregularities, until, in traversing plains along the latter part of their course, their fall usually becomes quite gentle. accordingly, in large basins, rivers in most cases begin as torrents with a variable flow, and end as gently flowing rivers with a comparatively regular discharge. the irregular flow of rivers throughout their course forms one of the main difficulties in devising works for mitigating inundations or for increasing the navigable capabilities of rivers. in tropical countries subject to periodical rains, the rivers are in flood during the rainy season and have hardly any flow during the rest of the year, while in temperate regions, where the rainfall is more evenly distributed throughout the year, evaporation causes the available rainfall to be much less in hot summer weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with
earliest record of a ship under sail is that of a nile boat dating to around 7, 000 bce. from prehistoric times, egyptians likely used the power of the annual flooding of the nile to irrigate their lands, gradually learning to regulate much of it through purposely built irrigation channels and " catch " basins. the ancient sumerians in mesopotamia used a complex system of canals and levees to divert water from the tigris and euphrates rivers for irrigation. archaeologists estimate that the wheel was invented independently and concurrently in mesopotamia ( in present - day iraq ), the northern caucasus ( maykop culture ), and central europe. time estimates range from 5, 500 to 3, 000 bce with most experts putting it closer to 4, 000 bce. the oldest artifacts with drawings depicting wheeled carts date from about 3, 500 bce. more recently, the oldest - known wooden wheel in the world as of 2024 was found in the ljubljana marsh of slovenia ; austrian experts have established that the wheel is between 5, 100 and 5, 350 years old. the invention of the wheel revolutionized trade and war. it did not take long to discover that wheeled wagons could be used to carry heavy loads. the ancient sumerians used a potter ' s wheel and may have invented it. a stone pottery wheel found in the city - state of ur dates to around 3, 429 bce, and even older fragments of wheel - thrown pottery have been found in the same area. fast ( rotary ) potters ' wheels enabled early mass production of pottery, but it was the use of the wheel as a transformer of energy ( through water wheels, windmills, and even treadmills ) that revolutionized the application of nonhuman power sources. the first two - wheeled carts were derived from travois and were first used in mesopotamia and iran in around 3, 000 bce. the oldest known constructed roadways are the stone - paved streets of the city - state of ur, dating to c. 4, 000 bce, and timber roads leading through the swamps of glastonbury, england, dating to around the same period. the first long - distance road, which came into use around 3, 500 bce, spanned 2, 400 km from the persian gulf to the mediterranean sea, but was not paved and was only partially maintained. in around 2, 000 bce, the minoans on the greek island of crete built a 50 km road leading from the palace of gortyn on the south side of the island, through the mountains,
becomes quite gentle. accordingly, in large basins, rivers in most cases begin as torrents with a variable flow, and end as gently flowing rivers with a comparatively regular discharge. the irregular flow of rivers throughout their course forms one of the main difficulties in devising works for mitigating inundations or for increasing the navigable capabilities of rivers. in tropical countries subject to periodical rains, the rivers are in flood during the rainy season and have hardly any flow during the rest of the year, while in temperate regions, where the rainfall is more evenly distributed throughout the year, evaporation causes the available rainfall to be much less in hot summer weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under
. forensic histopathology is the application of histological techniques and examination to forensic pathology practice. forensic limnology is the analysis of evidence collected from crime scenes in or around fresh - water sources. examination of biological organisms, in particular diatoms, can be useful in connecting suspects with victims. forensic linguistics deals with issues in the legal system that requires linguistic expertise. forensic meteorology is a site - specific analysis of past weather conditions for a point of loss. forensic metrology is the application of metrology to assess the reliability of scientific evidence obtained through measurements forensic microbiology is the study of the necrobiome. forensic nursing is the application of nursing sciences to abusive crimes, like child abuse, or sexual abuse. categorization of wounds and traumas, collection of bodily fluids and emotional support are some of the duties of forensic nurses. forensic odontology is the study of the uniqueness of dentition, better known as the study of teeth. forensic optometry is the study of glasses and other eyewear relating to crime scenes and criminal investigations. forensic pathology is a field in which the principles of medicine and pathology are applied to determine a cause of death or injury in the context of a legal inquiry. forensic podiatry is an application of the study of feet footprint or footwear and their traces to analyze scene of crime and to establish personal identity in forensic examinations. forensic psychiatry is a specialized branch of psychiatry as applied to and based on scientific criminology. forensic psychology is the study of the mind of an individual, using forensic methods. usually it determines the circumstances behind a criminal ' s behavior. forensic seismology is the study of techniques to distinguish the seismic signals generated by underground nuclear explosions from those generated by earthquakes. forensic serology is the study of the body fluids. forensic social work is the specialist study of social work theories and their applications to a clinical, criminal justice or psychiatric setting. practitioners of forensic social work connected with the criminal justice system are often termed social supervisors, whilst the remaining use the interchangeable titles forensic social worker, approved mental health professional or forensic practitioner and they conduct specialist assessments of risk, care planning and act as an officer of the court. forensic toxicology is the study of the effect of drugs and poisons on / in the human body. forensic video analysis is the scientific examination, comparison and evaluation of video in legal matters. mobile device forensics is the scientific examination and evaluation of evidence found in mobile phones, e. g. call history and deleted sms, and includes sim card forensics
Question: Scientists recorded the hourly temperature at a weather station for the month of July and want to quickly measure a trend over time in temperature changes. Which of these formats would be the most appropriate representation of the temperature data to quickly measure any trend?
A) bar graph
B) line graph
C) pie chart
D) data table
|
B) line graph
|
Context:
the third millennium bc in palmela, portugal, los millares, spain, and stonehenge, united kingdom. the precise beginnings, however, have not be clearly ascertained and new discoveries are both continuous and ongoing. in approximately 1900 bc, ancient iron smelting sites existed in tamil nadu. in the near east, about 3, 500 bc, it was discovered that by combining copper and tin, a superior metal could be made, an alloy called bronze. this represented a major technological shift known as the bronze age. the extraction of iron from its ore into a workable metal is much more difficult than for copper or tin. the process appears to have been invented by the hittites in about 1200 bc, beginning the iron age. the secret of extracting and working iron was a key factor in the success of the philistines. historical developments in ferrous metallurgy can be found in a wide variety of past cultures and civilizations. this includes the ancient and medieval kingdoms and empires of the middle east and near east, ancient iran, ancient egypt, ancient nubia, and anatolia in present - day turkey, ancient nok, carthage, the celts, greeks and romans of ancient europe, medieval europe, ancient and medieval china, ancient and medieval india, ancient and medieval japan, amongst others. a 16th century book by georg agricola, de re metallica, describes the highly developed and complex processes of mining metal ores, metal extraction, and metallurgy of the time. agricola has been described as the " father of metallurgy ". = = extraction = = extractive metallurgy is the practice of removing valuable metals from an ore and refining the extracted raw metals into a purer form. in order to convert a metal oxide or sulphide to a purer metal, the ore must be reduced physically, chemically, or electrolytically. extractive metallurgists are interested in three primary streams : feed, concentrate ( metal oxide / sulphide ) and tailings ( waste ). after mining, large pieces of the ore feed are broken through crushing or grinding in order to obtain particles small enough, where each particle is either mostly valuable or mostly waste. concentrating the particles of value in a form supporting separation enables the desired metal to be removed from waste products. mining may not be necessary, if the ore body and physical environment are conducive to leaching. leaching dissolves minerals in an ore body and results in an enriched solution. the solution
process of lactic acid fermentation, which produced other preserved foods, such as soy sauce. fermentation was also used in this time period to produce leavened bread. although the process of fermentation was not fully understood until louis pasteur ' s work in 1857, it is still the first use of biotechnology to convert a food source into another form. before the time of charles darwin ' s work and life, animal and plant scientists had already used selective breeding. darwin added to that body of work with his scientific observations about the ability of science to change species. these accounts contributed to darwin ' s theory of natural selection. for thousands of years, humans have used selective breeding to improve the production of crops and livestock to use them for food. in selective breeding, organisms with desirable characteristics are mated to produce offspring with the same characteristics. for example, this technique was used with corn to produce the largest and sweetest crops. in the early twentieth century scientists gained a greater understanding of microbiology and explored ways of manufacturing specific products. in 1917, chaim weizmann first used a pure microbiological culture in an industrial process, that of manufacturing corn starch using clostridium acetobutylicum, to produce acetone, which the united kingdom desperately needed to manufacture explosives during world war i. biotechnology has also led to the development of antibiotics. in 1928, alexander fleming discovered the mold penicillium. his work led to the purification of the antibiotic formed by the mold by howard florey, ernst boris chain and norman heatley β to form what we today know as penicillin. in 1940, penicillin became available for medicinal use to treat bacterial infections in humans. the field of modern biotechnology is generally thought of as having been born in 1971 when paul berg ' s ( stanford ) experiments in gene splicing had early success. herbert w. boyer ( univ. calif. at san francisco ) and stanley n. cohen ( stanford ) significantly advanced the new technology in 1972 by transferring genetic material into a bacterium, such that the imported material would be reproduced. the commercial viability of a biotechnology industry was significantly expanded on june 16, 1980, when the united states supreme court ruled that a genetically modified microorganism could be patented in the case of diamond v. chakrabarty. indian - born ananda chakrabarty, working for general electric, had modified a bacterium ( of the genus pseudomonas ) capable of breaking down crude oil, which he proposed to
three major planets, venus, earth, and mercury formed out of the solar nebula. a fourth planetesimal, theia, also formed near earth where it collided in a giant impact, rebounding as the planet mars. during this impact earth lost $ { \ approx } 4 $ \ % of its crust and mantle that is now is found on mars and the moon. at the antipode of the giant impact, $ \ approx $ 60 \ % of earth ' s crust, atmosphere, and a large amount of mantle were ejected into space forming the moon. the lost crust never reformed and became the earth ' s ocean basins. the theia impact site corresponds to indian ocean gravitational anomaly on earth and the hellas basin on mars. the dynamics of the giant impact are consistent with the rotational rates and axial tilts of both earth and mars. the giant impact removed sufficient co $ _ 2 $ from earth ' s atmosphere to avoid a runaway greenhouse effect, initiated plate tectonics, and gave life time to form near geothermal vents at the continental margins. mercury formed near venus where on a close approach it was slingshot into the sun ' s convective zone losing 94 \ % of its mass, much of which remains there today. black carbon, from co $ _ 2 $ decomposed by the intense heat, is still found on the surface of mercury. arriving at 616 km / s, mercury dramatically altered the sun ' s rotational energy, explaining both its anomalously slow rotation rate and axial tilt. these results are quantitatively supported by mass balances, the current locations of the terrestrial planets, and the orientations of their major orbital axes.
onset of electro - chemical corrosion. similar problems are encountered in coastal and offshore structures. = = = anti - fouling = = = anti - fouling is the process of eliminating obstructive organisms from essential components of seawater systems. depending on the nature and location of marine growth, this process is performed in a number of different ways : marine organisms may grow and attach to the surfaces of the outboard suction inlets used to obtain water for cooling systems. electro - chlorination involves running high electrical current through sea water, altering the water ' s chemical composition to create sodium hypochlorite, purging any bio - matter. an electrolytic method of anti - fouling involves running electrical current through two anodes ( scardino, 2009 ). these anodes typically consist of copper and aluminum ( or alternatively, iron ). the first metal, copper anode, releases its ion into the water, creating an environment that is too toxic for bio - matter. the second metal, aluminum, coats the inside of the pipes to prevent corrosion. other forms of marine growth such as mussels and algae may attach themselves to the bottom of a ship ' s hull. this growth interferes with the smoothness and uniformity of the ship ' s hull, causing the ship to have a less hydrodynamic shape that causes it to be slower and less fuel - efficient. marine growth on the hull can be remedied by using special paint that prevents the growth of such organisms. = = = pollution control = = = = = = = sulfur emission = = = = the burning of marine fuels releases harmful pollutants into the atmosphere. ships burn marine diesel in addition to heavy fuel oil. heavy fuel oil, being the heaviest of refined oils, releases sulfur dioxide when burned. sulfur dioxide emissions have the potential to raise atmospheric and ocean acidity causing harm to marine life. however, heavy fuel oil may only be burned in international waters due to the pollution created. it is commercially advantageous due to the cost effectiveness compared to other marine fuels. it is prospected that heavy fuel oil will be phased out of commercial use by the year 2020 ( smith, 2018 ). = = = = oil and water discharge = = = = water, oil, and other substances collect at the bottom of the ship in what is known as the bilge. bilge water is pumped overboard, but must pass a pollution threshold test of 15 ppm ( parts per million ) of oil to be discharged. water is tested
##lating and copying the genetic material of interest using recombinant dna methods or by artificially synthesising the dna. a construct is usually created and used to insert this dna into the host organism. the first recombinant dna molecule was made by paul berg in 1972 by combining dna from the monkey virus sv40 with the lambda virus. as well as inserting genes, the process can be used to remove, or " knock out ", genes. the new dna can be inserted randomly, or targeted to a specific part of the genome. an organism that is generated through genetic engineering is considered to be genetically modified ( gm ) and the resulting entity is a genetically modified organism ( gmo ). the first gmo was a bacterium generated by herbert boyer and stanley cohen in 1973. rudolf jaenisch created the first gm animal when he inserted foreign dna into a mouse in 1974. the first company to focus on genetic engineering, genentech, was founded in 1976 and started the production of human proteins. genetically engineered human insulin was produced in 1978 and insulin - producing bacteria were commercialised in 1982. genetically modified food has been sold since 1994, with the release of the flavr savr tomato. the flavr savr was engineered to have a longer shelf life, but most current gm crops are modified to increase resistance to insects and herbicides. glofish, the first gmo designed as a pet, was sold in the united states in december 2003. in 2016 salmon modified with a growth hormone were sold. genetic engineering has been applied in numerous fields including research, medicine, industrial biotechnology and agriculture. in research, gmos are used to study gene function and expression through loss of function, gain of function, tracking and expression experiments. by knocking out genes responsible for certain conditions it is possible to create animal model organisms of human diseases. as well as producing hormones, vaccines and other drugs, genetic engineering has the potential to cure genetic diseases through gene therapy. chinese hamster ovary ( cho ) cells are used in industrial genetic engineering. additionally mrna vaccines are made through genetic engineering to prevent infections by viruses such as covid - 19. the same techniques that are used to produce drugs can also have industrial applications such as producing enzymes for laundry detergent, cheeses and other products. the rise of commercialised genetically modified crops has provided economic benefit to farmers in many different countries, but has also been the source of most of the controversy surrounding the technology. this has been present since its early use ; the first field trials were destroyed by
it is believed that there may have been a large number of black holes formed in the very early universe. these would have quantised masses. a charged ` ` elementary black hole ' ' ( with the minimum possible mass ) can capture electrons, protons and other charged particles to form a ` ` black hole atom ' '. we find the spectrum of such an object with a view to laboratory and astronomical observation of them, and estimate the lifetime of the bound states. there is no limit to the charge of the black hole, which gives us the possibility of observing z > 137 bound states and transitions at the lower continuum. negatively charged black holes can capture protons. for z > 1, the orbiting protons will coalesce to form a nucleus ( after beta - decay of some protons to neutrons ), with a stability curve different to that of free nuclei. in this system there is also the distinct possibility of single quark capture. this leads to the formation of a coloured black hole that plays the role of an extremely heavy quark interacting strongly with the other two quarks. finally we consider atoms formed with much larger black holes.
with the lambda virus. as well as inserting genes, the process can be used to remove, or " knock out ", genes. the new dna can be inserted randomly, or targeted to a specific part of the genome. an organism that is generated through genetic engineering is considered to be genetically modified ( gm ) and the resulting entity is a genetically modified organism ( gmo ). the first gmo was a bacterium generated by herbert boyer and stanley cohen in 1973. rudolf jaenisch created the first gm animal when he inserted foreign dna into a mouse in 1974. the first company to focus on genetic engineering, genentech, was founded in 1976 and started the production of human proteins. genetically engineered human insulin was produced in 1978 and insulin - producing bacteria were commercialised in 1982. genetically modified food has been sold since 1994, with the release of the flavr savr tomato. the flavr savr was engineered to have a longer shelf life, but most current gm crops are modified to increase resistance to insects and herbicides. glofish, the first gmo designed as a pet, was sold in the united states in december 2003. in 2016 salmon modified with a growth hormone were sold. genetic engineering has been applied in numerous fields including research, medicine, industrial biotechnology and agriculture. in research, gmos are used to study gene function and expression through loss of function, gain of function, tracking and expression experiments. by knocking out genes responsible for certain conditions it is possible to create animal model organisms of human diseases. as well as producing hormones, vaccines and other drugs, genetic engineering has the potential to cure genetic diseases through gene therapy. chinese hamster ovary ( cho ) cells are used in industrial genetic engineering. additionally mrna vaccines are made through genetic engineering to prevent infections by viruses such as covid - 19. the same techniques that are used to produce drugs can also have industrial applications such as producing enzymes for laundry detergent, cheeses and other products. the rise of commercialised genetically modified crops has provided economic benefit to farmers in many different countries, but has also been the source of most of the controversy surrounding the technology. this has been present since its early use ; the first field trials were destroyed by anti - gm activists. although there is a scientific consensus that currently available food derived from gm crops poses no greater risk to human health than conventional food, critics consider gm food safety a leading concern. gene flow, impact on non - target organisms, control of the food supply and intellectual property rights have also been raised as potential
##rozoic eon that began 539 million years ago being subdivided into paleozoic, mesozoic, and cenozoic eras. these three eras together comprise eleven periods ( cambrian, ordovician, silurian, devonian, carboniferous, permian, triassic, jurassic, cretaceous, tertiary, and quaternary ). the similarities among all known present - day species indicate that they have diverged through the process of evolution from their common ancestor. biologists regard the ubiquity of the genetic code as evidence of universal common descent for all bacteria, archaea, and eukaryotes. microbial mats of coexisting bacteria and archaea were the dominant form of life in the early archean eon and many of the major steps in early evolution are thought to have taken place in this environment. the earliest evidence of eukaryotes dates from 1. 85 billion years ago, and while they may have been present earlier, their diversification accelerated when they started using oxygen in their metabolism. later, around 1. 7 billion years ago, multicellular organisms began to appear, with differentiated cells performing specialised functions. algae - like multicellular land plants are dated back to about 1 billion years ago, although evidence suggests that microorganisms formed the earliest terrestrial ecosystems, at least 2. 7 billion years ago. microorganisms are thought to have paved the way for the inception of land plants in the ordovician period. land plants were so successful that they are thought to have contributed to the late devonian extinction event. ediacara biota appear during the ediacaran period, while vertebrates, along with most other modern phyla originated about 525 million years ago during the cambrian explosion. during the permian period, synapsids, including the ancestors of mammals, dominated the land, but most of this group became extinct in the permian β triassic extinction event 252 million years ago. during the recovery from this catastrophe, archosaurs became the most abundant land vertebrates ; one archosaur group, the dinosaurs, dominated the jurassic and cretaceous periods. after the cretaceous β paleogene extinction event 66 million years ago killed off the non - avian dinosaurs, mammals increased rapidly in size and diversity. such mass extinctions may have accelerated evolution by providing opportunities for new groups of organisms to diversify. = = diversity = = = = = bacteria and archaea = = = bacteria are a type of cell that constitute a large domain of prokar
the resulting entity is a genetically modified organism ( gmo ). the first gmo was a bacterium generated by herbert boyer and stanley cohen in 1973. rudolf jaenisch created the first gm animal when he inserted foreign dna into a mouse in 1974. the first company to focus on genetic engineering, genentech, was founded in 1976 and started the production of human proteins. genetically engineered human insulin was produced in 1978 and insulin - producing bacteria were commercialised in 1982. genetically modified food has been sold since 1994, with the release of the flavr savr tomato. the flavr savr was engineered to have a longer shelf life, but most current gm crops are modified to increase resistance to insects and herbicides. glofish, the first gmo designed as a pet, was sold in the united states in december 2003. in 2016 salmon modified with a growth hormone were sold. genetic engineering has been applied in numerous fields including research, medicine, industrial biotechnology and agriculture. in research, gmos are used to study gene function and expression through loss of function, gain of function, tracking and expression experiments. by knocking out genes responsible for certain conditions it is possible to create animal model organisms of human diseases. as well as producing hormones, vaccines and other drugs, genetic engineering has the potential to cure genetic diseases through gene therapy. chinese hamster ovary ( cho ) cells are used in industrial genetic engineering. additionally mrna vaccines are made through genetic engineering to prevent infections by viruses such as covid - 19. the same techniques that are used to produce drugs can also have industrial applications such as producing enzymes for laundry detergent, cheeses and other products. the rise of commercialised genetically modified crops has provided economic benefit to farmers in many different countries, but has also been the source of most of the controversy surrounding the technology. this has been present since its early use ; the first field trials were destroyed by anti - gm activists. although there is a scientific consensus that currently available food derived from gm crops poses no greater risk to human health than conventional food, critics consider gm food safety a leading concern. gene flow, impact on non - target organisms, control of the food supply and intellectual property rights have also been raised as potential issues. these concerns have led to the development of a regulatory framework, which started in 1975. it has led to an international treaty, the cartagena protocol on biosafety, that was adopted in 2000. individual countries have developed their own regulatory systems regarding gmos, with the most marked differences occurring between the
biotechnology to convert a food source into another form. before the time of charles darwin ' s work and life, animal and plant scientists had already used selective breeding. darwin added to that body of work with his scientific observations about the ability of science to change species. these accounts contributed to darwin ' s theory of natural selection. for thousands of years, humans have used selective breeding to improve the production of crops and livestock to use them for food. in selective breeding, organisms with desirable characteristics are mated to produce offspring with the same characteristics. for example, this technique was used with corn to produce the largest and sweetest crops. in the early twentieth century scientists gained a greater understanding of microbiology and explored ways of manufacturing specific products. in 1917, chaim weizmann first used a pure microbiological culture in an industrial process, that of manufacturing corn starch using clostridium acetobutylicum, to produce acetone, which the united kingdom desperately needed to manufacture explosives during world war i. biotechnology has also led to the development of antibiotics. in 1928, alexander fleming discovered the mold penicillium. his work led to the purification of the antibiotic formed by the mold by howard florey, ernst boris chain and norman heatley β to form what we today know as penicillin. in 1940, penicillin became available for medicinal use to treat bacterial infections in humans. the field of modern biotechnology is generally thought of as having been born in 1971 when paul berg ' s ( stanford ) experiments in gene splicing had early success. herbert w. boyer ( univ. calif. at san francisco ) and stanley n. cohen ( stanford ) significantly advanced the new technology in 1972 by transferring genetic material into a bacterium, such that the imported material would be reproduced. the commercial viability of a biotechnology industry was significantly expanded on june 16, 1980, when the united states supreme court ruled that a genetically modified microorganism could be patented in the case of diamond v. chakrabarty. indian - born ananda chakrabarty, working for general electric, had modified a bacterium ( of the genus pseudomonas ) capable of breaking down crude oil, which he proposed to use in treating oil spills. ( chakrabarty ' s work did not involve gene manipulation but rather the transfer of entire organelles between strains of the pseudomonas bacterium ). the mosfet invented at bell labs between 1955 and 1960, two years later, leland c. clark and champ lyons invented the
Question: A factory in Japan had been dumping mercury into Minamata Bay. Certain microorganisms in the water were able to change the mercury into a form that the plankton could use. Which most likely happened next?
A) The organisms built up a resistance to the mercury.
B) The mercury passed through the food chain.
C) The plankton made the mercury safe.
D) The fish would not eat the plankton.
|
B) The mercury passed through the food chain.
|
Context:
generation times. corn has been used to study mechanisms of photosynthesis and phloem loading of sugar in c4 plants. the single celled green alga chlamydomonas reinhardtii, while not an embryophyte itself, contains a green - pigmented chloroplast related to that of land plants, making it useful for study. a red alga cyanidioschyzon merolae has also been used to study some basic chloroplast functions. spinach, peas, soybeans and a moss physcomitrella patens are commonly used to study plant cell biology. agrobacterium tumefaciens, a soil rhizosphere bacterium, can attach to plant cells and infect them with a callus - inducing ti plasmid by horizontal gene transfer, causing a callus infection called crown gall disease. schell and van montagu ( 1977 ) hypothesised that the ti plasmid could be a natural vector for introducing the nif gene responsible for nitrogen fixation in the root nodules of legumes and other plant species. today, genetic modification of the ti plasmid is one of the main techniques for introduction of transgenes to plants and the creation of genetically modified crops. = = = epigenetics = = = epigenetics is the study of heritable changes in gene function that cannot be explained by changes in the underlying dna sequence but cause the organism ' s genes to behave ( or " express themselves " ) differently. one example of epigenetic change is the marking of the genes by dna methylation which determines whether they will be expressed or not. gene expression can also be controlled by repressor proteins that attach to silencer regions of the dna and prevent that region of the dna code from being expressed. epigenetic marks may be added or removed from the dna during programmed stages of development of the plant, and are responsible, for example, for the differences between anthers, petals and normal leaves, despite the fact that they all have the same underlying genetic code. epigenetic changes may be temporary or may remain through successive cell divisions for the remainder of the cell ' s life. some epigenetic changes have been shown to be heritable, while others are reset in the germ cells. epigenetic changes in eukaryotic biology serve to regulate the process of cellular differentiation. during morphogenesis, totipotent stem cells become the various
##fer the validity of a general statement from a number of specific instances or infer the truth of a theory from a series of successful tests. for example, a chicken observes that each morning the farmer comes and gives it food, for hundreds of days in a row. the chicken may therefore use inductive reasoning to infer that the farmer will bring food every morning. however, one morning, the farmer comes and kills the chicken. how is scientific reasoning more trustworthy than the chicken ' s reasoning? one approach is to acknowledge that induction cannot achieve certainty, but observing more instances of a general statement can at least make the general statement more probable. so the chicken would be right to conclude from all those mornings that it is likely the farmer will come with food again the next morning, even if it cannot be certain. however, there remain difficult questions about the process of interpreting any given evidence into a probability that the general statement is true. one way out of these particular difficulties is to declare that all beliefs about scientific theories are subjective, or personal, and correct reasoning is merely about how evidence should change one ' s subjective beliefs over time. some argue that what scientists do is not inductive reasoning at all but rather abductive reasoning, or inference to the best explanation. in this account, science is not about generalizing specific instances but rather about hypothesizing explanations for what is observed. as discussed in the previous section, it is not always clear what is meant by the " best explanation ". ockham ' s razor, which counsels choosing the simplest available explanation, thus plays an important role in some versions of this approach. to return to the example of the chicken, would it be simpler to suppose that the farmer cares about it and will continue taking care of it indefinitely or that the farmer is fattening it up for slaughter? philosophers have tried to make this heuristic principle more precise regarding theoretical parsimony or other measures. yet, although various measures of simplicity have been brought forward as potential candidates, it is generally accepted that there is no such thing as a theory - independent measure of simplicity. in other words, there appear to be as many different measures of simplicity as there are theories themselves, and the task of choosing between measures of simplicity appears to be every bit as problematic as the job of choosing between theories. nicholas maxwell has argued for some decades that unity rather than simplicity is the key non - empirical factor in influencing the choice of theory in science, persistent preference for unified theories in effect committing science to the
on earth in suitable amounts. one isotope of uranium, namely uranium - 235, is naturally occurring and sufficiently unstable, but it is always found mixed with the more stable isotope uranium - 238. the latter accounts for more than 99 % of the weight of natural uranium. therefore, some method of isotope separation based on the weight of three neutrons must be performed to enrich ( isolate ) uranium - 235. alternatively, the element plutonium possesses an isotope that is sufficiently unstable for this process to be usable. terrestrial plutonium does not currently occur naturally in sufficient quantities for such use, so it must be manufactured in a nuclear reactor. ultimately, the manhattan project manufactured nuclear weapons based on each of these elements. they detonated the first nuclear weapon in a test code - named " trinity ", near alamogordo, new mexico, on july 16, 1945. the test was conducted to ensure that the implosion method of detonation would work, which it did. a uranium bomb, little boy, was dropped on the japanese city hiroshima on august 6, 1945, followed three days later by the plutonium - based fat man on nagasaki. in the wake of unprecedented devastation and casualties from a single weapon, the japanese government soon surrendered, ending world war ii. since these bombings, no nuclear weapons have been deployed offensively. nevertheless, they prompted an arms race to develop increasingly destructive bombs to provide a nuclear deterrent. just over four years later, on august 29, 1949, the soviet union detonated its first fission weapon. the united kingdom followed on october 2, 1952 ; france, on february 13, 1960 ; and china component to a nuclear weapon. approximately half of the deaths from hiroshima and nagasaki died two to five years afterward from radiation exposure. a radiological weapon is a type of nuclear weapon designed to distribute hazardous nuclear material in enemy areas. such a weapon would not have the explosive capability of a fission or fusion bomb, but would kill many people and contaminate a large area. a radiological weapon has never been deployed. while considered useless by a conventional military, such a weapon raises concerns over nuclear terrorism. there have been over 2, 000 nuclear tests conducted since 1945. in 1963, all nuclear and many non - nuclear states signed the limited test ban treaty, pledging to refrain from testing nuclear weapons in the atmosphere, underwater, or in outer space. the treaty permitted underground nuclear testing. france continued atmospheric testing until 1974, while china continued up until 1980. the last underground test by the united states was in 1992, the soviet union
of three neutrons must be performed to enrich ( isolate ) uranium - 235. alternatively, the element plutonium possesses an isotope that is sufficiently unstable for this process to be usable. terrestrial plutonium does not currently occur naturally in sufficient quantities for such use, so it must be manufactured in a nuclear reactor. ultimately, the manhattan project manufactured nuclear weapons based on each of these elements. they detonated the first nuclear weapon in a test code - named " trinity ", near alamogordo, new mexico, on july 16, 1945. the test was conducted to ensure that the implosion method of detonation would work, which it did. a uranium bomb, little boy, was dropped on the japanese city hiroshima on august 6, 1945, followed three days later by the plutonium - based fat man on nagasaki. in the wake of unprecedented devastation and casualties from a single weapon, the japanese government soon surrendered, ending world war ii. since these bombings, no nuclear weapons have been deployed offensively. nevertheless, they prompted an arms race to develop increasingly destructive bombs to provide a nuclear deterrent. just over four years later, on august 29, 1949, the soviet union detonated its first fission weapon. the united kingdom followed on october 2, 1952 ; france, on february 13, 1960 ; and china component to a nuclear weapon. approximately half of the deaths from hiroshima and nagasaki died two to five years afterward from radiation exposure. a radiological weapon is a type of nuclear weapon designed to distribute hazardous nuclear material in enemy areas. such a weapon would not have the explosive capability of a fission or fusion bomb, but would kill many people and contaminate a large area. a radiological weapon has never been deployed. while considered useless by a conventional military, such a weapon raises concerns over nuclear terrorism. there have been over 2, 000 nuclear tests conducted since 1945. in 1963, all nuclear and many non - nuclear states signed the limited test ban treaty, pledging to refrain from testing nuclear weapons in the atmosphere, underwater, or in outer space. the treaty permitted underground nuclear testing. france continued atmospheric testing until 1974, while china continued up until 1980. the last underground test by the united states was in 1992, the soviet union in 1990, the united kingdom in 1991, and both france and china continued testing until 1996. after signing the comprehensive test ban treaty in 1996 ( which had as of 2011 not entered into force ), all of these states have pledged to discontinue all nuclear testing. non - signatories india and pakistan last
the qcd instanton can be observed at relatively low energies at nica collider by studying the spin - spin correlations between the incoming proton and the produced hyperons.
; however, a successful large - scale industrial application of the process was the development of continuous freeze drying of coffee. high - temperature short time processing β these processes, for the most part, are characterized by rapid heating and cooling, holding for a short time at a relatively high temperature and filling aseptically into sterile containers. decaffeination of coffee and tea β decaffeinated coffee and tea was first developed on a commercial basis in europe around 1900. the process is described in u. s. patent 897, 763. green coffee beans are treated with water, heat and solvents to remove the caffeine from the beans. process optimization β food technology now allows production of foods to be more efficient, oil saving technologies are now available on different forms. production methods and methodology have also become increasingly sophisticated. aseptic packaging β the process of filling a commercially sterile product into a sterile container and hermetically sealing the containers so that re - infection is prevented. thus, this results into a shelf stable product at ambient conditions. food irradiation β the process of exposing food and food packaging to ionizing radiation can effectively destroy organisms responsible for spoilage and foodborne illness and inhibit sprouting, extending shelf life. commercial fruit ripening rooms using ethylene as a plant hormone. food delivery β an order is typically made either through a restaurant or grocer ' s website or mobile app, or through a food ordering company. the ordered food is typically delivered in boxes or bags to the customer ' s doorsteps. = = categories = = technology has innovated these categories from the food industry : agricultural technology β or agtech, it is the use of technology in agriculture, horticulture, and aquaculture with the aim of improving yield, efficiency, and profitability. agricultural technology can be products, services or applications derived from agriculture that improve various input / output processes. food science β technology in this sector focuses on the development of new functional ingredients and alternative proteins. foodservice β technology innovated the way establishments prepare, supply, and serve food outside the home. there ' s a tendency to create the conditions for the restaurant of the future with robotics and cloudkitchens. consumer tech β technology allows what we call consumer electronics, which is the equipment of consumers with devices that facilitates the cooking process. food delivery β as the food delivery market is growing, companies and startups are rapidly revolutionizing the communication process between consumers and food establishments, with platform - to - consumer delivery as the
of beliefs. an observation of a transit of venus requires a huge range of auxiliary beliefs, such as those that describe the optics of telescopes, the mechanics of the telescope mount, and an understanding of celestial mechanics. if the prediction fails and a transit is not observed, that is likely to occasion an adjustment in the system, a change in some auxiliary assumption, rather than a rejection of the theoretical system. according to the duhem β quine thesis, after pierre duhem and w. v. quine, it is impossible to test a theory in isolation. one must always add auxiliary hypotheses in order to make testable predictions. for example, to test newton ' s law of gravitation in the solar system, one needs information about the masses and positions of the sun and all the planets. famously, the failure to predict the orbit of uranus in the 19th century led not to the rejection of newton ' s law but rather to the rejection of the hypothesis that the solar system comprises only seven planets. the investigations that followed led to the discovery of an eighth planet, neptune. if a test fails, something is wrong. but there is a problem in figuring out what that something is : a missing planet, badly calibrated test equipment, an unsuspected curvature of space, or something else. one consequence of the duhem β quine thesis is that one can make any theory compatible with any empirical observation by the addition of a sufficient number of suitable ad hoc hypotheses. karl popper accepted this thesis, leading him to reject naive falsification. instead, he favored a " survival of the fittest " view in which the most falsifiable scientific theories are to be preferred. = = = anything goes methodology = = = paul feyerabend ( 1924 β 1994 ) argued that no description of scientific method could possibly be broad enough to include all the approaches and methods used by scientists, and that there are no useful and exception - free methodological rules governing the progress of science. he argued that " the only principle that does not inhibit progress is : anything goes ". feyerabend said that science started as a liberating movement, but that over time it had become increasingly dogmatic and rigid and had some oppressive features, and thus had become increasingly an ideology. because of this, he said it was impossible to come up with an unambiguous way to distinguish science from religion, magic, or mythology. he saw the exclusive dominance of science as a means of directing society as
was used before copper smelting was known. copper smelting is believed to have originated when the technology of pottery kilns allowed sufficiently high temperatures. the concentration of various elements such as arsenic increase with depth in copper ore deposits and smelting of these ores yields arsenical bronze, which can be sufficiently work hardened to be suitable for making tools. bronze is an alloy of copper with tin ; the latter being found in relatively few deposits globally caused a long time to elapse before true tin bronze became widespread. ( see : tin sources and trade in ancient times ) bronze was a major advancement over stone as a material for making tools, both because of its mechanical properties like strength and ductility and because it could be cast in molds to make intricately shaped objects. bronze significantly advanced shipbuilding technology with better tools and bronze nails. bronze nails replaced the old method of attaching boards of the hull with cord woven through drilled holes. better ships enabled long - distance trade and the advance of civilization. this technological trend apparently began in the fertile crescent and spread outward over time. these developments were not, and still are not, universal. the three - age system does not accurately describe the technology history of groups outside of eurasia, and does not apply at all in the case of some isolated populations, such as the spinifex people, the sentinelese, and various amazonian tribes, which still make use of stone age technology, and have not developed agricultural or metal technology. these villages preserve traditional customs in the face of global modernity, exhibiting a remarkable resistance to the rapid advancement of technology. = = = = iron age = = = = before iron smelting was developed the only iron was obtained from meteorites and is usually identified by having nickel content. meteoric iron was rare and valuable, but was sometimes used to make tools and other implements, such as fish hooks. the iron age involved the adoption of iron smelting technology. it generally replaced bronze and made it possible to produce tools which were stronger, lighter and cheaper to make than bronze equivalents. the raw materials to make iron, such as ore and limestone, are far more abundant than copper and especially tin ores. consequently, iron was produced in many areas. it was not possible to mass manufacture steel or pure iron because of the high temperatures required. furnaces could reach melting temperature but the crucibles and molds needed for melting and casting had not been developed. steel could be produced by forging bloomery iron to reduce the carbon content in a
alarm is triggered. = = = food processing and agriculture = = = in biology and agriculture, radiation is used to induce mutations to produce new or improved species, such as in atomic gardening. another use in insect control is the sterile insect technique, where male insects are sterilized by radiation and released, so they have no offspring, to reduce the population. in industrial and food applications, radiation is used for sterilization of tools and equipment. an advantage is that the object may be sealed in plastic before sterilization. an emerging use in food production is the sterilization of food using food irradiation. food irradiation is the process of exposing food to ionizing radiation in order to destroy microorganisms, bacteria, viruses, or insects that might be present in the food. the radiation sources used include radioisotope gamma ray sources, x - ray generators and electron accelerators. further applications include sprout inhibition, delay of ripening, increase of juice yield, and improvement of re - hydration. irradiation is a more general term of deliberate exposure of materials to radiation to achieve a technical goal ( in this context ' ionizing radiation ' is implied ). as such it is also used on non - food items, such as medical hardware, plastics, tubes for gas - pipelines, hoses for floor - heating, shrink - foils for food packaging, automobile parts, wires and cables ( isolation ), tires, and even gemstones. compared to the amount of food irradiated, the volume of those every - day applications is huge but not noticed by the consumer. the genuine effect of processing food by ionizing radiation relates to damages to the dna, the basic genetic information for life. microorganisms can no longer proliferate and continue their malignant or pathogenic activities. spoilage causing micro - organisms cannot continue their activities. insects do not survive or become incapable of procreation. plants cannot continue the natural ripening or aging process. all these effects are beneficial to the consumer and the food industry, likewise. the amount of energy imparted for effective food irradiation is low compared to cooking the same ; even at a typical dose of 10 kgy most food, which is ( with regard to warming ) physically equivalent to water, would warm by only about 2. 5 Β°c ( 4. 5 Β°f ). the specialty of processing food by ionizing radiation is the fact, that the energy density per atomic transition is very high, it can cleave
in this talk, i will explain how to reduce the spectral index to be n _ s = 0. 96 for supernatural inflation. i will also show the constraint to the reheating temperature from big bang nucleosynthesis of both thermal and non - thermal gravitino production.
Question: Students conducted an investigation with green bananas. They hypothesized that an onion placed near a banana could increase the rate at which the banana ripened. They placed one onion and one banana in each of three containers. In a fourth container they placed a banana, but no onion. Which statement best explains why the onion was left out of the fourth container?
A) The onion would contaminate the fourth trial.
B) The container without the onion is the control.
C) Three containers with onions are enough to conduct the test.
D) The container without the onion is the independent variable.
|
B) The container without the onion is the control.
|
Context:
weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under ordinary conditions, most of the materials brought down from the high lands by torrential water courses are carried forward by the main river to the sea, or partially strewn over flat alluvial plains during floods ; the size of the materials forming the bed of the river or borne along by the stream is gradually reduced on proceeding seawards, so that in the po river in italy, for instance, pebbles and gravel are found for about 140 miles below turin, sand along the next 100 miles, and silt and mud in the last 110 miles ( 176 km ). = = channelization = = the removal of obstructions, natural or artificial
light and cold extrasolar planets such as ogle 2005 - blg - 390lb, a 5. 5 earth - mass planet detected via microlensing, could be frequent in the galaxy according to some preliminary results from microlensing experiments. these planets can be frozen rocky - or ocean - planets, situated beyond the snow line and, therefore, beyond the habitable zone of their system. they can nonetheless host a layer of liquid water, heated by radiogenic energy, underneath an ice shell surface for billions of years, before freezing completely. these results suggest that oceans under ice, like those suspected to be present on icy moons in the solar system, could be a common feature of cold low - mass extrasolar planets.
navigable capabilities of rivers. in tropical countries subject to periodical rains, the rivers are in flood during the rainy season and have hardly any flow during the rest of the year, while in temperate regions, where the rainfall is more evenly distributed throughout the year, evaporation causes the available rainfall to be much less in hot summer weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under ordinary conditions, most of the materials brought down from the high lands by torrential water courses are carried forward by the main river to the sea, or partially strewn over flat alluvial plains during floods ; the size of the materials forming the bed of the river or borne along by the stream is gradually reduced on proceeding sea
general modes : static failure, and fatigue failure. static structural failure occurs when, upon being loaded ( having a force applied ) the object being analyzed either breaks or is deformed plastically, depending on the criterion for failure. fatigue failure occurs when an object fails after a number of repeated loading and unloading cycles. fatigue failure occurs because of imperfections in the object : a microscopic crack on the surface of the object, for instance, will grow slightly with each cycle ( propagation ) until the crack is large enough to cause ultimate failure. failure is not simply defined as when a part breaks, however ; it is defined as when a part does not operate as intended. some systems, such as the perforated top sections of some plastic bags, are designed to break. if these systems do not break, failure analysis might be employed to determine the cause. structural analysis is often used by mechanical engineers after a failure has occurred, or when designing to prevent failure. engineers often use online documents and books such as those published by asm to aid them in determining the type of failure and possible causes. once theory is applied to a mechanical design, physical testing is often performed to verify calculated results. structural analysis may be used in an office when designing parts, in the field to analyze failed parts, or in laboratories where parts might undergo controlled failure tests. = = = thermodynamics and thermo - science = = = thermodynamics is an applied science used in several branches of engineering, including mechanical and chemical engineering. at its simplest, thermodynamics is the study of energy, its use and transformation through a system. typically, engineering thermodynamics is concerned with changing energy from one form to another. as an example, automotive engines convert chemical energy ( enthalpy ) from the fuel into heat, and then into mechanical work that eventually turns the wheels. thermodynamics principles are used by mechanical engineers in the fields of heat transfer, thermofluids, and energy conversion. mechanical engineers use thermo - science to design engines and power plants, heating, ventilation, and air - conditioning ( hvac ) systems, heat exchangers, heat sinks, radiators, refrigeration, insulation, and others. = = = design and drafting = = = drafting or technical drawing is the means by which mechanical engineers design products and create instructions for manufacturing parts. a technical drawing can be a computer model or hand - drawn schematic showing all the dimensions necessary to manufacture a
are the cryosphere ( corresponding to ice ) as a distinct portion of the hydrosphere and the pedosphere ( corresponding to soil ) as an active and intermixed sphere. the following fields of science are generally categorized within the earth sciences : geology describes the rocky parts of the earth ' s crust ( or lithosphere ) and its historic development. major subdisciplines are mineralogy and petrology, geomorphology, paleontology, stratigraphy, structural geology, engineering geology, and sedimentology. physical geography focuses on geography as an earth science. physical geography is the study of earth ' s seasons, climate, atmosphere, soil, streams, landforms, and oceans. physical geography can be divided into several branches or related fields, as follows : geomorphology, biogeography, environmental geography, palaeogeography, climatology, meteorology, coastal geography, hydrology, ecology, glaciology. geophysics and geodesy investigate the shape of the earth, its reaction to forces and its magnetic and gravity fields. geophysicists explore the earth ' s core and mantle as well as the tectonic and seismic activity of the lithosphere. geophysics is commonly used to supplement the work of geologists in developing a comprehensive understanding of crustal geology, particularly in mineral and petroleum exploration. seismologists use geophysics to understand plate tectonic movement, as well as predict seismic activity. geochemistry studies the processes that control the abundance, composition, and distribution of chemical compounds and isotopes in geologic environments. geochemists use the tools and principles of chemistry to study the earth ' s composition, structure, processes, and other physical aspects. major subdisciplines are aqueous geochemistry, cosmochemistry, isotope geochemistry and biogeochemistry. soil science covers the outermost layer of the earth ' s crust that is subject to soil formation processes ( or pedosphere ). major subdivisions in this field of study include edaphology and pedology. ecology covers the interactions between organisms and their environment. this field of study differentiates the study of earth from other planets in the solar system, earth being the only planet teeming with life. hydrology, oceanography and limnology are studies which focus on the movement, distribution, and quality of the water and involve all the components of the hydrologic cycle on the earth and its atmosphere ( or hydrosphere ). "
##thic, or " old stone age ", and spans all of human history up to the development of agriculture approximately 12, 000 years ago. to make a stone tool, a " core " of hard stone with specific flaking properties ( such as flint ) was struck with a hammerstone. this flaking produced sharp edges which could be used as tools, primarily in the form of choppers or scrapers. these tools greatly aided the early humans in their hunter - gatherer lifestyle to perform a variety of tasks including butchering carcasses ( and breaking bones to get at the marrow ) ; chopping wood ; cracking open nuts ; skinning an animal for its hide, and even forming other tools out of softer materials such as bone and wood. the earliest stone tools were irrelevant, being little more than a fractured rock. in the acheulian era, beginning approximately 1. 65 million years ago, methods of working these stones into specific shapes, such as hand axes emerged. this early stone age is described as the lower paleolithic. the middle paleolithic, approximately 300, 000 years ago, saw the introduction of the prepared - core technique, where multiple blades could be rapidly formed from a single core stone. the upper paleolithic, beginning approximately 40, 000 years ago, saw the introduction of pressure flaking, where a wood, bone, or antler punch could be used to shape a stone very finely. the end of the last ice age about 10, 000 years ago is taken as the end point of the upper paleolithic and the beginning of the epipaleolithic / mesolithic. the mesolithic technology included the use of microliths as composite stone tools, along with wood, bone, and antler tools. the later stone age, during which the rudiments of agricultural technology were developed, is called the neolithic period. during this period, polished stone tools were made from a variety of hard rocks such as flint, jade, jadeite, and greenstone, largely by working exposures as quarries, but later the valuable rocks were pursued by tunneling underground, the first steps in mining technology. the polished axes were used for forest clearance and the establishment of crop farming and were so effective as to remain in use when bronze and iron appeared. these stone axes were used alongside a continued use of stone tools such as a range of projectiles, knives, and scrapers, as well as tools, made from organic materials such as wood, bone, and antler. stone age cultures
significantly greater strength and fracture toughness. another major change in the body during the firing or sintering process will be the establishment of the polycrystalline nature of the solid. significant grain growth tends to occur during sintering, with this growth depending on temperature and duration of the sintering process. the growth of grains will result in some form of grain size distribution, which will have a significant impact on the ultimate physical properties of the material. in particular, abnormal grain growth in which certain grains grow very large in a matrix of finer grains will significantly alter the physical and mechanical properties of the obtained ceramic. in the sintered body, grain sizes are a product of the thermal processing parameters as well as the initial particle size, or possibly the sizes of aggregates or particle clusters which arise during the initial stages of processing. the ultimate microstructure ( and thus the physical properties ) of the final product will be limited by and subject to the form of the structural template or precursor which is created in the initial stages of chemical synthesis and physical forming. hence the importance of chemical powder and polymer processing as it pertains to the synthesis of industrial ceramics, glasses and glass - ceramics. there are numerous possible refinements of the sintering process. some of the most common involve pressing the green body to give the densification a head start and reduce the sintering time needed. sometimes organic binders such as polyvinyl alcohol are added to hold the green body together ; these burn out during the firing ( at 200 β 350 Β°c ). sometimes organic lubricants are added during pressing to increase densification. it is common to combine these, and add binders and lubricants to a powder, then press. ( the formulation of these organic chemical additives is an art in itself. this is particularly important in the manufacture of high performance ceramics such as those used by the billions for electronics, in capacitors, inductors, sensors, etc. ) a slurry can be used in place of a powder, and then cast into a desired shape, dried and then sintered. indeed, traditional pottery is done with this type of method, using a plastic mixture worked with the hands. if a mixture of different materials is used together in a ceramic, the sintering temperature is sometimes above the melting point of one minor component β a liquid phase sintering. this results in shorter sintering times compared to solid state sintering. such liquid phase sintering involves in faster diffusion processes and may result in abnormal grain
floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under ordinary conditions, most of the materials brought down from the high lands by torrential water courses are carried forward by the main river to the sea, or partially strewn over flat alluvial plains during floods ; the size of the materials forming the bed of the river or borne along by the stream is gradually reduced on proceeding seawards, so that in the po river in italy, for instance, pebbles and gravel are found for about 140 miles below turin, sand along the next 100 miles, and silt and mud in the last 110 miles ( 176 km ). = = channelization = = the removal of obstructions, natural or artificial ( e. g., trunks of trees, boulders and accumulations of gravel ) from a river bed furnishes a simple and efficient means of increasing the discharging capacity of its channel. such removals will consequently lower the height of floods upstream. every impediment to the flow, in proportion to its extent, raises the level of the river above it so as to produce the additional artificial fall necessary to convey the flow through the restricted channel, thereby reducing the total available fall. reducing the length of the channel by substituting straight cuts for a winding course is the only way in which the effective fall can be increased.
, crystal structure, hazards associated with minerals, and the physical and chemical properties of minerals. petrology is the study of rocks, including the formation and composition of rocks. petrography is a branch of petrology that studies the typology and classification of rocks. = = earth ' s interior = = plate tectonics, mountain ranges, volcanoes, and earthquakes are geological phenomena that can be explained in terms of physical and chemical processes in the earth ' s crust. beneath the earth ' s crust lies the mantle which is heated by the radioactive decay of heavy elements. the mantle is not quite solid and consists of magma which is in a state of semi - perpetual convection. this convection process causes the lithospheric plates to move, albeit slowly. the resulting process is known as plate tectonics. areas of the crust where new crust is created are called divergent boundaries, those where it is brought back into the earth are convergent boundaries and those where plates slide past each other, but no new lithospheric material is created or destroyed, are referred to as transform ( or conservative ) boundaries. earthquakes result from the movement of the lithospheric plates, and they often occur near convergent boundaries where parts of the crust are forced into the earth as part of subduction. plate tectonics might be thought of as the process by which the earth is resurfaced. as the result of seafloor spreading, new crust and lithosphere is created by the flow of magma from the mantle to the near surface, through fissures, where it cools and solidifies. through subduction, oceanic crust and lithosphere vehemently returns to the convecting mantle. volcanoes result primarily from the melting of subducted crust material. crust material that is forced into the asthenosphere melts, and some portion of the melted material becomes light enough to rise to the surface β giving birth to volcanoes. = = atmospheric science = = atmospheric science initially developed in the late - 19th century as a means to forecast the weather through meteorology, the study of weather. atmospheric chemistry was developed in the 20th century to measure air pollution and expanded in the 1970s in response to acid rain. climatology studies the climate and climate change. the troposphere, stratosphere, mesosphere, thermosphere, and exosphere are the five layers which make up earth ' s atmosphere. 75 % of the mass in the atmosphere is located within the troposphere, the lowest
becomes quite gentle. accordingly, in large basins, rivers in most cases begin as torrents with a variable flow, and end as gently flowing rivers with a comparatively regular discharge. the irregular flow of rivers throughout their course forms one of the main difficulties in devising works for mitigating inundations or for increasing the navigable capabilities of rivers. in tropical countries subject to periodical rains, the rivers are in flood during the rainy season and have hardly any flow during the rest of the year, while in temperate regions, where the rainfall is more evenly distributed throughout the year, evaporation causes the available rainfall to be much less in hot summer weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under
Question: When cold weather freezes water in the cracks of rocks, which would most likely happen?
A) The rocks would become rounded.
B) The rocks would be used for shelter.
C) The rocks would be moved by the wind.
D) The rocks would break into smaller pieces.
|
D) The rocks would break into smaller pieces.
|
Context:
is also higher at high temperature, as shown by carnot ' s theorem. in a conventional metallic engine, much of the energy released from the fuel must be dissipated as waste heat in order to prevent a meltdown of the metallic parts. despite all of these desirable properties, such engines are not in production because the manufacturing of ceramic parts in the requisite precision and durability is difficult. imperfection in the ceramic leads to cracks, which can lead to potentially dangerous equipment failure. such engines are possible in laboratory settings, but mass - production is not feasible with current technology. work is being done in developing ceramic parts for gas turbine engines. currently, even blades made of advanced metal alloys used in the engines ' hot section require cooling and careful limiting of operating temperatures. turbine engines made with ceramics could operate more efficiently, giving aircraft greater range and payload for a set amount of fuel. recently, there have been advances in ceramics which include bio - ceramics, such as dental implants and synthetic bones. hydroxyapatite, the natural mineral component of bone, has been made synthetically from a number of biological and chemical sources and can be formed into ceramic materials. orthopedic implants made from these materials bond readily to bone and other tissues in the body without rejection or inflammatory reactions. because of this, they are of great interest for gene delivery and tissue engineering scaffolds. most hydroxyapatite ceramics are very porous and lack mechanical strength and are used to coat metal orthopedic devices to aid in forming a bond to bone or as bone fillers. they are also used as fillers for orthopedic plastic screws to aid in reducing the inflammation and increase absorption of these plastic materials. work is being done to make strong, fully dense nano crystalline hydroxyapatite ceramic materials for orthopedic weight bearing devices, replacing foreign metal and plastic orthopedic materials with a synthetic, but naturally occurring, bone mineral. ultimately these ceramic materials may be used as bone replacements or with the incorporation of protein collagens, synthetic bones. durable actinide - containing ceramic materials have many applications such as in nuclear fuels for burning excess pu and in chemically - inert sources of alpha irradiation for power supply of unmanned space vehicles or to produce electricity for microelectronic devices. both use and disposal of radioactive actinides require their immobilization in a durable host material. nuclear waste long - lived radionuclides such as actinides are immobilized using chemical
a comparison of the sensitivities of methods which allow us to determine the coordinates of a moving hot body is made.
10 kgy most food, which is ( with regard to warming ) physically equivalent to water, would warm by only about 2. 5 Β°c ( 4. 5 Β°f ). the specialty of processing food by ionizing radiation is the fact, that the energy density per atomic transition is very high, it can cleave molecules and induce ionization ( hence the name ) which cannot be achieved by mere heating. this is the reason for new beneficial effects, however at the same time, for new concerns. the treatment of solid food by ionizing radiation can provide an effect similar to heat pasteurization of liquids, such as milk. however, the use of the term, cold pasteurization, to describe irradiated foods is controversial, because pasteurization and irradiation are fundamentally different processes, although the intended end results can in some cases be similar. detractors of food irradiation have concerns about the health hazards of induced radioactivity. a report for the industry advocacy group american council on science and health entitled " irradiated foods " states : " the types of radiation sources approved for the treatment of foods have specific energy levels well below that which would cause any element in food to become radioactive. food undergoing irradiation does not become any more radioactive than luggage passing through an airport x - ray scanner or teeth that have been x - rayed. " food irradiation is currently permitted by over 40 countries and volumes are estimated to exceed 500, 000 metric tons ( 490, 000 long tons ; 550, 000 short tons ) annually worldwide. food irradiation is essentially a non - nuclear technology ; it relies on the use of ionizing radiation which may be generated by accelerators for electrons and conversion into bremsstrahlung, but which may use also gamma - rays from nuclear decay. there is a worldwide industry for processing by ionizing radiation, the majority by number and by processing power using accelerators. food irradiation is only a niche application compared to medical supplies, plastic materials, raw materials, gemstones, cables and wires, etc. = = accidents = = nuclear accidents, because of the powerful forces involved, are often very dangerous. historically, the first incidents involved fatal radiation exposure. marie curie died from aplastic anemia which resulted from her high levels of exposure. two scientists, an american and canadian respectively, harry daghlian and louis slotin, died after mishandling the same plutonium mass. unlike conventional weapons, the intense light, heat, and explosive force is
assuming only statistical mechanics and general relativity, we calculate the maximal temperature of gas of particles placed in ads space - time. if two particles with a given center of mass energy come close enough, according to classical gravity they will form a black hole. we focus only on the black holes with hawking temperature lower than the environment, because they do not disappear. the number density of such black holes grows with the temperature in the system. at a certain finite temperature, the thermodynamical system will be dominated by black holes. this critical temperature is lower than the planck temperature for the values of the ads vacuum energy density below the planck density. this result might be interesting from the ads / cft correspondence point of view, since it is different from the hawking - page phase transition, and it is not immediately clear what effect dynamically limits the maximal temperature of the thermal state on the cft side of the correspondence.
. currently, even blades made of advanced metal alloys used in the engines ' hot section require cooling and careful limiting of operating temperatures. turbine engines made with ceramics could operate more efficiently, giving aircraft greater range and payload for a set amount of fuel. recently, there have been advances in ceramics which include bio - ceramics, such as dental implants and synthetic bones. hydroxyapatite, the natural mineral component of bone, has been made synthetically from a number of biological and chemical sources and can be formed into ceramic materials. orthopedic implants made from these materials bond readily to bone and other tissues in the body without rejection or inflammatory reactions. because of this, they are of great interest for gene delivery and tissue engineering scaffolds. most hydroxyapatite ceramics are very porous and lack mechanical strength and are used to coat metal orthopedic devices to aid in forming a bond to bone or as bone fillers. they are also used as fillers for orthopedic plastic screws to aid in reducing the inflammation and increase absorption of these plastic materials. work is being done to make strong, fully dense nano crystalline hydroxyapatite ceramic materials for orthopedic weight bearing devices, replacing foreign metal and plastic orthopedic materials with a synthetic, but naturally occurring, bone mineral. ultimately these ceramic materials may be used as bone replacements or with the incorporation of protein collagens, synthetic bones. durable actinide - containing ceramic materials have many applications such as in nuclear fuels for burning excess pu and in chemically - inert sources of alpha irradiation for power supply of unmanned space vehicles or to produce electricity for microelectronic devices. both use and disposal of radioactive actinides require their immobilization in a durable host material. nuclear waste long - lived radionuclides such as actinides are immobilized using chemically - durable crystalline materials based on polycrystalline ceramics and large single crystals. alumina ceramics are widely utilized in the chemical industry due to their excellent chemical stability and high resistance to corrosion. it is used as acid - resistant pump impellers and pump bodies, ensuring long - lasting performance in transferring aggressive fluids. they are also used in acid - carrying pipe linings to prevent contamination and maintain fluid purity, which is crucial in industries like pharmaceuticals and food processing. valves made from alumina ceramics demonstrate exceptional durability and resistance to chemical attack, making them reliable for controlling the flow of corrosive liquids. =
higher concentrations of atmospheric nitrous oxide ( n2o ) are expected to slightly warm earth ' s surface because of increases in radiative forcing. radiative forcing is the difference in the net upward thermal radiation flux from the earth through a transparent atmosphere and radiation through an otherwise identical atmosphere with greenhouse gases. radiative forcing, normally measured in w / m ^ 2, depends on latitude, longitude and altitude, but it is often quoted for the tropopause, about 11 km of altitude for temperate latitudes, or for the top of the atmosphere at around 90 km. for current concentrations of greenhouse gases, the radiative forcing per added n2o molecule is about 230 times larger than the forcing per added carbon dioxide ( co2 ) molecule. this is due to the heavy saturation of the absorption band of the relatively abundant greenhouse gas, co2, compared to the much smaller saturation of the absorption bands of the trace greenhouse gas n2o. but the rate of increase of co2 molecules, about 2. 5 ppm / year ( ppm = part per million by mole ), is about 3000 times larger than the rate of increase of n2o molecules, which has held steady at around 0. 00085 ppm / year since 1985. so, the contribution of nitrous oxide to the annual increase in forcing is 230 / 3000 or about 1 / 13 that of co2. if the main greenhouse gases, co2, ch4 and n2o have contributed about 0. 1 c / decade of the warming observed over the past few decades, this would correspond to about 0. 00064 k per year or 0. 064 k per century of warming from n2o. proposals to place harsh restrictions on nitrous oxide emissions because of warming fears are not justified by these facts. restrictions would cause serious harm ; for example, by jeopardizing world food supplies.
do not survive or become incapable of procreation. plants cannot continue the natural ripening or aging process. all these effects are beneficial to the consumer and the food industry, likewise. the amount of energy imparted for effective food irradiation is low compared to cooking the same ; even at a typical dose of 10 kgy most food, which is ( with regard to warming ) physically equivalent to water, would warm by only about 2. 5 Β°c ( 4. 5 Β°f ). the specialty of processing food by ionizing radiation is the fact, that the energy density per atomic transition is very high, it can cleave molecules and induce ionization ( hence the name ) which cannot be achieved by mere heating. this is the reason for new beneficial effects, however at the same time, for new concerns. the treatment of solid food by ionizing radiation can provide an effect similar to heat pasteurization of liquids, such as milk. however, the use of the term, cold pasteurization, to describe irradiated foods is controversial, because pasteurization and irradiation are fundamentally different processes, although the intended end results can in some cases be similar. detractors of food irradiation have concerns about the health hazards of induced radioactivity. a report for the industry advocacy group american council on science and health entitled " irradiated foods " states : " the types of radiation sources approved for the treatment of foods have specific energy levels well below that which would cause any element in food to become radioactive. food undergoing irradiation does not become any more radioactive than luggage passing through an airport x - ray scanner or teeth that have been x - rayed. " food irradiation is currently permitted by over 40 countries and volumes are estimated to exceed 500, 000 metric tons ( 490, 000 long tons ; 550, 000 short tons ) annually worldwide. food irradiation is essentially a non - nuclear technology ; it relies on the use of ionizing radiation which may be generated by accelerators for electrons and conversion into bremsstrahlung, but which may use also gamma - rays from nuclear decay. there is a worldwide industry for processing by ionizing radiation, the majority by number and by processing power using accelerators. food irradiation is only a niche application compared to medical supplies, plastic materials, raw materials, gemstones, cables and wires, etc. = = accidents = = nuclear accidents, because of the powerful forces involved, are often very dangerous. historically, the first incidents involved fatal
, heat from friction during rolling can cause problems for metal bearings ; problems which are reduced by the use of ceramics. ceramics are also more chemically resistant and can be used in wet environments where steel bearings would rust. the major drawback to using ceramics is a significantly higher cost. in many cases their electrically insulating properties may also be valuable in bearings. in the early 1980s, toyota researched production of an adiabatic ceramic engine which can run at a temperature of over 6000 Β°f ( 3300 Β°c ). ceramic engines do not require a cooling system and hence allow a major weight reduction and therefore greater fuel efficiency. fuel efficiency of the engine is also higher at high temperature, as shown by carnot ' s theorem. in a conventional metallic engine, much of the energy released from the fuel must be dissipated as waste heat in order to prevent a meltdown of the metallic parts. despite all of these desirable properties, such engines are not in production because the manufacturing of ceramic parts in the requisite precision and durability is difficult. imperfection in the ceramic leads to cracks, which can lead to potentially dangerous equipment failure. such engines are possible in laboratory settings, but mass - production is not feasible with current technology. work is being done in developing ceramic parts for gas turbine engines. currently, even blades made of advanced metal alloys used in the engines ' hot section require cooling and careful limiting of operating temperatures. turbine engines made with ceramics could operate more efficiently, giving aircraft greater range and payload for a set amount of fuel. recently, there have been advances in ceramics which include bio - ceramics, such as dental implants and synthetic bones. hydroxyapatite, the natural mineral component of bone, has been made synthetically from a number of biological and chemical sources and can be formed into ceramic materials. orthopedic implants made from these materials bond readily to bone and other tissues in the body without rejection or inflammatory reactions. because of this, they are of great interest for gene delivery and tissue engineering scaffolds. most hydroxyapatite ceramics are very porous and lack mechanical strength and are used to coat metal orthopedic devices to aid in forming a bond to bone or as bone fillers. they are also used as fillers for orthopedic plastic screws to aid in reducing the inflammation and increase absorption of these plastic materials. work is being done to make strong, fully dense nano crystalline hydroxyapatite ceramic materials for orthopedic weight bearing devices, replacing foreign metal and plastic orthopedic materials
background : african swine fever is among the most devastating viral diseases of pigs. despite nearly a century of research, there is still no safe and effective vaccine available. the current situation is that either vaccines are safe but not effective, or they are effective but not safe. findings : the asf vaccine prepared using the inactivation method with propiolactone provided 98. 6 % protection within 100 days after three intranasal immunizations, spaced 7 days apart. conclusions : an inactivated vaccine made from complete african swine fever virus particles using propiolactone is safe and effective for controlling asf through mucosal immunity.
techniques that provide heart and lung support. it is used primarily to support the lungs for a prolonged but still temporary timeframe ( 1 β 30 days ) and allow for recovery from reversible diseases. robert bartlett is known as the father of ecmo and performed the first treatment of a newborn using an ecmo machine in 1975. skin tissue - engineered skin is a type of bioartificial organ that is often used to treat burns, diabetic foot ulcers, or other large wounds that cannot heal well on their own. artificial skin can be made from autografts, allografts, and xenografts. autografted skin comes from a patient ' s own skin, which allows the dermis to have a faster healing rate, and the donor site can be re - harvested a few times. allograft skin often comes from cadaver skin and is mostly used to treat burn victims. lastly, xenografted skin comes from animals and provides a temporary healing structure for the skin. they assist in dermal regeneration, but cannot become part of the host skin. tissue - engineered skin is now available in commercial products. integra, originally used to only treat burns, consists of a collagen matrix and chondroitin sulfate that can be used as a skin replacement. the chondroitin sulfate functions as a component of proteoglycans, which helps to form the extracellular matrix. integra can be repopulated and revascularized while maintaining its dermal collagen architecture, making it a bioartificial organ dermagraft, another commercial - made tissue - engineered skin product, is made out of living fibroblasts. these fibroblasts proliferate and produce growth factors, collagen, and ecm proteins, that help build granulation tissue. = = = = heart = = = = since the number of patients awaiting a heart transplant is continuously increasing over time, and the number of patients on the waiting list surpasses the organ availability, artificial organs used as replacement therapy for terminal heart failure would help alleviate this difficulty. artificial hearts are usually used to bridge the heart transplantation or can be applied as replacement therapy for terminal heart malfunction. the total artificial heart ( tah ), first introduced by dr. vladimir p. demikhov in 1937, emerged as an ideal alternative. since then it has been developed and improved as a mechanical pump that provides long - term circulatory support and
Question: What would best regulate the temperature of an overheated body?
A) a decrease in sweating
B) a decrease in the breathing rate
C) a dilation of blood vessels in the skin
D) an increase in heart rate
|
C) a dilation of blood vessels in the skin
|
Context:
##sphere ( or lithosphere ). earth science can be considered to be a branch of planetary science but with a much older history. = = geology = = geology is broadly the study of earth ' s structure, substance, and processes. geology is largely the study of the lithosphere, or earth ' s surface, including the crust and rocks. it includes the physical characteristics and processes that occur in the lithosphere as well as how they are affected by geothermal energy. it incorporates aspects of chemistry, physics, and biology as elements of geology interact. historical geology is the application of geology to interpret earth history and how it has changed over time. geochemistry studies the chemical components and processes of the earth. geophysics studies the physical properties of the earth. paleontology studies fossilized biological material in the lithosphere. planetary geology studies geoscience as it pertains to extraterrestrial bodies. geomorphology studies the origin of landscapes. structural geology studies the deformation of rocks to produce mountains and lowlands. resource geology studies how energy resources can be obtained from minerals. environmental geology studies how pollution and contaminants affect soil and rock. mineralogy is the study of minerals and includes the study of mineral formation, crystal structure, hazards associated with minerals, and the physical and chemical properties of minerals. petrology is the study of rocks, including the formation and composition of rocks. petrography is a branch of petrology that studies the typology and classification of rocks. = = earth ' s interior = = plate tectonics, mountain ranges, volcanoes, and earthquakes are geological phenomena that can be explained in terms of physical and chemical processes in the earth ' s crust. beneath the earth ' s crust lies the mantle which is heated by the radioactive decay of heavy elements. the mantle is not quite solid and consists of magma which is in a state of semi - perpetual convection. this convection process causes the lithospheric plates to move, albeit slowly. the resulting process is known as plate tectonics. areas of the crust where new crust is created are called divergent boundaries, those where it is brought back into the earth are convergent boundaries and those where plates slide past each other, but no new lithospheric material is created or destroyed, are referred to as transform ( or conservative ) boundaries. earthquakes result from the movement of the lithospheric plates, and they often occur near convergent boundaries where parts of the crust are forced into the earth as
have evolved from the earliest emergence of life to present day. earth formed about 4. 5 billion years ago and all life on earth, both living and extinct, descended from a last universal common ancestor that lived about 3. 5 billion years ago. geologists have developed a geologic time scale that divides the history of the earth into major divisions, starting with four eons ( hadean, archean, proterozoic, and phanerozoic ), the first three of which are collectively known as the precambrian, which lasted approximately 4 billion years. each eon can be divided into eras, with the phanerozoic eon that began 539 million years ago being subdivided into paleozoic, mesozoic, and cenozoic eras. these three eras together comprise eleven periods ( cambrian, ordovician, silurian, devonian, carboniferous, permian, triassic, jurassic, cretaceous, tertiary, and quaternary ). the similarities among all known present - day species indicate that they have diverged through the process of evolution from their common ancestor. biologists regard the ubiquity of the genetic code as evidence of universal common descent for all bacteria, archaea, and eukaryotes. microbial mats of coexisting bacteria and archaea were the dominant form of life in the early archean eon and many of the major steps in early evolution are thought to have taken place in this environment. the earliest evidence of eukaryotes dates from 1. 85 billion years ago, and while they may have been present earlier, their diversification accelerated when they started using oxygen in their metabolism. later, around 1. 7 billion years ago, multicellular organisms began to appear, with differentiated cells performing specialised functions. algae - like multicellular land plants are dated back to about 1 billion years ago, although evidence suggests that microorganisms formed the earliest terrestrial ecosystems, at least 2. 7 billion years ago. microorganisms are thought to have paved the way for the inception of land plants in the ordovician period. land plants were so successful that they are thought to have contributed to the late devonian extinction event. ediacara biota appear during the ediacaran period, while vertebrates, along with most other modern phyla originated about 525 million years ago during the cambrian explosion. during the permian period, synapsids, including the ancestors of mammals, dominated the land, but most of this group became
##ctonics, mountain ranges, volcanoes, and earthquakes are geological phenomena that can be explained in terms of physical and chemical processes in the earth ' s crust. beneath the earth ' s crust lies the mantle which is heated by the radioactive decay of heavy elements. the mantle is not quite solid and consists of magma which is in a state of semi - perpetual convection. this convection process causes the lithospheric plates to move, albeit slowly. the resulting process is known as plate tectonics. areas of the crust where new crust is created are called divergent boundaries, those where it is brought back into the earth are convergent boundaries and those where plates slide past each other, but no new lithospheric material is created or destroyed, are referred to as transform ( or conservative ) boundaries. earthquakes result from the movement of the lithospheric plates, and they often occur near convergent boundaries where parts of the crust are forced into the earth as part of subduction. plate tectonics might be thought of as the process by which the earth is resurfaced. as the result of seafloor spreading, new crust and lithosphere is created by the flow of magma from the mantle to the near surface, through fissures, where it cools and solidifies. through subduction, oceanic crust and lithosphere vehemently returns to the convecting mantle. volcanoes result primarily from the melting of subducted crust material. crust material that is forced into the asthenosphere melts, and some portion of the melted material becomes light enough to rise to the surface β giving birth to volcanoes. = = atmospheric science = = atmospheric science initially developed in the late - 19th century as a means to forecast the weather through meteorology, the study of weather. atmospheric chemistry was developed in the 20th century to measure air pollution and expanded in the 1970s in response to acid rain. climatology studies the climate and climate change. the troposphere, stratosphere, mesosphere, thermosphere, and exosphere are the five layers which make up earth ' s atmosphere. 75 % of the mass in the atmosphere is located within the troposphere, the lowest layer. in all, the atmosphere is made up of about 78. 0 % nitrogen, 20. 9 % oxygen, and 0. 92 % argon, and small amounts of other gases including co2 and water vapor. water vapor and co2 cause the earth ' s atmosphere to catch and hold the sun ' s
a state of semi - perpetual convection. this convection process causes the lithospheric plates to move, albeit slowly. the resulting process is known as plate tectonics. areas of the crust where new crust is created are called divergent boundaries, those where it is brought back into the earth are convergent boundaries and those where plates slide past each other, but no new lithospheric material is created or destroyed, are referred to as transform ( or conservative ) boundaries. earthquakes result from the movement of the lithospheric plates, and they often occur near convergent boundaries where parts of the crust are forced into the earth as part of subduction. plate tectonics might be thought of as the process by which the earth is resurfaced. as the result of seafloor spreading, new crust and lithosphere is created by the flow of magma from the mantle to the near surface, through fissures, where it cools and solidifies. through subduction, oceanic crust and lithosphere vehemently returns to the convecting mantle. volcanoes result primarily from the melting of subducted crust material. crust material that is forced into the asthenosphere melts, and some portion of the melted material becomes light enough to rise to the surface β giving birth to volcanoes. = = atmospheric science = = atmospheric science initially developed in the late - 19th century as a means to forecast the weather through meteorology, the study of weather. atmospheric chemistry was developed in the 20th century to measure air pollution and expanded in the 1970s in response to acid rain. climatology studies the climate and climate change. the troposphere, stratosphere, mesosphere, thermosphere, and exosphere are the five layers which make up earth ' s atmosphere. 75 % of the mass in the atmosphere is located within the troposphere, the lowest layer. in all, the atmosphere is made up of about 78. 0 % nitrogen, 20. 9 % oxygen, and 0. 92 % argon, and small amounts of other gases including co2 and water vapor. water vapor and co2 cause the earth ' s atmosphere to catch and hold the sun ' s energy through the greenhouse effect. this makes earth ' s surface warm enough for liquid water and life. in addition to trapping heat, the atmosphere also protects living organisms by shielding the earth ' s surface from cosmic rays. the magnetic field β created by the internal motions of the core β produces the magnetosphere which protects earth '
variation in total solar irradiance is thought to have little effect on the earth ' s surface temperature because of the thermal time constant - - the characteristic response time of the earth ' s global surface temperature to changes in forcing. this time constant is large enough to smooth annual variations but not necessarily variations having a longer period such as those due to solar inertial motion ; the magnitude of these surface temperature variations is estimated.
, crystal structure, hazards associated with minerals, and the physical and chemical properties of minerals. petrology is the study of rocks, including the formation and composition of rocks. petrography is a branch of petrology that studies the typology and classification of rocks. = = earth ' s interior = = plate tectonics, mountain ranges, volcanoes, and earthquakes are geological phenomena that can be explained in terms of physical and chemical processes in the earth ' s crust. beneath the earth ' s crust lies the mantle which is heated by the radioactive decay of heavy elements. the mantle is not quite solid and consists of magma which is in a state of semi - perpetual convection. this convection process causes the lithospheric plates to move, albeit slowly. the resulting process is known as plate tectonics. areas of the crust where new crust is created are called divergent boundaries, those where it is brought back into the earth are convergent boundaries and those where plates slide past each other, but no new lithospheric material is created or destroyed, are referred to as transform ( or conservative ) boundaries. earthquakes result from the movement of the lithospheric plates, and they often occur near convergent boundaries where parts of the crust are forced into the earth as part of subduction. plate tectonics might be thought of as the process by which the earth is resurfaced. as the result of seafloor spreading, new crust and lithosphere is created by the flow of magma from the mantle to the near surface, through fissures, where it cools and solidifies. through subduction, oceanic crust and lithosphere vehemently returns to the convecting mantle. volcanoes result primarily from the melting of subducted crust material. crust material that is forced into the asthenosphere melts, and some portion of the melted material becomes light enough to rise to the surface β giving birth to volcanoes. = = atmospheric science = = atmospheric science initially developed in the late - 19th century as a means to forecast the weather through meteorology, the study of weather. atmospheric chemistry was developed in the 20th century to measure air pollution and expanded in the 1970s in response to acid rain. climatology studies the climate and climate change. the troposphere, stratosphere, mesosphere, thermosphere, and exosphere are the five layers which make up earth ' s atmosphere. 75 % of the mass in the atmosphere is located within the troposphere, the lowest
earth science or geoscience includes all fields of natural science related to the planet earth. this is a branch of science dealing with the physical, chemical, and biological complex constitutions and synergistic linkages of earth ' s four spheres : the biosphere, hydrosphere / cryosphere, atmosphere, and geosphere ( or lithosphere ). earth science can be considered to be a branch of planetary science but with a much older history. = = geology = = geology is broadly the study of earth ' s structure, substance, and processes. geology is largely the study of the lithosphere, or earth ' s surface, including the crust and rocks. it includes the physical characteristics and processes that occur in the lithosphere as well as how they are affected by geothermal energy. it incorporates aspects of chemistry, physics, and biology as elements of geology interact. historical geology is the application of geology to interpret earth history and how it has changed over time. geochemistry studies the chemical components and processes of the earth. geophysics studies the physical properties of the earth. paleontology studies fossilized biological material in the lithosphere. planetary geology studies geoscience as it pertains to extraterrestrial bodies. geomorphology studies the origin of landscapes. structural geology studies the deformation of rocks to produce mountains and lowlands. resource geology studies how energy resources can be obtained from minerals. environmental geology studies how pollution and contaminants affect soil and rock. mineralogy is the study of minerals and includes the study of mineral formation, crystal structure, hazards associated with minerals, and the physical and chemical properties of minerals. petrology is the study of rocks, including the formation and composition of rocks. petrography is a branch of petrology that studies the typology and classification of rocks. = = earth ' s interior = = plate tectonics, mountain ranges, volcanoes, and earthquakes are geological phenomena that can be explained in terms of physical and chemical processes in the earth ' s crust. beneath the earth ' s crust lies the mantle which is heated by the radioactive decay of heavy elements. the mantle is not quite solid and consists of magma which is in a state of semi - perpetual convection. this convection process causes the lithospheric plates to move, albeit slowly. the resulting process is known as plate tectonics. areas of the crust where new crust is created are called divergent boundaries, those where it is brought back into the earth are convergent boundaries and
be the more significant to modern soil theory than fallou ' s. previously, soil had been considered a product of chemical transformations of rocks, a dead substrate from which plants derive nutritious elements. soil and bedrock were in fact equated. dokuchaev considers the soil as a natural body having its own genesis and its own history of development, a body with complex and multiform processes taking place within it. the soil is considered as different from bedrock. the latter becomes soil under the influence of a series of soil - formation factors ( climate, vegetation, country, relief and age ). according to him, soil should be called the " daily " or outward horizons of rocks regardless of the type ; they are changed naturally by the common effect of water, air and various kinds of living and dead organisms. a 1914 encyclopedic definition : " the different forms of earth on the surface of the rocks, formed by the breaking down or weathering of rocks ". serves to illustrate the historic view of soil which persisted from the 19th century. dokuchaev ' s late 19th century soil concept developed in the 20th century to one of soil as earthy material that has been altered by living processes. a corollary concept is that soil without a living component is simply a part of earth ' s outer layer. further refinement of the soil concept is occurring in view of an appreciation of energy transport and transformation within soil. the term is popularly applied to the material on the surface of the earth ' s moon and mars, a usage acceptable within a portion of the scientific community. accurate to this modern understanding of soil is nikiforoff ' s 1959 definition of soil as the " excited skin of the sub aerial part of the earth ' s crust ". = = areas of practice = = academically, soil scientists tend to be drawn to one of five areas of specialization : microbiology, pedology, edaphology, physics, or chemistry. yet the work specifics are very much dictated by the challenges facing our civilization ' s desire to sustain the land that supports it, and the distinctions between the sub - disciplines of soil science often blur in the process. soil science professionals commonly stay current in soil chemistry, soil physics, soil microbiology, pedology, and applied soil science in related disciplines. one exciting effort drawing in soil scientists in the u. s. as of 2004 is the soil quality initiative. central to the soil quality initiative is developing indices of soil health and then monitoring them in a way
##morphology studies the origin of landscapes. structural geology studies the deformation of rocks to produce mountains and lowlands. resource geology studies how energy resources can be obtained from minerals. environmental geology studies how pollution and contaminants affect soil and rock. mineralogy is the study of minerals and includes the study of mineral formation, crystal structure, hazards associated with minerals, and the physical and chemical properties of minerals. petrology is the study of rocks, including the formation and composition of rocks. petrography is a branch of petrology that studies the typology and classification of rocks. = = earth ' s interior = = plate tectonics, mountain ranges, volcanoes, and earthquakes are geological phenomena that can be explained in terms of physical and chemical processes in the earth ' s crust. beneath the earth ' s crust lies the mantle which is heated by the radioactive decay of heavy elements. the mantle is not quite solid and consists of magma which is in a state of semi - perpetual convection. this convection process causes the lithospheric plates to move, albeit slowly. the resulting process is known as plate tectonics. areas of the crust where new crust is created are called divergent boundaries, those where it is brought back into the earth are convergent boundaries and those where plates slide past each other, but no new lithospheric material is created or destroyed, are referred to as transform ( or conservative ) boundaries. earthquakes result from the movement of the lithospheric plates, and they often occur near convergent boundaries where parts of the crust are forced into the earth as part of subduction. plate tectonics might be thought of as the process by which the earth is resurfaced. as the result of seafloor spreading, new crust and lithosphere is created by the flow of magma from the mantle to the near surface, through fissures, where it cools and solidifies. through subduction, oceanic crust and lithosphere vehemently returns to the convecting mantle. volcanoes result primarily from the melting of subducted crust material. crust material that is forced into the asthenosphere melts, and some portion of the melted material becomes light enough to rise to the surface β giving birth to volcanoes. = = atmospheric science = = atmospheric science initially developed in the late - 19th century as a means to forecast the weather through meteorology, the study of weather. atmospheric chemistry was developed in the 20th century to measure air pollution and expanded in the 1970s in response to
the fundamental constants could not influence different elements uniformly, and a comparison between each of the elements ' resulting unique chronological timescales would then give inconsistent time estimates. in refutation of young earth claims of inconstant decay rates affecting the reliability of radiometric dating, roger c. wiens, a physicist specializing in isotope dating states : there are only three quite technical instances where a half - life changes, and these do not affect the dating methods : " only one technical exception occurs under terrestrial conditions, and this is not for an isotope used for dating.... the artificially - produced isotope, beryllium - 7 has been shown to change by up to 1. 5 %, depending on its chemical environment.... heavier atoms are even less subject to these minute changes, so the dates of rocks made by electron - capture decays would only be off by at most a few hundredths of a percent. " "... another case is material inside of stars, which is in a plasma state where electrons are not bound to atoms. in the extremely hot stellar environment, a completely different kind of decay can occur. ' bound - state beta decay ' occurs when the nucleus emits an electron into a bound electronic state close to the nucleus.... all normal matter, such as everything on earth, the moon, meteorites, etc. has electrons in normal positions, so these instances never apply to rocks, or anything colder than several hundred thousand degrees. " " the last case also involves very fast - moving matter. it has been demonstrated by atomic clocks in very fast spacecraft. these atomic clocks slow down very slightly ( only a second or so per year ) as predicted by einstein ' s theory of relativity. no rocks in our solar system are going fast enough to make a noticeable change in their dates. " = = = = radiohaloes = = = = in the 1970s, young earth creationist robert v. gentry proposed that radiohaloes in certain granites represented evidence for the earth being created instantaneously rather than gradually. this idea has been criticized by physicists and geologists on many grounds including that the rocks gentry studied were not primordial and that the radionuclides in question need not have been in the rocks initially. thomas a. baillieul, a geologist and retired senior environmental scientist with the united states department of energy, disputed gentry ' s claims in an article entitled, " ' polonium haloes ' refuted : a review of ' radioactive halos in a radio
Question: The surface of Earth is constantly changing. Which of these findings provides the best evidence that the crustal surface of Earth has changed over time?
A) Some rock layers are very thick.
B) Some species have become extinct.
C) Marine fossils are found in mountain areas.
D) Igneous rock is found on all tectonic plates.
|
C) Marine fossils are found in mountain areas.
|
Context:
##itive material by selective exposure to a radiation source such as light. a photosensitive material is a material that experiences a change in its physical properties when exposed to a radiation source. if a photosensitive material is selectively exposed to radiation ( e. g. by masking some of the radiation ) the pattern of the radiation on the material is transferred to the material exposed, as the properties of the exposed and unexposed regions differs. this exposed region can then be removed or treated providing a mask for the underlying substrate. photolithography is typically used with metal or other thin film deposition, wet and dry etching. sometimes, photolithography is used to create structure without any kind of post etching. one example is su8 based lens where su8 based square blocks are generated. then the photoresist is melted to form a semi - sphere which acts as a lens. electron beam lithography ( often abbreviated as e - beam lithography ) is the practice of scanning a beam of electrons in a patterned fashion across a surface covered with a film ( called the resist ), ( " exposing " the resist ) and of selectively removing either exposed or non - exposed regions of the resist ( " developing " ). the purpose, as with photolithography, is to create very small structures in the resist that can subsequently be transferred to the substrate material, often by etching. it was developed for manufacturing integrated circuits, and is also used for creating nanotechnology architectures. the primary advantage of electron beam lithography is that it is one of the ways to beat the diffraction limit of light and make features in the nanometer range. this form of maskless lithography has found wide usage in photomask - making used in photolithography, low - volume production of semiconductor components, and research & development. the key limitation of electron beam lithography is throughput, i. e., the very long time it takes to expose an entire silicon wafer or glass substrate. a long exposure time leaves the user vulnerable to beam drift or instability which may occur during the exposure. also, the turn - around time for reworking or re - design is lengthened unnecessarily if the pattern is not being changed the second time. it is known that focused - ion beam lithography has the capability of writing extremely fine lines ( less than 50 nm line and space has been achieved ) without proximity effect. however, because the writing field in ion - beam lit
, natural phenomena on earth only involve gravity and electromagnetism, and not nuclear reactions. this is because atomic nuclei are generally kept apart because they contain positive electrical charges and therefore repel each other. in 1896, henri becquerel was investigating phosphorescence in uranium salts when he discovered a new phenomenon which came to be called radioactivity. he, pierre curie and marie curie began investigating the phenomenon. in the process, they isolated the element radium, which is highly radioactive. they discovered that radioactive materials produce intense, penetrating rays of three distinct sorts, which they labeled alpha, beta, and gamma after the first three greek letters. some of these kinds of radiation could pass through ordinary matter, and all of them could be harmful in large amounts. all of the early researchers received various radiation burns, much like sunburn, and thought little of it. the new phenomenon of radioactivity was seized upon by the manufacturers of quack medicine ( as had the discoveries of electricity and magnetism, earlier ), and a number of patent medicines and treatments involving radioactivity were put forward. gradually it was realized that the radiation produced by radioactive decay was ionizing radiation, and that even quantities too small to burn could pose a severe long - term hazard. many of the scientists working on radioactivity died of cancer as a result of their exposure. radioactive patent medicines mostly disappeared, but other applications of radioactive materials persisted, such as the use of radium salts to produce glowing dials on meters. as the atom came to be better understood, the nature of radioactivity became clearer. some larger atomic nuclei are unstable, and so decay ( release matter or energy ) after a random interval. the three forms of radiation that becquerel and the curies discovered are also more fully understood. alpha decay is when a nucleus releases an alpha particle, which is two protons and two neutrons, equivalent to a helium nucleus. beta decay is the release of a beta particle, a high - energy electron. gamma decay releases gamma rays, which unlike alpha and beta radiation are not matter but electromagnetic radiation of very high frequency, and therefore energy. this type of radiation is the most dangerous and most difficult to block. all three types of radiation occur naturally in certain elements. it has also become clear that the ultimate source of most terrestrial energy is nuclear, either through radiation from the sun caused by stellar thermonuclear reactions or by radioactive decay of uranium within the earth, the principal source of geothermal energy.
which came to be called radioactivity. he, pierre curie and marie curie began investigating the phenomenon. in the process, they isolated the element radium, which is highly radioactive. they discovered that radioactive materials produce intense, penetrating rays of three distinct sorts, which they labeled alpha, beta, and gamma after the first three greek letters. some of these kinds of radiation could pass through ordinary matter, and all of them could be harmful in large amounts. all of the early researchers received various radiation burns, much like sunburn, and thought little of it. the new phenomenon of radioactivity was seized upon by the manufacturers of quack medicine ( as had the discoveries of electricity and magnetism, earlier ), and a number of patent medicines and treatments involving radioactivity were put forward. gradually it was realized that the radiation produced by radioactive decay was ionizing radiation, and that even quantities too small to burn could pose a severe long - term hazard. many of the scientists working on radioactivity died of cancer as a result of their exposure. radioactive patent medicines mostly disappeared, but other applications of radioactive materials persisted, such as the use of radium salts to produce glowing dials on meters. as the atom came to be better understood, the nature of radioactivity became clearer. some larger atomic nuclei are unstable, and so decay ( release matter or energy ) after a random interval. the three forms of radiation that becquerel and the curies discovered are also more fully understood. alpha decay is when a nucleus releases an alpha particle, which is two protons and two neutrons, equivalent to a helium nucleus. beta decay is the release of a beta particle, a high - energy electron. gamma decay releases gamma rays, which unlike alpha and beta radiation are not matter but electromagnetic radiation of very high frequency, and therefore energy. this type of radiation is the most dangerous and most difficult to block. all three types of radiation occur naturally in certain elements. it has also become clear that the ultimate source of most terrestrial energy is nuclear, either through radiation from the sun caused by stellar thermonuclear reactions or by radioactive decay of uranium within the earth, the principal source of geothermal energy. = = = nuclear fission = = = in natural nuclear radiation, the byproducts are very small compared to the nuclei from which they originate. nuclear fission is the process of splitting a nucleus into roughly equal parts, and releasing energy and neutrons in the process. if these neutrons are captured by another unstable nucleus
the first three greek letters. some of these kinds of radiation could pass through ordinary matter, and all of them could be harmful in large amounts. all of the early researchers received various radiation burns, much like sunburn, and thought little of it. the new phenomenon of radioactivity was seized upon by the manufacturers of quack medicine ( as had the discoveries of electricity and magnetism, earlier ), and a number of patent medicines and treatments involving radioactivity were put forward. gradually it was realized that the radiation produced by radioactive decay was ionizing radiation, and that even quantities too small to burn could pose a severe long - term hazard. many of the scientists working on radioactivity died of cancer as a result of their exposure. radioactive patent medicines mostly disappeared, but other applications of radioactive materials persisted, such as the use of radium salts to produce glowing dials on meters. as the atom came to be better understood, the nature of radioactivity became clearer. some larger atomic nuclei are unstable, and so decay ( release matter or energy ) after a random interval. the three forms of radiation that becquerel and the curies discovered are also more fully understood. alpha decay is when a nucleus releases an alpha particle, which is two protons and two neutrons, equivalent to a helium nucleus. beta decay is the release of a beta particle, a high - energy electron. gamma decay releases gamma rays, which unlike alpha and beta radiation are not matter but electromagnetic radiation of very high frequency, and therefore energy. this type of radiation is the most dangerous and most difficult to block. all three types of radiation occur naturally in certain elements. it has also become clear that the ultimate source of most terrestrial energy is nuclear, either through radiation from the sun caused by stellar thermonuclear reactions or by radioactive decay of uranium within the earth, the principal source of geothermal energy. = = = nuclear fission = = = in natural nuclear radiation, the byproducts are very small compared to the nuclei from which they originate. nuclear fission is the process of splitting a nucleus into roughly equal parts, and releasing energy and neutrons in the process. if these neutrons are captured by another unstable nucleus, they can fission as well, leading to a chain reaction. the average number of neutrons released per nucleus that go on to fission another nucleus is referred to as k. values of k larger than 1 mean that the fission reaction is releasing more neutrons than it absorbs, and therefore is referred to as a self
variation in total solar irradiance is thought to have little effect on the earth ' s surface temperature because of the thermal time constant - - the characteristic response time of the earth ' s global surface temperature to changes in forcing. this time constant is large enough to smooth annual variations but not necessarily variations having a longer period such as those due to solar inertial motion ; the magnitude of these surface temperature variations is estimated.
the motion of celestial bodies through a higher power such as god. aristotle did not have the technological advancements that would have explained the motion of celestial bodies. in addition, aristotle had many views on the elements. he believed that everything was derived of the elements earth, water, air, fire, and lastly the aether. the aether was a celestial element, and therefore made up the matter of the celestial bodies. the elements of earth, water, air and fire were derived of a combination of two of the characteristics of hot, wet, cold, and dry, and all had their inevitable place and motion. the motion of these elements begins with earth being the closest to " the earth, " then water, air, fire, and finally aether. in addition to the makeup of all things, aristotle came up with theories as to why things did not return to their natural motion. he understood that water sits above earth, air above water, and fire above air in their natural state. he explained that although all elements must return to their natural state, the human body and other living things have a constraint on the elements β thus not allowing the elements making one who they are to return to their natural state. the important legacy of this period included substantial advances in factual knowledge, especially in anatomy, zoology, botany, mineralogy, geography, mathematics and astronomy ; an awareness of the importance of certain scientific problems, especially those related to the problem of change and its causes ; and a recognition of the methodological importance of applying mathematics to natural phenomena and of undertaking empirical research. in the hellenistic age scholars frequently employed the principles developed in earlier greek thought : the application of mathematics and deliberate empirical research, in their scientific investigations. thus, clear unbroken lines of influence lead from ancient greek and hellenistic philosophers, to medieval muslim philosophers and scientists, to the european renaissance and enlightenment, to the secular sciences of the modern day. neither reason nor inquiry began with the ancient greeks, but the socratic method did, along with the idea of forms, give great advances in geometry, logic, and the natural sciences. according to benjamin farrington, former professor of classics at swansea university : " men were weighing for thousands of years before archimedes worked out the laws of equilibrium ; they must have had practical and intuitional knowledge of the principals involved. what archimedes did was to sort out the theoretical implications of this practical knowledge and present the resulting body of knowledge as a logically coherent system. " and again : " with astonishment we find ourselves on the threshold of modern science
harding - cox presidential election. = = technology = = radio waves are radiated by electric charges undergoing acceleration. they are generated artificially by time - varying electric currents, consisting of electrons flowing back and forth in a metal conductor called an antenna. as they travel farther from the transmitting antenna, radio waves spread out so their signal strength ( intensity in watts per square meter ) decreases ( see inverse - square law ), so radio transmissions can only be received within a limited range of the transmitter, the distance depending on the transmitter power, the antenna radiation pattern, receiver sensitivity, background noise level, and presence of obstructions between transmitter and receiver. an omnidirectional antenna transmits or receives radio waves in all directions, while a directional antenna transmits radio waves in a beam in a particular direction, or receives waves from only one direction. radio waves travel at the speed of light in vacuum and at slightly lower velocity in air. the other types of electromagnetic waves besides radio waves, infrared, visible light, ultraviolet, x - rays and gamma rays, can also carry information and be used for communication. the wide use of radio waves for telecommunication is mainly due to their desirable propagation properties stemming from their longer wavelength. radio waves have the ability to pass through the atmosphere in any weather, foliage, and at longer wavelengths through most building materials. by diffraction, longer wavelengths can bend around obstructions, and unlike other electromagnetic waves they tend to be scattered rather than absorbed by objects larger than their wavelength. = = radio communication = = in radio communication systems, information is carried across space using radio waves. at the sending end, the information to be sent is converted by some type of transducer to a time - varying electrical signal called the modulation signal. the modulation signal may be an audio signal representing sound from a microphone, a video signal representing moving images from a video camera, or a digital signal consisting of a sequence of bits representing binary data from a computer. the modulation signal is applied to a radio transmitter. in the transmitter, an electronic oscillator generates an alternating current oscillating at a radio frequency, called the carrier wave because it serves to generate the radio waves that carry the information through the air. the modulation signal is used to modulate the carrier, varying some aspect of the carrier wave, impressing the information in the modulation signal onto the carrier. different radio systems use different modulation methods : amplitude modulation ( am ) β in an am transmitter, the amplitude ( strength ) of the radio carrier wave is varied by the modulation
two neutrons, equivalent to a helium nucleus. beta decay is the release of a beta particle, a high - energy electron. gamma decay releases gamma rays, which unlike alpha and beta radiation are not matter but electromagnetic radiation of very high frequency, and therefore energy. this type of radiation is the most dangerous and most difficult to block. all three types of radiation occur naturally in certain elements. it has also become clear that the ultimate source of most terrestrial energy is nuclear, either through radiation from the sun caused by stellar thermonuclear reactions or by radioactive decay of uranium within the earth, the principal source of geothermal energy. = = = nuclear fission = = = in natural nuclear radiation, the byproducts are very small compared to the nuclei from which they originate. nuclear fission is the process of splitting a nucleus into roughly equal parts, and releasing energy and neutrons in the process. if these neutrons are captured by another unstable nucleus, they can fission as well, leading to a chain reaction. the average number of neutrons released per nucleus that go on to fission another nucleus is referred to as k. values of k larger than 1 mean that the fission reaction is releasing more neutrons than it absorbs, and therefore is referred to as a self - sustaining chain reaction. a mass of fissile material large enough ( and in a suitable configuration ) to induce a self - sustaining chain reaction is called a critical mass. when a neutron is captured by a suitable nucleus, fission may occur immediately, or the nucleus may persist in an unstable state for a short time. if there are enough immediate decays to carry on the chain reaction, the mass is said to be prompt critical, and the energy release will grow rapidly and uncontrollably, usually leading to an explosion. when discovered on the eve of world war ii, this insight led multiple countries to begin programs investigating the possibility of constructing an atomic bomb β a weapon which utilized fission reactions to generate far more energy than could be created with chemical explosives. the manhattan project, run by the united states with the help of the united kingdom and canada, developed multiple fission weapons which were used against japan in 1945 at hiroshima and nagasaki. during the project, the first fission reactors were developed as well, though they were primarily for weapons manufacture and did not generate electricity. in 1951, the first nuclear fission power plant was the first to produce electricity at the experimental breeder reactor no. 1 ( ebr - 1 ), in arco, idaho, ushering in the "
radiation exposure. marie curie died from aplastic anemia which resulted from her high levels of exposure. two scientists, an american and canadian respectively, harry daghlian and louis slotin, died after mishandling the same plutonium mass. unlike conventional weapons, the intense light, heat, and explosive force is not the only deadly component to a nuclear weapon. approximately half of the deaths from hiroshima and nagasaki died two to five years afterward from radiation exposure. civilian nuclear and radiological accidents primarily involve nuclear power plants. most common are nuclear leaks that expose workers to hazardous material. a nuclear meltdown refers to the more serious hazard of releasing nuclear material into the surrounding environment. the most significant meltdowns occurred at three mile island in pennsylvania and chernobyl in the soviet ukraine. the earthquake and tsunami on march 11, 2011 caused serious damage to three nuclear reactors and a spent fuel storage pond at the fukushima daiichi nuclear power plant in japan. military reactors that experienced similar accidents were windscale in the united kingdom and sl - 1 in the united states. military accidents usually involve the loss or unexpected detonation of nuclear weapons. the castle bravo test in 1954 produced a larger yield than expected, which contaminated nearby islands, a japanese fishing boat ( with one fatality ), and raised concerns about contaminated fish in japan. in the 1950s through 1970s, several nuclear bombs were lost from submarines and aircraft, some of which have never been recovered. the last twenty years have seen a marked decline in such accidents. = = examples of environmental benefits = = proponents of nuclear energy note that annually, nuclear - generated electricity reduces 470 million metric tons of carbon dioxide emissions that would otherwise come from fossil fuels. additionally, the amount of comparatively low waste that nuclear energy does create is safely disposed of by the large scale nuclear energy production facilities or it is repurposed / recycled for other energy uses. proponents of nuclear energy also bring to attention the opportunity cost of utilizing other forms of electricity. for example, the environmental protection agency estimates that coal kills 30, 000 people a year, as a result of its environmental impact, while 60 people died in the chernobyl disaster. a real world example of impact provided by proponents of nuclear energy is the 650, 000 ton increase in carbon emissions in the two months following the closure of the vermont yankee nuclear plant. = = see also = = atomic age lists of nuclear disasters and radioactive incidents nuclear power debate outline of nuclear technology radiology = = references = = = = external links = = nuclear energy institute β beneficial uses
stations located in places like light poles or building roofs. in the past, 4g networking had to rely on large cell towers in order to transmit signals over large distances. with the introduction of 5g networking, it is imperative that small cell stations are used because the mm wave spectrum, which is the specific type of band used in 5g services, strictly travels over short distances. if the distances between cell stations were longer, signals may suffer from interference from inclimate weather, or other objects such as houses, buildings, trees, and much more. in 5g networking, there are 3 main kinds of 5g : low - band, mid - band, and high - band. low - band frequencies operate below 2 ghz, mid - band frequencies operate between 2 β 10 ghz, and high - band frequencies operate between 20 and 100 ghz. verizon have seen outrageous numbers on their high - band 5g service, which they deem " ultraband ", which hit speeds of over 3 gbit / s. the main advantage of 5g networks is that the data transmission rate is much higher than the previous cellular network, up to 10 gbit / s, which is faster than the current wired internet and 100 times faster than the previous 4g lte cellular network. another advantage is lower network latency ( faster response time ), less than 1 millisecond, and 4g is 30 - 70 milliseconds. the peak rate needs to reach the gbit / s standard to meet the high data volume of high - definition video, virtual reality and so on. the air interface delay level needs to be around 1ms, which meets real - time applications such as autonomous driving and telemedicine. large network capacity, providing the connection capacity of 100 billion devices to meet iot communication. the spectrum efficiency is 10 times higher than lte. with continuous wide area coverage and high mobility, the user experience rate reaches 100 mbit / s. the flow density and the number of connections are greatly increased. since 5g is a relatively new type of service, only phones which are newly released or are upcoming can support 5g service. some of these phones include the iphone 12 / 13 ; select samsung devices such as the s21 series, note series, flip / fold series, a series ; google pixel 4a / 5 ; and a few more devices from other manufacturers. the first ever 5g smartphone, the samsung galaxy s20, was released by samsung in march 2020. following the release of samsung ' s s
Question: Solar radiation is a combination of light with different wavelengths. Which type of ray can harm human skin the most?
A) infrared
B) microwave
C) ultraviolet
D) visible
|
C) ultraviolet
|
Context:
remediation include ; soil contamination, hazardous waste, groundwater contamination, oil, gas and chemical spills. there are three most common types of environmental remediation. these include soil, water, and sediment remediation. soil remediation consists of removing contaminants in soil, as these pose great risks to humans and the ecosystem. some examples of this are heavy metals, pesticides, and radioactive materials. depending on the contaminant the remedial processes can be physical, chemical, thermal, or biological. water remediation is one of the most important considering water is an essential natural resource. depending on the source of water there will be different contaminants. surface water contamination mainly consists of agricultural, animal, and industrial waste, as well as acid mine drainage. there has been a rise in the need for water remediation due to the increased discharge of industrial waste, leading to a demand for sustainable water solutions. the market for water remediation is expected to consistently increase to $ 19. 6 billion by 2030. sediment remediation consists of removing contaminated sediments. is it almost similar to soil remediation except it is often more sophisticated as it involves additional contaminants. to reduce the contaminants it is likely to use physical, chemical, and biological processes that help with source control, but if these processes are executed correctly, there ' s a risk of contamination resurfacing. = = = solid waste management = = = solid waste management is the purification, consumption, reuse, disposal, and treatment of solid waste that is undertaken by the government or the ruling bodies of a city / town. it refers to the collection, treatment, and disposal of non - soluble, solid waste material. solid waste is associated with both industrial, institutional, commercial and residential activities. hazardous solid waste, when improperly disposed can encourage the infestation of insects and rodents, contributing to the spread of diseases. some of the most common types of solid waste management include ; landfills, vermicomposting, composting, recycling, and incineration. however, a major barrier for solid waste management practices is the high costs associated with recycling and the risks of creating more pollution. = = = e - waste recycling = = = the recycling of electronic waste ( e - waste ) has seen significant technological advancements due to increasing environmental concerns and the growing volume of electronic product disposals. traditional e - waste recycling methods, which often involve manual disassemb
= = = = = = environmental remediation = = = environmental remediation is the process through which contaminants or pollutants in soil, water and other media are removed to improve environmental quality. the main focus is the reduction of hazardous substances within the environment. some of the areas involved in environmental remediation include ; soil contamination, hazardous waste, groundwater contamination, oil, gas and chemical spills. there are three most common types of environmental remediation. these include soil, water, and sediment remediation. soil remediation consists of removing contaminants in soil, as these pose great risks to humans and the ecosystem. some examples of this are heavy metals, pesticides, and radioactive materials. depending on the contaminant the remedial processes can be physical, chemical, thermal, or biological. water remediation is one of the most important considering water is an essential natural resource. depending on the source of water there will be different contaminants. surface water contamination mainly consists of agricultural, animal, and industrial waste, as well as acid mine drainage. there has been a rise in the need for water remediation due to the increased discharge of industrial waste, leading to a demand for sustainable water solutions. the market for water remediation is expected to consistently increase to $ 19. 6 billion by 2030. sediment remediation consists of removing contaminated sediments. is it almost similar to soil remediation except it is often more sophisticated as it involves additional contaminants. to reduce the contaminants it is likely to use physical, chemical, and biological processes that help with source control, but if these processes are executed correctly, there ' s a risk of contamination resurfacing. = = = solid waste management = = = solid waste management is the purification, consumption, reuse, disposal, and treatment of solid waste that is undertaken by the government or the ruling bodies of a city / town. it refers to the collection, treatment, and disposal of non - soluble, solid waste material. solid waste is associated with both industrial, institutional, commercial and residential activities. hazardous solid waste, when improperly disposed can encourage the infestation of insects and rodents, contributing to the spread of diseases. some of the most common types of solid waste management include ; landfills, vermicomposting, composting, recycling, and incineration. however, a major barrier for solid waste management practices is the high costs associated with recycling
earth. it emphasizes the study of how humans use and interact with freshwater supplies. study of water ' s movement is closely related to geomorphology and other branches of earth science. applied hydrology involves engineering to maintain aquatic environments and distribute water supplies. subdisciplines of hydrology include oceanography, hydrogeology, ecohydrology, and glaciology. oceanography is the study of oceans. hydrogeology is the study of groundwater. it includes the mapping of groundwater supplies and the analysis of groundwater contaminants. applied hydrogeology seeks to prevent contamination of groundwater and mineral springs and make it available as drinking water. the earliest exploitation of groundwater resources dates back to 3000 bc, and hydrogeology as a science was developed by hydrologists beginning in the 17th century. ecohydrology is the study of ecological systems in the hydrosphere. it can be divided into the physical study of aquatic ecosystems and the biological study of aquatic organisms. ecohydrology includes the effects that organisms and aquatic ecosystems have on one another as well as how these ecoystems are affected by humans. glaciology is the study of the cryosphere, including glaciers and coverage of the earth by ice and snow. concerns of glaciology include access to glacial freshwater, mitigation of glacial hazards, obtaining resources that exist beneath frozen land, and addressing the effects of climate change on the cryosphere. = = ecology = = ecology is the study of the biosphere. this includes the study of nature and of how living things interact with the earth and one another and the consequences of that. it considers how living things use resources such as oxygen, water, and nutrients from the earth to sustain themselves. it also considers how humans and other living creatures cause changes to nature. = = physical geography = = physical geography is the study of earth ' s systems and how they interact with one another as part of a single self - contained system. it incorporates astronomy, mathematical geography, meteorology, climatology, geology, geomorphology, biology, biogeography, pedology, and soils geography. physical geography is distinct from human geography, which studies the human populations on earth, though it does include human effects on the environment. = = methodology = = methodologies vary depending on the nature of the subjects being studied. studies typically fall into one of three categories : observational, experimental, or theoretical. earth scientists often conduct sophisticated computer analysis or visit an interesting location to study earth phenomena (
becomes quite gentle. accordingly, in large basins, rivers in most cases begin as torrents with a variable flow, and end as gently flowing rivers with a comparatively regular discharge. the irregular flow of rivers throughout their course forms one of the main difficulties in devising works for mitigating inundations or for increasing the navigable capabilities of rivers. in tropical countries subject to periodical rains, the rivers are in flood during the rainy season and have hardly any flow during the rest of the year, while in temperate regions, where the rainfall is more evenly distributed throughout the year, evaporation causes the available rainfall to be much less in hot summer weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under
equalizing the flow at these places, produces a lowering of the river above the rapids by facilitating the efflux, which may result in the appearance of fresh shoals at the low stage of the river. where, however, narrow rocky reefs or other hard shoals stretch across the bottom of a river and present obstacles to the erosion by the current of the soft materials forming the bed of the river above and below, their removal may result in permanent improvement by enabling the river to deepen its bed by natural scour. the capability of a river to provide a waterway for navigation during the summer or throughout the dry season depends on the depth that can be secured in the channel at the lowest stage. the problem in the dry season is the small discharge and deficiency in scour during this period. a typical solution is to restrict the width of the low - water channel, concentrate all of the flow in it, and also to fix its position so that it is scoured out every year by the floods which follow the deepest part of the bed along the line of the strongest current. this can be effected by closing subsidiary low - water channels with dikes across them, and narrowing the channel at the low stage by low - dipping cross dikes extending from the river banks down the slope and pointing slightly up - stream so as to direct the water flowing over them into a central channel. = = estuarine works = = the needs of navigation may also require that a stable, continuous, navigable channel is prolonged from the navigable river to deep water at the mouth of the estuary. the interaction of river flow and tide needs to be modeled by computer or using scale models, moulded to the configuration of the estuary under consideration and reproducing in miniature the tidal ebb and flow and fresh - water discharge over a bed of fine sand, in which various lines of training walls can be successively inserted. the models should be capable of furnishing valuable indications of the respective effects and comparative merits of the different schemes proposed for works. = = see also = = bridge scour flood control = = references = = = = external links = = u. s. army corps of engineers β civil works program river morphology and stream restoration references - wildland hydrology at the library of congress web archives ( archived 2002 - 08 - 13 )
water content and the internal evolution of terrestrial planets and icy bodies are closely linked. the distribution of water in planetary systems is controlled by the temperature structure in the protoplanetary disk and dynamics and migration of planetesimals and planetary embryos. this results in the formation of planetesimals and planetary embryos with a great variety of compositions, water contents and degrees of oxidation. the internal evolution and especially the formation time of planetesimals relative to the timescale of radiogenic heating by short - lived 26al decay may govern the amount of hydrous silicates and leftover rock - ice mixtures available in the late stages of their evolution. in turn, water content may affect the early internal evolution of the planetesimals and in particular metal - silicate separation processes. moreover, water content may contribute to an increase of oxygen fugacity and thus affect the concentrations of siderophile elements within the silicate reservoirs of solar system objects. finally, the water content strongly influences the differentiation rate of the icy moons, controls their internal evolution and governs the alteration processes occurring in their deep interiors.
the injuries of the inundations they have been designed to prevent, as the escape of floods from the raised river must occur sooner or later. inadequate planning controls which have permitted development on floodplains have been blamed for the flooding of domestic properties. channelization was done under the auspices or overall direction of engineers employed by the local authority or the national government. one of the most heavily channelized areas in the united states is west tennessee, where every major stream with one exception ( the hatchie river ) has been partially or completely channelized. channelization of a stream may be undertaken for several reasons. one is to make a stream more suitable for navigation or for navigation by larger vessels with deep draughts. another is to restrict water to a certain area of a stream ' s natural bottom lands so that the bulk of such lands can be made available for agriculture. a third reason is flood control, with the idea of giving a stream a sufficiently large and deep channel so that flooding beyond those limits will be minimal or nonexistent, at least on a routine basis. one major reason is to reduce natural erosion ; as a natural waterway curves back and forth, it usually deposits sand and gravel on the inside of the corners where the water flows slowly, and cuts sand, gravel, subsoil, and precious topsoil from the outside corners where it flows rapidly due to a change in direction. unlike sand and gravel, the topsoil that is eroded does not get deposited on the inside of the next corner of the river. it simply washes away. = = loss of wetlands = = channelization has several predictable and negative effects. one of them is loss of wetlands. wetlands are an excellent habitat for multiple forms of wildlife, and additionally serve as a " filter " for much of the world ' s surface fresh water. another is the fact that channelized streams are almost invariably straightened. for example, the channelization of florida ' s kissimmee river has been cited as a cause contributing to the loss of wetlands. this straightening causes the streams to flow more rapidly, which can, in some instances, vastly increase soil erosion. it can also increase flooding downstream from the channelized area, as larger volumes of water traveling more rapidly than normal can reach choke points over a shorter period of time than they otherwise would, with a net effect of flood control in one area coming at the expense of aggravated flooding in another. in addition, studies have shown that stream channelization results in declines of river fish populations. : 3 - 1ff a
) : the reason for the current medical visit. these are the symptoms. they are in the patient ' s own words and are recorded along with the duration of each one. also called chief concern or presenting complaint. current activity : occupation, hobbies, what the patient actually does. family history ( fh ) : listing of diseases in the family that may impact the patient. a family tree is sometimes used. history of present illness ( hpi ) : the chronological order of events of symptoms and further clarification of each symptom. distinguishable from history of previous illness, often called past medical history ( pmh ). medical history comprises hpi and pmh. medications ( rx ) : what drugs the patient takes including prescribed, over - the - counter, and home remedies, as well as alternative and herbal medicines or remedies. allergies are also recorded. past medical history ( pmh / pmhx ) : concurrent medical problems, past hospitalizations and operations, injuries, past infectious diseases or vaccinations, history of known allergies. review of systems ( ros ) or systems inquiry : a set of additional questions to ask, which may be missed on hpi : a general enquiry ( have you noticed any weight loss, change in sleep quality, fevers, lumps and bumps? etc. ), followed by questions on the body ' s main organ systems ( heart, lungs, digestive tract, urinary tract, etc. ). social history ( sh ) : birthplace, residences, marital history, social and economic status, habits ( including diet, medications, tobacco, alcohol ). the physical examination is the examination of the patient for medical signs of disease that are objective and observable, in contrast to symptoms that are volunteered by the patient and are not necessarily objectively observable. the healthcare provider uses sight, hearing, touch, and sometimes smell ( e. g., in infection, uremia, diabetic ketoacidosis ). four actions are the basis of physical examination : inspection, palpation ( feel ), percussion ( tap to determine resonance characteristics ), and auscultation ( listen ), generally in that order, although auscultation occurs prior to percussion and palpation for abdominal assessments. the clinical examination involves the study of : abdomen and rectum cardiovascular ( heart and blood vessels ) general appearance of the patient and specific indicators of disease ( nutritional status, presence of jaundice,
use less energy than conventional thermal separation processes such as distillation, sublimation or crystallization. the separation process is purely physical and both fractions ( permeate and retentate ) can be obtained as useful products. cold separation using membrane technology is widely used in the food technology, biotechnology and pharmaceutical industries. furthermore, using membranes enables separations to take place that would be impossible using thermal separation methods. for example, it is impossible to separate the constituents of azeotropic liquids or solutes which form isomorphic crystals by distillation or recrystallization but such separations can be achieved using membrane technology. depending on the type of membrane, the selective separation of certain individual substances or substance mixtures is possible. important technical applications include the production of drinking water by reverse osmosis. in waste water treatment, membrane technology is becoming increasingly important. ultra / microfiltration can be very effective in removing colloids and macromolecules from wastewater. this is needed if wastewater is discharged into sensitive waters especially those designated for contact water sports and recreation. about half of the market is in medical applications such as artificial kidneys to remove toxic substances by hemodialysis and as artificial lung for bubble - free supply of oxygen in the blood. the importance of membrane technology is growing in the field of environmental protection ( nano - mem - pro ippc database ). even in modern energy recovery techniques, membranes are increasingly used, for example in fuel cells and in osmotic power plants. = = mass transfer = = two basic models can be distinguished for mass transfer through the membrane : the solution - diffusion model and the hydrodynamic model. in real membranes, these two transport mechanisms certainly occur side by side, especially during ultra - filtration. = = = solution - diffusion model = = = in the solution - diffusion model, transport occurs only by diffusion. the component that needs to be transported must first be dissolved in the membrane. the general approach of the solution - diffusion model is to assume that the chemical potential of the feed and permeate fluids are in equilibrium with the adjacent membrane surfaces such that appropriate expressions for the chemical potential in the fluid and membrane phases can be equated at the solution - membrane interface. this principle is more important for dense membranes without natural pores such as those used for reverse osmosis and in fuel cells. during the filtration process a boundary layer forms on the membrane. this concentration gradient is created by molecules which cannot pass through the membrane. the
navigable capabilities of rivers. in tropical countries subject to periodical rains, the rivers are in flood during the rainy season and have hardly any flow during the rest of the year, while in temperate regions, where the rainfall is more evenly distributed throughout the year, evaporation causes the available rainfall to be much less in hot summer weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under ordinary conditions, most of the materials brought down from the high lands by torrential water courses are carried forward by the main river to the sea, or partially strewn over flat alluvial plains during floods ; the size of the materials forming the bed of the river or borne along by the stream is gradually reduced on proceeding sea
Question: Water is often put through a treatment system before it is used in homes as drinking water. Which is the most likely reason water is treated before people drink it?
A) to remove harmful substances
B) to add color
C) to increase its supply
D) to decrease its temperature
|
A) to remove harmful substances
|
Context:
the structural components of cells. as a by - product of photosynthesis, plants release oxygen into the atmosphere, a gas that is required by nearly all living things to carry out cellular respiration. in addition, they are influential in the global carbon and water cycles and plant roots bind and stabilise soils, preventing soil erosion. plants are crucial to the future of human society as they provide food, oxygen, biochemicals, and products for people, as well as creating and preserving soil. historically, all living things were classified as either animals or plants and botany covered the study of all organisms not considered animals. botanists examine both the internal functions and processes within plant organelles, cells, tissues, whole plants, plant populations and plant communities. at each of these levels, a botanist may be concerned with the classification ( taxonomy ), phylogeny and evolution, structure ( anatomy and morphology ), or function ( physiology ) of plant life. the strictest definition of " plant " includes only the " land plants " or embryophytes, which include seed plants ( gymnosperms, including the pines, and flowering plants ) and the free - sporing cryptogams including ferns, clubmosses, liverworts, hornworts and mosses. embryophytes are multicellular eukaryotes descended from an ancestor that obtained its energy from sunlight by photosynthesis. they have life cycles with alternating haploid and diploid phases. the sexual haploid phase of embryophytes, known as the gametophyte, nurtures the developing diploid embryo sporophyte within its tissues for at least part of its life, even in the seed plants, where the gametophyte itself is nurtured by its parent sporophyte. other groups of organisms that were previously studied by botanists include bacteria ( now studied in bacteriology ), fungi ( mycology ) β including lichen - forming fungi ( lichenology ), non - chlorophyte algae ( phycology ), and viruses ( virology ). however, attention is still given to these groups by botanists, and fungi ( including lichens ) and photosynthetic protists are usually covered in introductory botany courses. palaeobotanists study ancient plants in the fossil record to provide information about the evolutionary history of plants. cyanobacteria, the first oxygen - releasing photosynthetic organisms on earth, are thought to have given rise to the
on a large scale provided protection from insect pests or tolerance to herbicides. fungal and virus resistant crops have also been developed or are in development. this makes the insect and weed management of crops easier and can indirectly increase crop yield. gm crops that directly improve yield by accelerating growth or making the plant more hardy ( by improving salt, cold or drought tolerance ) are also under development. in 2016 salmon have been genetically modified with growth hormones to reach normal adult size much faster. gmos have been developed that modify the quality of produce by increasing the nutritional value or providing more industrially useful qualities or quantities. the amflora potato produces a more industrially useful blend of starches. soybeans and canola have been genetically modified to produce more healthy oils. the first commercialised gm food was a tomato that had delayed ripening, increasing its shelf life. plants and animals have been engineered to produce materials they do not normally make. pharming uses crops and animals as bioreactors to produce vaccines, drug intermediates, or the drugs themselves ; the useful product is purified from the harvest and then used in the standard pharmaceutical production process. cows and goats have been engineered to express drugs and other proteins in their milk, and in 2009 the fda approved a drug produced in goat milk. = = = other applications = = = genetic engineering has potential applications in conservation and natural area management. gene transfer through viral vectors has been proposed as a means of controlling invasive species as well as vaccinating threatened fauna from disease. transgenic trees have been suggested as a way to confer resistance to pathogens in wild populations. with the increasing risks of maladaptation in organisms as a result of climate change and other perturbations, facilitated adaptation through gene tweaking could be one solution to reducing extinction risks. applications of genetic engineering in conservation are thus far mostly theoretical and have yet to be put into practice. genetic engineering is also being used to create microbial art. some bacteria have been genetically engineered to create black and white photographs. novelty items such as lavender - colored carnations, blue roses, and glowing fish, have also been produced through genetic engineering. = = regulation = = the regulation of genetic engineering concerns the approaches taken by governments to assess and manage the risks associated with the development and release of gmos. the development of a regulatory framework began in 1975, at asilomar, california. the asilomar meeting recommended a set of voluntary guidelines regarding the use of recombinant technology. as the technology improved
pectin and xyloglucan from which the land plant cell wall is constructed. vascular land plants make lignin, a polymer used to strengthen the secondary cell walls of xylem tracheids and vessels to keep them from collapsing when a plant sucks water through them under water stress. lignin is also used in other cell types like sclerenchyma fibres that provide structural support for a plant and is a major constituent of wood. sporopollenin is a chemically resistant polymer found in the outer cell walls of spores and pollen of land plants responsible for the survival of early land plant spores and the pollen of seed plants in the fossil record. it is widely regarded as a marker for the start of land plant evolution during the ordovician period. the concentration of carbon dioxide in the atmosphere today is much lower than it was when plants emerged onto land during the ordovician and silurian periods. many monocots like maize and the pineapple and some dicots like the asteraceae have since independently evolved pathways like crassulacean acid metabolism and the c4 carbon fixation pathway for photosynthesis which avoid the losses resulting from photorespiration in the more common c3 carbon fixation pathway. these biochemical strategies are unique to land plants. = = = medicine and materials = = = phytochemistry is a branch of plant biochemistry primarily concerned with the chemical substances produced by plants during secondary metabolism. some of these compounds are toxins such as the alkaloid coniine from hemlock. others, such as the essential oils peppermint oil and lemon oil are useful for their aroma, as flavourings and spices ( e. g., capsaicin ), and in medicine as pharmaceuticals as in opium from opium poppies. many medicinal and recreational drugs, such as tetrahydrocannabinol ( active ingredient in cannabis ), caffeine, morphine and nicotine come directly from plants. others are simple derivatives of botanical natural products. for example, the pain killer aspirin is the acetyl ester of salicylic acid, originally isolated from the bark of willow trees, and a wide range of opiate painkillers like heroin are obtained by chemical modification of morphine obtained from the opium poppy. popular stimulants come from plants, such as caffeine from coffee, tea and chocolate, and nicotine from tobacco. most alcoholic beverages come from fermentation of carbohy
and their competitive or mutualistic interactions with other species. some ecologists even rely on empirical data from indigenous people that is gathered by ethnobotanists. this information can relay a great deal of information on how the land once was thousands of years ago and how it has changed over that time. the goals of plant ecology are to understand the causes of their distribution patterns, productivity, environmental impact, evolution, and responses to environmental change. plants depend on certain edaphic ( soil ) and climatic factors in their environment but can modify these factors too. for example, they can change their environment ' s albedo, increase runoff interception, stabilise mineral soils and develop their organic content, and affect local temperature. plants compete with other organisms in their ecosystem for resources. they interact with their neighbours at a variety of spatial scales in groups, populations and communities that collectively constitute vegetation. regions with characteristic vegetation types and dominant plants as well as similar abiotic and biotic factors, climate, and geography make up biomes like tundra or tropical rainforest. herbivores eat plants, but plants can defend themselves and some species are parasitic or even carnivorous. other organisms form mutually beneficial relationships with plants. for example, mycorrhizal fungi and rhizobia provide plants with nutrients in exchange for food, ants are recruited by ant plants to provide protection, honey bees, bats and other animals pollinate flowers and humans and other animals act as dispersal vectors to spread spores and seeds. = = = plants, climate and environmental change = = = plant responses to climate and other environmental changes can inform our understanding of how these changes affect ecosystem function and productivity. for example, plant phenology can be a useful proxy for temperature in historical climatology, and the biological impact of climate change and global warming. palynology, the analysis of fossil pollen deposits in sediments from thousands or millions of years ago allows the reconstruction of past climates. estimates of atmospheric co2 concentrations since the palaeozoic have been obtained from stomatal densities and the leaf shapes and sizes of ancient land plants. ozone depletion can expose plants to higher levels of ultraviolet radiation - b ( uv - b ), resulting in lower growth rates. moreover, information from studies of community ecology, plant systematics, and taxonomy is essential to understanding vegetation change, habitat destruction and species extinction. = = genetics = = inheritance in plants follows the same fundamental principles of genetics as in other multicellular organisms. gregor mendel discovered the genetic laws of inheritance by studying
soil erosion. plants are crucial to the future of human society as they provide food, oxygen, biochemicals, and products for people, as well as creating and preserving soil. historically, all living things were classified as either animals or plants and botany covered the study of all organisms not considered animals. botanists examine both the internal functions and processes within plant organelles, cells, tissues, whole plants, plant populations and plant communities. at each of these levels, a botanist may be concerned with the classification ( taxonomy ), phylogeny and evolution, structure ( anatomy and morphology ), or function ( physiology ) of plant life. the strictest definition of " plant " includes only the " land plants " or embryophytes, which include seed plants ( gymnosperms, including the pines, and flowering plants ) and the free - sporing cryptogams including ferns, clubmosses, liverworts, hornworts and mosses. embryophytes are multicellular eukaryotes descended from an ancestor that obtained its energy from sunlight by photosynthesis. they have life cycles with alternating haploid and diploid phases. the sexual haploid phase of embryophytes, known as the gametophyte, nurtures the developing diploid embryo sporophyte within its tissues for at least part of its life, even in the seed plants, where the gametophyte itself is nurtured by its parent sporophyte. other groups of organisms that were previously studied by botanists include bacteria ( now studied in bacteriology ), fungi ( mycology ) β including lichen - forming fungi ( lichenology ), non - chlorophyte algae ( phycology ), and viruses ( virology ). however, attention is still given to these groups by botanists, and fungi ( including lichens ) and photosynthetic protists are usually covered in introductory botany courses. palaeobotanists study ancient plants in the fossil record to provide information about the evolutionary history of plants. cyanobacteria, the first oxygen - releasing photosynthetic organisms on earth, are thought to have given rise to the ancestor of plants by entering into an endosymbiotic relationship with an early eukaryote, ultimately becoming the chloroplasts in plant cells. the new photosynthetic plants ( along with their algal relatives ) accelerated the rise in atmospheric oxygen started by the cyanobacteria, changing the
higher concentrations of atmospheric nitrous oxide ( n2o ) are expected to slightly warm earth ' s surface because of increases in radiative forcing. radiative forcing is the difference in the net upward thermal radiation flux from the earth through a transparent atmosphere and radiation through an otherwise identical atmosphere with greenhouse gases. radiative forcing, normally measured in w / m ^ 2, depends on latitude, longitude and altitude, but it is often quoted for the tropopause, about 11 km of altitude for temperate latitudes, or for the top of the atmosphere at around 90 km. for current concentrations of greenhouse gases, the radiative forcing per added n2o molecule is about 230 times larger than the forcing per added carbon dioxide ( co2 ) molecule. this is due to the heavy saturation of the absorption band of the relatively abundant greenhouse gas, co2, compared to the much smaller saturation of the absorption bands of the trace greenhouse gas n2o. but the rate of increase of co2 molecules, about 2. 5 ppm / year ( ppm = part per million by mole ), is about 3000 times larger than the rate of increase of n2o molecules, which has held steady at around 0. 00085 ppm / year since 1985. so, the contribution of nitrous oxide to the annual increase in forcing is 230 / 3000 or about 1 / 13 that of co2. if the main greenhouse gases, co2, ch4 and n2o have contributed about 0. 1 c / decade of the warming observed over the past few decades, this would correspond to about 0. 00064 k per year or 0. 064 k per century of warming from n2o. proposals to place harsh restrictions on nitrous oxide emissions because of warming fears are not justified by these facts. restrictions would cause serious harm ; for example, by jeopardizing world food supplies.
eat them. plants and other photosynthetic organisms are at the base of most food chains because they use the energy from the sun and nutrients from the soil and atmosphere, converting them into a form that can be used by animals. this is what ecologists call the first trophic level. the modern forms of the major staple foods, such as hemp, teff, maize, rice, wheat and other cereal grasses, pulses, bananas and plantains, as well as hemp, flax and cotton grown for their fibres, are the outcome of prehistoric selection over thousands of years from among wild ancestral plants with the most desirable characteristics. botanists study how plants produce food and how to increase yields, for example through plant breeding, making their work important to humanity ' s ability to feed the world and provide food security for future generations. botanists also study weeds, which are a considerable problem in agriculture, and the biology and control of plant pathogens in agriculture and natural ecosystems. ethnobotany is the study of the relationships between plants and people. when applied to the investigation of historical plant β people relationships ethnobotany may be referred to as archaeobotany or palaeoethnobotany. some of the earliest plant - people relationships arose between the indigenous people of canada in identifying edible plants from inedible plants. this relationship the indigenous people had with plants was recorded by ethnobotanists. = = plant biochemistry = = plant biochemistry is the study of the chemical processes used by plants. some of these processes are used in their primary metabolism like the photosynthetic calvin cycle and crassulacean acid metabolism. others make specialised materials like the cellulose and lignin used to build their bodies, and secondary products like resins and aroma compounds. plants and various other groups of photosynthetic eukaryotes collectively known as " algae " have unique organelles known as chloroplasts. chloroplasts are thought to be descended from cyanobacteria that formed endosymbiotic relationships with ancient plant and algal ancestors. chloroplasts and cyanobacteria contain the blue - green pigment chlorophyll a. chlorophyll a ( as well as its plant and green algal - specific cousin chlorophyll b ) absorbs light in the blue - violet and orange / red parts of the spectrum while reflecting and transmitting the green light that we see as the characteristic colour
sugar ) for export to the rest of the plant. unlike in animals ( which lack chloroplasts ), plants and their eukaryote relatives have delegated many biochemical roles to their chloroplasts, including synthesising all their fatty acids, and most amino acids. the fatty acids that chloroplasts make are used for many things, such as providing material to build cell membranes out of and making the polymer cutin which is found in the plant cuticle that protects land plants from drying out. plants synthesise a number of unique polymers like the polysaccharide molecules cellulose, pectin and xyloglucan from which the land plant cell wall is constructed. vascular land plants make lignin, a polymer used to strengthen the secondary cell walls of xylem tracheids and vessels to keep them from collapsing when a plant sucks water through them under water stress. lignin is also used in other cell types like sclerenchyma fibres that provide structural support for a plant and is a major constituent of wood. sporopollenin is a chemically resistant polymer found in the outer cell walls of spores and pollen of land plants responsible for the survival of early land plant spores and the pollen of seed plants in the fossil record. it is widely regarded as a marker for the start of land plant evolution during the ordovician period. the concentration of carbon dioxide in the atmosphere today is much lower than it was when plants emerged onto land during the ordovician and silurian periods. many monocots like maize and the pineapple and some dicots like the asteraceae have since independently evolved pathways like crassulacean acid metabolism and the c4 carbon fixation pathway for photosynthesis which avoid the losses resulting from photorespiration in the more common c3 carbon fixation pathway. these biochemical strategies are unique to land plants. = = = medicine and materials = = = phytochemistry is a branch of plant biochemistry primarily concerned with the chemical substances produced by plants during secondary metabolism. some of these compounds are toxins such as the alkaloid coniine from hemlock. others, such as the essential oils peppermint oil and lemon oil are useful for their aroma, as flavourings and spices ( e. g., capsaicin ), and in medicine as pharmaceuticals as in opium from opium poppies. many medicinal and recreational drugs, such as tetrahydrocannabino
venus flytrap and bladderworts, and the pollinia of orchids. the hypothesis that plant growth and development is coordinated by plant hormones or plant growth regulators first emerged in the late 19th century. darwin experimented on the movements of plant shoots and roots towards light and gravity, and concluded " it is hardly an exaggeration to say that the tip of the radicle.. acts like the brain of one of the lower animals.. directing the several movements ". about the same time, the role of auxins ( from the greek auxein, to grow ) in control of plant growth was first outlined by the dutch scientist frits went. the first known auxin, indole - 3 - acetic acid ( iaa ), which promotes cell growth, was only isolated from plants about 50 years later. this compound mediates the tropic responses of shoots and roots towards light and gravity. the finding in 1939 that plant callus could be maintained in culture containing iaa, followed by the observation in 1947 that it could be induced to form roots and shoots by controlling the concentration of growth hormones were key steps in the development of plant biotechnology and genetic modification. cytokinins are a class of plant hormones named for their control of cell division ( especially cytokinesis ). the natural cytokinin zeatin was discovered in corn, zea mays, and is a derivative of the purine adenine. zeatin is produced in roots and transported to shoots in the xylem where it promotes cell division, bud development, and the greening of chloroplasts. the gibberelins, such as gibberelic acid are diterpenes synthesised from acetyl coa via the mevalonate pathway. they are involved in the promotion of germination and dormancy - breaking in seeds, in regulation of plant height by controlling stem elongation and the control of flowering. abscisic acid ( aba ) occurs in all land plants except liverworts, and is synthesised from carotenoids in the chloroplasts and other plastids. it inhibits cell division, promotes seed maturation, and dormancy, and promotes stomatal closure. it was so named because it was originally thought to control abscission. ethylene is a gaseous hormone that is produced in all higher plant tissues from methionine. it is now known to be the hormone that stimulates or regulates fruit ripening and abscission,
energy they need to exist. plants, algae and cyanobacteria are the major groups of organisms that carry out photosynthesis, a process that uses the energy of sunlight to convert water and carbon dioxide into sugars that can be used both as a source of chemical energy and of organic molecules that are used in the structural components of cells. as a by - product of photosynthesis, plants release oxygen into the atmosphere, a gas that is required by nearly all living things to carry out cellular respiration. in addition, they are influential in the global carbon and water cycles and plant roots bind and stabilise soils, preventing soil erosion. plants are crucial to the future of human society as they provide food, oxygen, biochemicals, and products for people, as well as creating and preserving soil. historically, all living things were classified as either animals or plants and botany covered the study of all organisms not considered animals. botanists examine both the internal functions and processes within plant organelles, cells, tissues, whole plants, plant populations and plant communities. at each of these levels, a botanist may be concerned with the classification ( taxonomy ), phylogeny and evolution, structure ( anatomy and morphology ), or function ( physiology ) of plant life. the strictest definition of " plant " includes only the " land plants " or embryophytes, which include seed plants ( gymnosperms, including the pines, and flowering plants ) and the free - sporing cryptogams including ferns, clubmosses, liverworts, hornworts and mosses. embryophytes are multicellular eukaryotes descended from an ancestor that obtained its energy from sunlight by photosynthesis. they have life cycles with alternating haploid and diploid phases. the sexual haploid phase of embryophytes, known as the gametophyte, nurtures the developing diploid embryo sporophyte within its tissues for at least part of its life, even in the seed plants, where the gametophyte itself is nurtured by its parent sporophyte. other groups of organisms that were previously studied by botanists include bacteria ( now studied in bacteriology ), fungi ( mycology ) β including lichen - forming fungi ( lichenology ), non - chlorophyte algae ( phycology ), and viruses ( virology ). however, attention is still given to these groups by botanists, and fungi ( including lichens ) and photos
Question: How can plants most likely improve air quality?
A) by absorbing nutrients
B) by producing oxygen
C) by producing sugar
D) by absorbing water
|
B) by producing oxygen
|
Context:
, subsequent switching to inbreeding becomes disadvantageous since it allows expression of the previously masked deleterious recessive mutations, commonly referred to as inbreeding depression. unlike in higher animals, where parthenogenesis is rare, asexual reproduction may occur in plants by several different mechanisms. the formation of stem tubers in potato is one example. particularly in arctic or alpine habitats, where opportunities for fertilisation of flowers by animals are rare, plantlets or bulbs, may develop instead of flowers, replacing sexual reproduction with asexual reproduction and giving rise to clonal populations genetically identical to the parent. this is one of several types of apomixis that occur in plants. apomixis can also happen in a seed, producing a seed that contains an embryo genetically identical to the parent. most sexually reproducing organisms are diploid, with paired chromosomes, but doubling of their chromosome number may occur due to errors in cytokinesis. this can occur early in development to produce an autopolyploid or partly autopolyploid organism, or during normal processes of cellular differentiation to produce some cell types that are polyploid ( endopolyploidy ), or during gamete formation. an allopolyploid plant may result from a hybridisation event between two different species. both autopolyploid and allopolyploid plants can often reproduce normally, but may be unable to cross - breed successfully with the parent population because there is a mismatch in chromosome numbers. these plants that are reproductively isolated from the parent species but live within the same geographical area, may be sufficiently successful to form a new species. some otherwise sterile plant polyploids can still reproduce vegetatively or by seed apomixis, forming clonal populations of identical individuals. durum wheat is a fertile tetraploid allopolyploid, while bread wheat is a fertile hexaploid. the commercial banana is an example of a sterile, seedless triploid hybrid. common dandelion is a triploid that produces viable seeds by apomictic seed. as in other eukaryotes, the inheritance of endosymbiotic organelles like mitochondria and chloroplasts in plants is non - mendelian. chloroplasts are inherited through the male parent in gymnosperms but often through the female parent in flowering plants. = = = molecular genetics = = = a considerable amount of new knowledge about plant function comes from
the broad definition of " utilizing a biotechnological system to make products ". indeed, the cultivation of plants may be viewed as the earliest biotechnological enterprise. agriculture has been theorized to have become the dominant way of producing food since the neolithic revolution. through early biotechnology, the earliest farmers selected and bred the best - suited crops ( e. g., those with the highest yields ) to produce enough food to support a growing population. as crops and fields became increasingly large and difficult to maintain, it was discovered that specific organisms and their by - products could effectively fertilize, restore nitrogen, and control pests. throughout the history of agriculture, farmers have inadvertently altered the genetics of their crops through introducing them to new environments and breeding them with other plants β one of the first forms of biotechnology. these processes also were included in early fermentation of beer. these processes were introduced in early mesopotamia, egypt, china and india, and still use the same basic biological methods. in brewing, malted grains ( containing enzymes ) convert starch from grains into sugar and then adding specific yeasts to produce beer. in this process, carbohydrates in the grains broke down into alcohols, such as ethanol. later, other cultures produced the process of lactic acid fermentation, which produced other preserved foods, such as soy sauce. fermentation was also used in this time period to produce leavened bread. although the process of fermentation was not fully understood until louis pasteur ' s work in 1857, it is still the first use of biotechnology to convert a food source into another form. before the time of charles darwin ' s work and life, animal and plant scientists had already used selective breeding. darwin added to that body of work with his scientific observations about the ability of science to change species. these accounts contributed to darwin ' s theory of natural selection. for thousands of years, humans have used selective breeding to improve the production of crops and livestock to use them for food. in selective breeding, organisms with desirable characteristics are mated to produce offspring with the same characteristics. for example, this technique was used with corn to produce the largest and sweetest crops. in the early twentieth century scientists gained a greater understanding of microbiology and explored ways of manufacturing specific products. in 1917, chaim weizmann first used a pure microbiological culture in an industrial process, that of manufacturing corn starch using clostridium acetobutylicum, to produce acetone, which the united
ranks varying from family to subgenus have terms for their study, including agrostology ( or graminology ) for the study of grasses, synantherology for the study of composites, and batology for the study of brambles. study can also be divided by guild rather than clade or grade. for example, dendrology is the study of woody plants. many divisions of biology have botanical subfields. these are commonly denoted by prefixing the word plant ( e. g. plant taxonomy, plant ecology, plant anatomy, plant morphology, plant systematics ), or prefixing or substituting the prefix phyto - ( e. g. phytochemistry, phytogeography ). the study of fossil plants is called palaeobotany. other fields are denoted by adding or substituting the word botany ( e. g. systematic botany ). phytosociology is a subfield of plant ecology that classifies and studies communities of plants. the intersection of fields from the above pair of categories gives rise to fields such as bryogeography, the study of the distribution of mosses. different parts of plants also give rise to their own subfields, including xylology, carpology ( or fructology ), and palynology, these being the study of wood, fruit and pollen / spores respectively. botany also overlaps on the one hand with agriculture, horticulture and silviculture, and on the other hand with medicine and pharmacology, giving rise to fields such as agronomy, horticultural botany, phytopathology, and phytopharmacology. = = scope and importance = = the study of plants is vital because they underpin almost all animal life on earth by generating a large proportion of the oxygen and food that provide humans and other organisms with aerobic respiration with the chemical energy they need to exist. plants, algae and cyanobacteria are the major groups of organisms that carry out photosynthesis, a process that uses the energy of sunlight to convert water and carbon dioxide into sugars that can be used both as a source of chemical energy and of organic molecules that are used in the structural components of cells. as a by - product of photosynthesis, plants release oxygen into the atmosphere, a gas that is required by nearly all living things to carry out cellular respiration. in addition, they are influential in the global carbon and water cycles and plant roots bind and stabilise soils, preventing
. an example of this would be bt corn. whether or not green biotechnology products such as this are ultimately more environmentally friendly is a topic of considerable debate. it is commonly considered as the next phase of green revolution, which can be seen as a platform to eradicate world hunger by using technologies which enable the production of more fertile and resistant, towards biotic and abiotic stress, plants and ensures application of environmentally friendly fertilizers and the use of biopesticides, it is mainly focused on the development of agriculture. on the other hand, some of the uses of green biotechnology involve microorganisms to clean and reduce waste. red biotechnology is the use of biotechnology in the medical and pharmaceutical industries, and health preservation. this branch involves the production of vaccines and antibiotics, regenerative therapies, creation of artificial organs and new diagnostics of diseases. as well as the development of hormones, stem cells, antibodies, sirna and diagnostic tests. white biotechnology, also known as industrial biotechnology, is biotechnology applied to industrial processes. an example is the designing of an organism to produce a useful chemical. another example is the using of enzymes as industrial catalysts to either produce valuable chemicals or destroy hazardous / polluting chemicals. white biotechnology tends to consume less in resources than traditional processes used to produce industrial goods. yellow biotechnology refers to the use of biotechnology in food production ( food industry ), for example in making wine ( winemaking ), cheese ( cheesemaking ), and beer ( brewing ) by fermentation. it has also been used to refer to biotechnology applied to insects. this includes biotechnology - based approaches for the control of harmful insects, the characterisation and utilisation of active ingredients or genes of insects for research, or application in agriculture and medicine and various other approaches. gray biotechnology is dedicated to environmental applications, and focused on the maintenance of biodiversity and the remotion of pollutants. brown biotechnology is related to the management of arid lands and deserts. one application is the creation of enhanced seeds that resist extreme environmental conditions of arid regions, which is related to the innovation, creation of agriculture techniques and management of resources. violet biotechnology is related to law, ethical and philosophical issues around biotechnology. microbial biotechnology has been proposed for the rapidly emerging area of biotechnology applications in space and microgravity ( space bioeconomy ) dark biotechnology is the color associated with bioterrorism or biological weapons and biowarfare which uses microorganisms, and toxins to cause diseases and death in humans, livestock and
fertile and resistant, towards biotic and abiotic stress, plants and ensures application of environmentally friendly fertilizers and the use of biopesticides, it is mainly focused on the development of agriculture. on the other hand, some of the uses of green biotechnology involve microorganisms to clean and reduce waste. red biotechnology is the use of biotechnology in the medical and pharmaceutical industries, and health preservation. this branch involves the production of vaccines and antibiotics, regenerative therapies, creation of artificial organs and new diagnostics of diseases. as well as the development of hormones, stem cells, antibodies, sirna and diagnostic tests. white biotechnology, also known as industrial biotechnology, is biotechnology applied to industrial processes. an example is the designing of an organism to produce a useful chemical. another example is the using of enzymes as industrial catalysts to either produce valuable chemicals or destroy hazardous / polluting chemicals. white biotechnology tends to consume less in resources than traditional processes used to produce industrial goods. yellow biotechnology refers to the use of biotechnology in food production ( food industry ), for example in making wine ( winemaking ), cheese ( cheesemaking ), and beer ( brewing ) by fermentation. it has also been used to refer to biotechnology applied to insects. this includes biotechnology - based approaches for the control of harmful insects, the characterisation and utilisation of active ingredients or genes of insects for research, or application in agriculture and medicine and various other approaches. gray biotechnology is dedicated to environmental applications, and focused on the maintenance of biodiversity and the remotion of pollutants. brown biotechnology is related to the management of arid lands and deserts. one application is the creation of enhanced seeds that resist extreme environmental conditions of arid regions, which is related to the innovation, creation of agriculture techniques and management of resources. violet biotechnology is related to law, ethical and philosophical issues around biotechnology. microbial biotechnology has been proposed for the rapidly emerging area of biotechnology applications in space and microgravity ( space bioeconomy ) dark biotechnology is the color associated with bioterrorism or biological weapons and biowarfare which uses microorganisms, and toxins to cause diseases and death in humans, livestock and crops. = = = medicine = = = in medicine, modern biotechnology has many applications in areas such as pharmaceutical drug discoveries and production, pharmacogenomics, and genetic testing ( or genetic screening ). in 2021, nearly 40 % of the total company value of pharmaceutical biotech companies worldwide were active in oncology
often injurious, at least with the plants on which i experimented. " an important adaptive benefit of outcrossing is that it allows the masking of deleterious mutations in the genome of progeny. this beneficial effect is also known as hybrid vigor or heterosis. once outcrossing is established, subsequent switching to inbreeding becomes disadvantageous since it allows expression of the previously masked deleterious recessive mutations, commonly referred to as inbreeding depression. unlike in higher animals, where parthenogenesis is rare, asexual reproduction may occur in plants by several different mechanisms. the formation of stem tubers in potato is one example. particularly in arctic or alpine habitats, where opportunities for fertilisation of flowers by animals are rare, plantlets or bulbs, may develop instead of flowers, replacing sexual reproduction with asexual reproduction and giving rise to clonal populations genetically identical to the parent. this is one of several types of apomixis that occur in plants. apomixis can also happen in a seed, producing a seed that contains an embryo genetically identical to the parent. most sexually reproducing organisms are diploid, with paired chromosomes, but doubling of their chromosome number may occur due to errors in cytokinesis. this can occur early in development to produce an autopolyploid or partly autopolyploid organism, or during normal processes of cellular differentiation to produce some cell types that are polyploid ( endopolyploidy ), or during gamete formation. an allopolyploid plant may result from a hybridisation event between two different species. both autopolyploid and allopolyploid plants can often reproduce normally, but may be unable to cross - breed successfully with the parent population because there is a mismatch in chromosome numbers. these plants that are reproductively isolated from the parent species but live within the same geographical area, may be sufficiently successful to form a new species. some otherwise sterile plant polyploids can still reproduce vegetatively or by seed apomixis, forming clonal populations of identical individuals. durum wheat is a fertile tetraploid allopolyploid, while bread wheat is a fertile hexaploid. the commercial banana is an example of a sterile, seedless triploid hybrid. common dandelion is a triploid that produces viable seeds by apomictic seed. as in other eukaryotes, the inheritance of endosymbiotic organelles like
and measuring radiation levels. the surveyor program conducted uncrewed lunar landings and takeoffs, as well as taking surface and regolith observations. despite the setback caused by the apollo 1 fire, which killed three astronauts, the program proceeded. apollo 8 was the first crewed spacecraft to leave low earth orbit and the first human spaceflight to reach the moon. the crew orbited the moon ten times on december 24 and 25, 1968, and then traveled safely back to earth. the three apollo 8 astronauts β frank borman, james lovell, and william anders β were the first humans to see the earth as a globe in space, the first to witness an earthrise, and the first to see and manually photograph the far side of the moon. the first lunar landing was conducted by apollo 11. commanded by neil armstrong with astronauts buzz aldrin and michael collins, apollo 11 was one of the most significant missions in nasa ' s history, marking the end of the space race when the soviet union gave up its lunar ambitions. as the first human to step on the surface of the moon, neil armstrong uttered the now famous words : that ' s one small step for man, one giant leap for mankind. nasa would conduct six total lunar landings as part of the apollo program, with apollo 17 concluding the program in 1972. = = = = end of apollo = = = = wernher von braun had advocated for nasa to develop a space station since the agency was created. in 1973, following the end of the apollo lunar missions, nasa launched its first space station, skylab, on the final launch of the saturn v. skylab reused a significant amount of apollo and saturn hardware, with a repurposed saturn v third stage serving as the primary module for the space station. damage to skylab during its launch required spacewalks to be performed by the first crew to make it habitable and operational. skylab hosted nine missions and was decommissioned in 1974 and deorbited in 1979, two years prior to the first launch of the space shuttle and any possibility of boosting its orbit. in 1975, the apollo β soyuz mission was the first ever international spaceflight and a major diplomatic accomplishment between the cold war rivals, which also marked the last flight of the apollo capsule. flown in 1975, a us apollo spacecraft docked with a soviet soyuz capsule. = = = interplanetary exploration and space science = = = during the 1960s, nasa started its space science and interplanetary probe program. the mariner program was its flagship
the best - suited crops ( e. g., those with the highest yields ) to produce enough food to support a growing population. as crops and fields became increasingly large and difficult to maintain, it was discovered that specific organisms and their by - products could effectively fertilize, restore nitrogen, and control pests. throughout the history of agriculture, farmers have inadvertently altered the genetics of their crops through introducing them to new environments and breeding them with other plants β one of the first forms of biotechnology. these processes also were included in early fermentation of beer. these processes were introduced in early mesopotamia, egypt, china and india, and still use the same basic biological methods. in brewing, malted grains ( containing enzymes ) convert starch from grains into sugar and then adding specific yeasts to produce beer. in this process, carbohydrates in the grains broke down into alcohols, such as ethanol. later, other cultures produced the process of lactic acid fermentation, which produced other preserved foods, such as soy sauce. fermentation was also used in this time period to produce leavened bread. although the process of fermentation was not fully understood until louis pasteur ' s work in 1857, it is still the first use of biotechnology to convert a food source into another form. before the time of charles darwin ' s work and life, animal and plant scientists had already used selective breeding. darwin added to that body of work with his scientific observations about the ability of science to change species. these accounts contributed to darwin ' s theory of natural selection. for thousands of years, humans have used selective breeding to improve the production of crops and livestock to use them for food. in selective breeding, organisms with desirable characteristics are mated to produce offspring with the same characteristics. for example, this technique was used with corn to produce the largest and sweetest crops. in the early twentieth century scientists gained a greater understanding of microbiology and explored ways of manufacturing specific products. in 1917, chaim weizmann first used a pure microbiological culture in an industrial process, that of manufacturing corn starch using clostridium acetobutylicum, to produce acetone, which the united kingdom desperately needed to manufacture explosives during world war i. biotechnology has also led to the development of antibiotics. in 1928, alexander fleming discovered the mold penicillium. his work led to the purification of the antibiotic formed by the mold by howard florey, ernst boris chain and norman heatley β to form
this extra strength allows some structural components to fail without bridge collapse. the extra strength used in the design is called the margin of safety. eyes and ears provide working examples of passive redundancy. vision loss in one eye does not cause blindness but depth perception is impaired. hearing loss in one ear does not cause deafness but directionality is lost. performance decline is commonly associated with passive redundancy when a limited number of failures occur. active redundancy eliminates performance declines by monitoring the performance of individual devices, and this monitoring is used in voting logic. the voting logic is linked to switching that automatically reconfigures the components. error detection and correction and the global positioning system ( gps ) are two examples of active redundancy. electrical power distribution provides an example of active redundancy. several power lines connect each generation facility with customers. each power line includes monitors that detect overload. each power line also includes circuit breakers. the combination of power lines provides excess capacity. circuit breakers disconnect a power line when monitors detect an overload. power is redistributed across the remaining lines. at the toronto airport, there are 4 redundant electrical lines. each of the 4 lines supply enough power for the entire airport. a spot network substation uses reverse current relays to open breakers to lines that fail, but lets power continue to flow the airport. electrical power systems use power scheduling to reconfigure active redundancy. computing systems adjust the production output of each generating facility when other generating facilities are suddenly lost. this prevents blackout conditions during major events such as an earthquake. = = disadvantages = = charles perrow, author of normal accidents, has said that sometimes redundancies backfire and produce less, not more reliability. this may happen in three ways : first, redundant safety devices result in a more complex system, more prone to errors and accidents. second, redundancy may lead to shirking of responsibility among workers. third, redundancy may lead to increased production pressures, resulting in a system that operates at higher speeds, but less safely. = = voting logic = = voting logic uses performance monitoring to determine how to reconfigure individual components so that operation continues without violating specification limitations of the overall system. voting logic often involves computers, but systems composed of items other than computers may be reconfigured using voting logic. circuit breakers are an example of a form of non - computer voting logic. the simplest voting logic in computing systems involves two components :
the gas giant planets in the solar system have a retinue of icy moons, and we expect giant exoplanets to have similar satellite systems. if a jupiter - like planet were to migrate toward its parent star the icy moons orbiting it would evaporate, creating atmospheres and possible habitable surface oceans. here, we examine how long the surface ice and possible oceans would last before being hydrodynamically lost to space. the hydrodynamic loss rate from the moons is determined, in large part, by the stellar flux available for absorption, which increases as the giant planet and icy moons migrate closer to the star. at some planet - star distance the stellar flux incident on the icy moons becomes so great that they enter a runaway greenhouse state. this runaway greenhouse state rapidly transfers all available surface water to the atmosphere as vapor, where it is easily lost from the small moons. however, for icy moons of ganymede ' s size around a sun - like star we found that surface water ( either ice or liquid ) can persist indefinitely outside the runaway greenhouse orbital distance. in contrast, the surface water on smaller moons of europa ' s size will only persist on timescales greater than 1 gyr at distances ranging 1. 49 to 0. 74 au around a sun - like star for bond albedos of 0. 2 and 0. 8, where the lower albedo becomes relevant if ice melts. consequently, small moons can lose their icy shells, which would create a torus of h atoms around their host planet that might be detectable in future observations.
Question: A condition in bees called "colony collapse disorder" occurs when worker bees leave the hive to collect nectar and never return. How would a decline in the bee population most likely affect a farm ecosystem?
A) There would be a decrease in soil thickness.
B) There would be a decline in fruit production.
C) There would be an increase in plant diseases.
D) There would be an increase in insect pest populations.
|
B) There would be a decline in fruit production.
|
Context:
is the genus, and columbianum the specific epithet. the combination is the name of the species. when writing the scientific name of an organism, it is proper to capitalise the first letter in the genus and put all of the specific epithet in lowercase. additionally, the entire term is ordinarily italicised ( or underlined when italics are not available ). the evolutionary relationships and heredity of a group of organisms is called its phylogeny. phylogenetic studies attempt to discover phylogenies. the basic approach is to use similarities based on shared inheritance to determine relationships. as an example, species of pereskia are trees or bushes with prominent leaves. they do not obviously resemble a typical leafless cactus such as an echinocactus. however, both pereskia and echinocactus have spines produced from areoles ( highly specialised pad - like structures ) suggesting that the two genera are indeed related. judging relationships based on shared characters requires care, since plants may resemble one another through convergent evolution in which characters have arisen independently. some euphorbias have leafless, rounded bodies adapted to water conservation similar to those of globular cacti, but characters such as the structure of their flowers make it clear that the two groups are not closely related. the cladistic method takes a systematic approach to characters, distinguishing between those that carry no information about shared evolutionary history β such as those evolved separately in different groups ( homoplasies ) or those left over from ancestors ( plesiomorphies ) β and derived characters, which have been passed down from innovations in a shared ancestor ( apomorphies ). only derived characters, such as the spine - producing areoles of cacti, provide evidence for descent from a common ancestor. the results of cladistic analyses are expressed as cladograms : tree - like diagrams showing the pattern of evolutionary branching and descent. from the 1990s onwards, the predominant approach to constructing phylogenies for living plants has been molecular phylogenetics, which uses molecular characters, particularly dna sequences, rather than morphological characters like the presence or absence of spines and areoles. the difference is that the genetic code itself is used to decide evolutionary relationships, instead of being used indirectly via the characters it gives rise to. clive stace describes this as having " direct access to the genetic basis of evolution. " as a simple example, prior to the use of genetic evidence, fungi were thought either to be plants or to be more closely related to plants
participates as a consumer, resource, or both in consumer β resource interactions, which form the core of food chains or food webs. there are different trophic levels within any food web, with the lowest level being the primary producers ( or autotrophs ) such as plants and algae that convert energy and inorganic material into organic compounds, which can then be used by the rest of the community. at the next level are the heterotrophs, which are the species that obtain energy by breaking apart organic compounds from other organisms. heterotrophs that consume plants are primary consumers ( or herbivores ) whereas heterotrophs that consume herbivores are secondary consumers ( or carnivores ). and those that eat secondary consumers are tertiary consumers and so on. omnivorous heterotrophs are able to consume at multiple levels. finally, there are decomposers that feed on the waste products or dead bodies of organisms. on average, the total amount of energy incorporated into the biomass of a trophic level per unit of time is about one - tenth of the energy of the trophic level that it consumes. waste and dead material used by decomposers as well as heat lost from metabolism make up the other ninety percent of energy that is not consumed by the next trophic level. = = = biosphere = = = in the global ecosystem or biosphere, matter exists as different interacting compartments, which can be biotic or abiotic as well as accessible or inaccessible, depending on their forms and locations. for example, matter from terrestrial autotrophs are both biotic and accessible to other organisms whereas the matter in rocks and minerals are abiotic and inaccessible. a biogeochemical cycle is a pathway by which specific elements of matter are turned over or moved through the biotic ( biosphere ) and the abiotic ( lithosphere, atmosphere, and hydrosphere ) compartments of earth. there are biogeochemical cycles for nitrogen, carbon, and water. = = = conservation = = = conservation biology is the study of the conservation of earth ' s biodiversity with the aim of protecting species, their habitats, and ecosystems from excessive rates of extinction and the erosion of biotic interactions. it is concerned with factors that influence the maintenance, loss, and restoration of biodiversity and the science of sustaining evolutionary processes that engender genetic, population, species, and ecosystem diversity. the concern stems from estimates suggesting that up to 50 % of all species on the planet
by ancestry rather than superficial characteristics. while scientists do not always agree on how to classify organisms, molecular phylogenetics, which uses dna sequences as data, has driven many recent revisions along evolutionary lines and is likely to continue to do so. the dominant classification system is called linnaean taxonomy. it includes ranks and binomial nomenclature. the nomenclature of botanical organisms is codified in the international code of nomenclature for algae, fungi, and plants ( icn ) and administered by the international botanical congress. kingdom plantae belongs to domain eukaryota and is broken down recursively until each species is separately classified. the order is : kingdom ; phylum ( or division ) ; class ; order ; family ; genus ( plural genera ) ; species. the scientific name of a plant represents its genus and its species within the genus, resulting in a single worldwide name for each organism. for example, the tiger lily is lilium columbianum. lilium is the genus, and columbianum the specific epithet. the combination is the name of the species. when writing the scientific name of an organism, it is proper to capitalise the first letter in the genus and put all of the specific epithet in lowercase. additionally, the entire term is ordinarily italicised ( or underlined when italics are not available ). the evolutionary relationships and heredity of a group of organisms is called its phylogeny. phylogenetic studies attempt to discover phylogenies. the basic approach is to use similarities based on shared inheritance to determine relationships. as an example, species of pereskia are trees or bushes with prominent leaves. they do not obviously resemble a typical leafless cactus such as an echinocactus. however, both pereskia and echinocactus have spines produced from areoles ( highly specialised pad - like structures ) suggesting that the two genera are indeed related. judging relationships based on shared characters requires care, since plants may resemble one another through convergent evolution in which characters have arisen independently. some euphorbias have leafless, rounded bodies adapted to water conservation similar to those of globular cacti, but characters such as the structure of their flowers make it clear that the two groups are not closely related. the cladistic method takes a systematic approach to characters, distinguishing between those that carry no information about shared evolutionary history β such as those evolved separately in different groups ( homoplasies ) or those left over from ancestors ( plesiomorphies ) β and derived characters, which
kingdom ; phylum ( or division ) ; class ; order ; family ; genus ( plural genera ) ; species. the scientific name of a plant represents its genus and its species within the genus, resulting in a single worldwide name for each organism. for example, the tiger lily is lilium columbianum. lilium is the genus, and columbianum the specific epithet. the combination is the name of the species. when writing the scientific name of an organism, it is proper to capitalise the first letter in the genus and put all of the specific epithet in lowercase. additionally, the entire term is ordinarily italicised ( or underlined when italics are not available ). the evolutionary relationships and heredity of a group of organisms is called its phylogeny. phylogenetic studies attempt to discover phylogenies. the basic approach is to use similarities based on shared inheritance to determine relationships. as an example, species of pereskia are trees or bushes with prominent leaves. they do not obviously resemble a typical leafless cactus such as an echinocactus. however, both pereskia and echinocactus have spines produced from areoles ( highly specialised pad - like structures ) suggesting that the two genera are indeed related. judging relationships based on shared characters requires care, since plants may resemble one another through convergent evolution in which characters have arisen independently. some euphorbias have leafless, rounded bodies adapted to water conservation similar to those of globular cacti, but characters such as the structure of their flowers make it clear that the two groups are not closely related. the cladistic method takes a systematic approach to characters, distinguishing between those that carry no information about shared evolutionary history β such as those evolved separately in different groups ( homoplasies ) or those left over from ancestors ( plesiomorphies ) β and derived characters, which have been passed down from innovations in a shared ancestor ( apomorphies ). only derived characters, such as the spine - producing areoles of cacti, provide evidence for descent from a common ancestor. the results of cladistic analyses are expressed as cladograms : tree - like diagrams showing the pattern of evolutionary branching and descent. from the 1990s onwards, the predominant approach to constructing phylogenies for living plants has been molecular phylogenetics, which uses molecular characters, particularly dna sequences, rather than morphological characters like the presence or absence of spines and areoles. the difference is that the genetic code itself is used
by which botanists group organisms into categories such as genera or species. biological classification is a form of scientific taxonomy. modern taxonomy is rooted in the work of carl linnaeus, who grouped species according to shared physical characteristics. these groupings have since been revised to align better with the darwinian principle of common descent β grouping organisms by ancestry rather than superficial characteristics. while scientists do not always agree on how to classify organisms, molecular phylogenetics, which uses dna sequences as data, has driven many recent revisions along evolutionary lines and is likely to continue to do so. the dominant classification system is called linnaean taxonomy. it includes ranks and binomial nomenclature. the nomenclature of botanical organisms is codified in the international code of nomenclature for algae, fungi, and plants ( icn ) and administered by the international botanical congress. kingdom plantae belongs to domain eukaryota and is broken down recursively until each species is separately classified. the order is : kingdom ; phylum ( or division ) ; class ; order ; family ; genus ( plural genera ) ; species. the scientific name of a plant represents its genus and its species within the genus, resulting in a single worldwide name for each organism. for example, the tiger lily is lilium columbianum. lilium is the genus, and columbianum the specific epithet. the combination is the name of the species. when writing the scientific name of an organism, it is proper to capitalise the first letter in the genus and put all of the specific epithet in lowercase. additionally, the entire term is ordinarily italicised ( or underlined when italics are not available ). the evolutionary relationships and heredity of a group of organisms is called its phylogeny. phylogenetic studies attempt to discover phylogenies. the basic approach is to use similarities based on shared inheritance to determine relationships. as an example, species of pereskia are trees or bushes with prominent leaves. they do not obviously resemble a typical leafless cactus such as an echinocactus. however, both pereskia and echinocactus have spines produced from areoles ( highly specialised pad - like structures ) suggesting that the two genera are indeed related. judging relationships based on shared characters requires care, since plants may resemble one another through convergent evolution in which characters have arisen independently. some euphorbias have leafless, rounded bodies adapted to water conservation similar to those of globular cacti, but characters such as the structure of their flowers make it clear that the
species occupying the same geographical area at the same time. a biological interaction is the effect that a pair of organisms living together in a community have on each other. they can be either of the same species ( intraspecific interactions ), or of different species ( interspecific interactions ). these effects may be short - term, like pollination and predation, or long - term ; both often strongly influence the evolution of the species involved. a long - term interaction is called a symbiosis. symbioses range from mutualism, beneficial to both partners, to competition, harmful to both partners. every species participates as a consumer, resource, or both in consumer β resource interactions, which form the core of food chains or food webs. there are different trophic levels within any food web, with the lowest level being the primary producers ( or autotrophs ) such as plants and algae that convert energy and inorganic material into organic compounds, which can then be used by the rest of the community. at the next level are the heterotrophs, which are the species that obtain energy by breaking apart organic compounds from other organisms. heterotrophs that consume plants are primary consumers ( or herbivores ) whereas heterotrophs that consume herbivores are secondary consumers ( or carnivores ). and those that eat secondary consumers are tertiary consumers and so on. omnivorous heterotrophs are able to consume at multiple levels. finally, there are decomposers that feed on the waste products or dead bodies of organisms. on average, the total amount of energy incorporated into the biomass of a trophic level per unit of time is about one - tenth of the energy of the trophic level that it consumes. waste and dead material used by decomposers as well as heat lost from metabolism make up the other ninety percent of energy that is not consumed by the next trophic level. = = = biosphere = = = in the global ecosystem or biosphere, matter exists as different interacting compartments, which can be biotic or abiotic as well as accessible or inaccessible, depending on their forms and locations. for example, matter from terrestrial autotrophs are both biotic and accessible to other organisms whereas the matter in rocks and minerals are abiotic and inaccessible. a biogeochemical cycle is a pathway by which specific elements of matter are turned over or moved through the biotic ( biosphere ) and the abiotic ( lithos
anticommutative engel algebras of the first five degeneration levels are classified. all algebras appearing in this classification are nilpotent malcev algebras.
and their competitive or mutualistic interactions with other species. some ecologists even rely on empirical data from indigenous people that is gathered by ethnobotanists. this information can relay a great deal of information on how the land once was thousands of years ago and how it has changed over that time. the goals of plant ecology are to understand the causes of their distribution patterns, productivity, environmental impact, evolution, and responses to environmental change. plants depend on certain edaphic ( soil ) and climatic factors in their environment but can modify these factors too. for example, they can change their environment ' s albedo, increase runoff interception, stabilise mineral soils and develop their organic content, and affect local temperature. plants compete with other organisms in their ecosystem for resources. they interact with their neighbours at a variety of spatial scales in groups, populations and communities that collectively constitute vegetation. regions with characteristic vegetation types and dominant plants as well as similar abiotic and biotic factors, climate, and geography make up biomes like tundra or tropical rainforest. herbivores eat plants, but plants can defend themselves and some species are parasitic or even carnivorous. other organisms form mutually beneficial relationships with plants. for example, mycorrhizal fungi and rhizobia provide plants with nutrients in exchange for food, ants are recruited by ant plants to provide protection, honey bees, bats and other animals pollinate flowers and humans and other animals act as dispersal vectors to spread spores and seeds. = = = plants, climate and environmental change = = = plant responses to climate and other environmental changes can inform our understanding of how these changes affect ecosystem function and productivity. for example, plant phenology can be a useful proxy for temperature in historical climatology, and the biological impact of climate change and global warming. palynology, the analysis of fossil pollen deposits in sediments from thousands or millions of years ago allows the reconstruction of past climates. estimates of atmospheric co2 concentrations since the palaeozoic have been obtained from stomatal densities and the leaf shapes and sizes of ancient land plants. ozone depletion can expose plants to higher levels of ultraviolet radiation - b ( uv - b ), resulting in lower growth rates. moreover, information from studies of community ecology, plant systematics, and taxonomy is essential to understanding vegetation change, habitat destruction and species extinction. = = genetics = = inheritance in plants follows the same fundamental principles of genetics as in other multicellular organisms. gregor mendel discovered the genetic laws of inheritance by studying
##al nomenclature. the nomenclature of botanical organisms is codified in the international code of nomenclature for algae, fungi, and plants ( icn ) and administered by the international botanical congress. kingdom plantae belongs to domain eukaryota and is broken down recursively until each species is separately classified. the order is : kingdom ; phylum ( or division ) ; class ; order ; family ; genus ( plural genera ) ; species. the scientific name of a plant represents its genus and its species within the genus, resulting in a single worldwide name for each organism. for example, the tiger lily is lilium columbianum. lilium is the genus, and columbianum the specific epithet. the combination is the name of the species. when writing the scientific name of an organism, it is proper to capitalise the first letter in the genus and put all of the specific epithet in lowercase. additionally, the entire term is ordinarily italicised ( or underlined when italics are not available ). the evolutionary relationships and heredity of a group of organisms is called its phylogeny. phylogenetic studies attempt to discover phylogenies. the basic approach is to use similarities based on shared inheritance to determine relationships. as an example, species of pereskia are trees or bushes with prominent leaves. they do not obviously resemble a typical leafless cactus such as an echinocactus. however, both pereskia and echinocactus have spines produced from areoles ( highly specialised pad - like structures ) suggesting that the two genera are indeed related. judging relationships based on shared characters requires care, since plants may resemble one another through convergent evolution in which characters have arisen independently. some euphorbias have leafless, rounded bodies adapted to water conservation similar to those of globular cacti, but characters such as the structure of their flowers make it clear that the two groups are not closely related. the cladistic method takes a systematic approach to characters, distinguishing between those that carry no information about shared evolutionary history β such as those evolved separately in different groups ( homoplasies ) or those left over from ancestors ( plesiomorphies ) β and derived characters, which have been passed down from innovations in a shared ancestor ( apomorphies ). only derived characters, such as the spine - producing areoles of cacti, provide evidence for descent from a common ancestor. the results of cladistic analyses are expressed as cladograms : tree - like diagrams showing the
aquatic and most of the aquatic photosynthetic eukaryotic organisms are collectively described as algae, which is a term of convenience as not all algae are closely related. algae comprise several distinct clades such as glaucophytes, which are microscopic freshwater algae that may have resembled in form to the early unicellular ancestor of plantae. unlike glaucophytes, the other algal clades such as red and green algae are multicellular. green algae comprise three major clades : chlorophytes, coleochaetophytes, and stoneworts. fungi are eukaryotes that digest foods outside their bodies, secreting digestive enzymes that break down large food molecules before absorbing them through their cell membranes. many fungi are also saprobes, feeding on dead organic matter, making them important decomposers in ecological systems. animals are multicellular eukaryotes. with few exceptions, animals consume organic material, breathe oxygen, are able to move, can reproduce sexually, and grow from a hollow sphere of cells, the blastula, during embryonic development. over 1. 5 million living animal species have been described β of which around 1 million are insects β but it has been estimated there are over 7 million animal species in total. they have complex interactions with each other and their environments, forming intricate food webs. = = = viruses = = = viruses are submicroscopic infectious agents that replicate inside the cells of organisms. viruses infect all types of life forms, from animals and plants to microorganisms, including bacteria and archaea. more than 6, 000 virus species have been described in detail. viruses are found in almost every ecosystem on earth and are the most numerous type of biological entity. the origins of viruses in the evolutionary history of life are unclear : some may have evolved from plasmids β pieces of dna that can move between cells β while others may have evolved from bacteria. in evolution, viruses are an important means of horizontal gene transfer, which increases genetic diversity in a way analogous to sexual reproduction. because viruses possess some but not all characteristics of life, they have been described as " organisms at the edge of life ", and as self - replicators. = = ecology = = ecology is the study of the distribution and abundance of life, the interaction between organisms and their environment. = = = ecosystems = = = the community of living ( biotic ) organisms in conjunction with the nonliving ( abiotic ) components ( e.
Question: How are animals that prey on other animals classified?
A) herbivores
B) carnivores
C) autotrophs
D) decomposers
|
B) carnivores
|
Context:
use less energy than conventional thermal separation processes such as distillation, sublimation or crystallization. the separation process is purely physical and both fractions ( permeate and retentate ) can be obtained as useful products. cold separation using membrane technology is widely used in the food technology, biotechnology and pharmaceutical industries. furthermore, using membranes enables separations to take place that would be impossible using thermal separation methods. for example, it is impossible to separate the constituents of azeotropic liquids or solutes which form isomorphic crystals by distillation or recrystallization but such separations can be achieved using membrane technology. depending on the type of membrane, the selective separation of certain individual substances or substance mixtures is possible. important technical applications include the production of drinking water by reverse osmosis. in waste water treatment, membrane technology is becoming increasingly important. ultra / microfiltration can be very effective in removing colloids and macromolecules from wastewater. this is needed if wastewater is discharged into sensitive waters especially those designated for contact water sports and recreation. about half of the market is in medical applications such as artificial kidneys to remove toxic substances by hemodialysis and as artificial lung for bubble - free supply of oxygen in the blood. the importance of membrane technology is growing in the field of environmental protection ( nano - mem - pro ippc database ). even in modern energy recovery techniques, membranes are increasingly used, for example in fuel cells and in osmotic power plants. = = mass transfer = = two basic models can be distinguished for mass transfer through the membrane : the solution - diffusion model and the hydrodynamic model. in real membranes, these two transport mechanisms certainly occur side by side, especially during ultra - filtration. = = = solution - diffusion model = = = in the solution - diffusion model, transport occurs only by diffusion. the component that needs to be transported must first be dissolved in the membrane. the general approach of the solution - diffusion model is to assume that the chemical potential of the feed and permeate fluids are in equilibrium with the adjacent membrane surfaces such that appropriate expressions for the chemical potential in the fluid and membrane phases can be equated at the solution - membrane interface. this principle is more important for dense membranes without natural pores such as those used for reverse osmosis and in fuel cells. during the filtration process a boundary layer forms on the membrane. this concentration gradient is created by molecules which cannot pass through the membrane. the
and still use the same basic biological methods. in brewing, malted grains ( containing enzymes ) convert starch from grains into sugar and then adding specific yeasts to produce beer. in this process, carbohydrates in the grains broke down into alcohols, such as ethanol. later, other cultures produced the process of lactic acid fermentation, which produced other preserved foods, such as soy sauce. fermentation was also used in this time period to produce leavened bread. although the process of fermentation was not fully understood until louis pasteur ' s work in 1857, it is still the first use of biotechnology to convert a food source into another form. before the time of charles darwin ' s work and life, animal and plant scientists had already used selective breeding. darwin added to that body of work with his scientific observations about the ability of science to change species. these accounts contributed to darwin ' s theory of natural selection. for thousands of years, humans have used selective breeding to improve the production of crops and livestock to use them for food. in selective breeding, organisms with desirable characteristics are mated to produce offspring with the same characteristics. for example, this technique was used with corn to produce the largest and sweetest crops. in the early twentieth century scientists gained a greater understanding of microbiology and explored ways of manufacturing specific products. in 1917, chaim weizmann first used a pure microbiological culture in an industrial process, that of manufacturing corn starch using clostridium acetobutylicum, to produce acetone, which the united kingdom desperately needed to manufacture explosives during world war i. biotechnology has also led to the development of antibiotics. in 1928, alexander fleming discovered the mold penicillium. his work led to the purification of the antibiotic formed by the mold by howard florey, ernst boris chain and norman heatley β to form what we today know as penicillin. in 1940, penicillin became available for medicinal use to treat bacterial infections in humans. the field of modern biotechnology is generally thought of as having been born in 1971 when paul berg ' s ( stanford ) experiments in gene splicing had early success. herbert w. boyer ( univ. calif. at san francisco ) and stanley n. cohen ( stanford ) significantly advanced the new technology in 1972 by transferring genetic material into a bacterium, such that the imported material would be reproduced. the commercial viability of a biotechnology industry was significantly expanded on june 16, 1980, when the united states
. throughout the history of agriculture, farmers have inadvertently altered the genetics of their crops through introducing them to new environments and breeding them with other plants β one of the first forms of biotechnology. these processes also were included in early fermentation of beer. these processes were introduced in early mesopotamia, egypt, china and india, and still use the same basic biological methods. in brewing, malted grains ( containing enzymes ) convert starch from grains into sugar and then adding specific yeasts to produce beer. in this process, carbohydrates in the grains broke down into alcohols, such as ethanol. later, other cultures produced the process of lactic acid fermentation, which produced other preserved foods, such as soy sauce. fermentation was also used in this time period to produce leavened bread. although the process of fermentation was not fully understood until louis pasteur ' s work in 1857, it is still the first use of biotechnology to convert a food source into another form. before the time of charles darwin ' s work and life, animal and plant scientists had already used selective breeding. darwin added to that body of work with his scientific observations about the ability of science to change species. these accounts contributed to darwin ' s theory of natural selection. for thousands of years, humans have used selective breeding to improve the production of crops and livestock to use them for food. in selective breeding, organisms with desirable characteristics are mated to produce offspring with the same characteristics. for example, this technique was used with corn to produce the largest and sweetest crops. in the early twentieth century scientists gained a greater understanding of microbiology and explored ways of manufacturing specific products. in 1917, chaim weizmann first used a pure microbiological culture in an industrial process, that of manufacturing corn starch using clostridium acetobutylicum, to produce acetone, which the united kingdom desperately needed to manufacture explosives during world war i. biotechnology has also led to the development of antibiotics. in 1928, alexander fleming discovered the mold penicillium. his work led to the purification of the antibiotic formed by the mold by howard florey, ernst boris chain and norman heatley β to form what we today know as penicillin. in 1940, penicillin became available for medicinal use to treat bacterial infections in humans. the field of modern biotechnology is generally thought of as having been born in 1971 when paul berg ' s ( stanford ) experiments in gene splicing had early success. herbert w. boyer
to dye denim and the artist ' s pigments gamboge and rose madder. sugar, starch, cotton, linen, hemp, some types of rope, wood and particle boards, papyrus and paper, vegetable oils, wax, and natural rubber are examples of commercially important materials made from plant tissues or their secondary products. charcoal, a pure form of carbon made by pyrolysis of wood, has a long history as a metal - smelting fuel, as a filter material and adsorbent and as an artist ' s material and is one of the three ingredients of gunpowder. cellulose, the world ' s most abundant organic polymer, can be converted into energy, fuels, materials and chemical feedstock. products made from cellulose include rayon and cellophane, wallpaper paste, biobutanol and gun cotton. sugarcane, rapeseed and soy are some of the plants with a highly fermentable sugar or oil content that are used as sources of biofuels, important alternatives to fossil fuels, such as biodiesel. sweetgrass was used by native americans to ward off bugs like mosquitoes. these bug repelling properties of sweetgrass were later found by the american chemical society in the molecules phytol and coumarin. = = plant ecology = = plant ecology is the science of the functional relationships between plants and their habitats β the environments where they complete their life cycles. plant ecologists study the composition of local and regional floras, their biodiversity, genetic diversity and fitness, the adaptation of plants to their environment, and their competitive or mutualistic interactions with other species. some ecologists even rely on empirical data from indigenous people that is gathered by ethnobotanists. this information can relay a great deal of information on how the land once was thousands of years ago and how it has changed over that time. the goals of plant ecology are to understand the causes of their distribution patterns, productivity, environmental impact, evolution, and responses to environmental change. plants depend on certain edaphic ( soil ) and climatic factors in their environment but can modify these factors too. for example, they can change their environment ' s albedo, increase runoff interception, stabilise mineral soils and develop their organic content, and affect local temperature. plants compete with other organisms in their ecosystem for resources. they interact with their neighbours at a variety of spatial scales in groups, populations and communities that collectively constitute vegetation. regions with characteristic vegetation types and dominant plants as well as similar abiot
which applies a forces that results in fracturing ), and impact ( which employs a milling medium or the particles themselves to cause fracturing ). attrition milling equipment includes the wet scrubber ( also called the planetary mill or wet attrition mill ), which has paddles in water creating vortexes in which the material collides and break up. compression mills include the jaw crusher, roller crusher and cone crusher. impact mills include the ball mill, which has media that tumble and fracture the material, or the resonantacoustic mixer. shaft impactors cause particle - to particle attrition and compression. batching is the process of weighing the oxides according to recipes, and preparing them for mixing and drying. mixing occurs after batching and is performed with various machines, such as dry mixing ribbon mixers ( a type of cement mixer ), resonantacoustic mixers, mueller mixers, and pug mills. wet mixing generally involves the same equipment. forming is making the mixed material into shapes, ranging from toilet bowls to spark plug insulators. forming can involve : ( 1 ) extrusion, such as extruding " slugs " to make bricks, ( 2 ) pressing to make shaped parts, ( 3 ) slip casting, as in making toilet bowls, wash basins and ornamentals like ceramic statues. forming produces a " green " part, ready for drying. green parts are soft, pliable, and over time will lose shape. handling the green product will change its shape. for example, a green brick can be " squeezed ", and after squeezing it will stay that way. drying is removing the water or binder from the formed material. spray drying is widely used to prepare powder for pressing operations. other dryers are tunnel dryers and periodic dryers. controlled heat is applied in this two - stage process. first, heat removes water. this step needs careful control, as rapid heating causes cracks and surface defects. the dried part is smaller than the green part, and is brittle, necessitating careful handling, since a small impact will cause crumbling and breaking. sintering is where the dried parts pass through a controlled heating process, and the oxides are chemically changed to cause bonding and densification. the fired part will be smaller than the dried part. = = forming methods = = ceramic forming techniques include throwing, slipcasting, tape casting, freeze - casting, injection molding, dry pressing, isostatic pressing, hot isostatic pressing
water, and used in the gristmilling and sugarcane industries. sugar mills first appeared in the medieval islamic world. they were first driven by watermills, and then windmills from the 9th and 10th centuries in what are today afghanistan, pakistan and iran. crops such as almonds and citrus fruit were brought to europe through al - andalus, and sugar cultivation was gradually adopted across europe. arab merchants dominated trade in the indian ocean until the arrival of the portuguese in the 16th century. the muslim world adopted papermaking from china. the earliest paper mills appeared in abbasid - era baghdad during 794 β 795. the knowledge of gunpowder was also transmitted from china via predominantly islamic countries, where formulas for pure potassium nitrate were developed. the spinning wheel was invented in the islamic world by the early 11th century. it was later widely adopted in europe, where it was adapted into the spinning jenny, a key device during the industrial revolution. the crankshaft was invented by al - jazari in 1206, and is central to modern machinery such as the steam engine, internal combustion engine and automatic controls. the camshaft was also first described by al - jazari in 1206. early programmable machines were also invented in the muslim world. the first music sequencer, a programmable musical instrument, was an automated flute player invented by the banu musa brothers, described in their book of ingenious devices, in the 9th century. in 1206, al - jazari invented programmable automata / robots. he described four automaton musicians, including two drummers operated by a programmable drum machine, where the drummer could be made to play different rhythms and different drum patterns. the castle clock, a hydropowered mechanical astronomical clock invented by al - jazari, was an early programmable analog computer. in the ottoman empire, a practical impulse steam turbine was invented in 1551 by taqi ad - din muhammad ibn ma ' ruf in ottoman egypt. he described a method for rotating a spit by means of a jet of steam playing on rotary vanes around the periphery of a wheel. known as a steam jack, a similar device for rotating a spit was also later described by john wilkins in 1648. = = = = medieval europe = = = = while medieval technology has been long depicted as a step backward in the evolution of western technology, a generation of medievalists ( like the american historian of science lynn white ) stressed from the 1940s onwards the innovative character of many medieval techniques. genuine medieval contributions include
power to watermills and water - raising machines. many of these technologies were transferred to medieval europe. wind - powered machines used to grind grain and pump water, the windmill and wind pump, first appeared in what are now iran, afghanistan and pakistan by the 9th century. they were used to grind grains and draw up water, and used in the gristmilling and sugarcane industries. sugar mills first appeared in the medieval islamic world. they were first driven by watermills, and then windmills from the 9th and 10th centuries in what are today afghanistan, pakistan and iran. crops such as almonds and citrus fruit were brought to europe through al - andalus, and sugar cultivation was gradually adopted across europe. arab merchants dominated trade in the indian ocean until the arrival of the portuguese in the 16th century. the muslim world adopted papermaking from china. the earliest paper mills appeared in abbasid - era baghdad during 794 β 795. the knowledge of gunpowder was also transmitted from china via predominantly islamic countries, where formulas for pure potassium nitrate were developed. the spinning wheel was invented in the islamic world by the early 11th century. it was later widely adopted in europe, where it was adapted into the spinning jenny, a key device during the industrial revolution. the crankshaft was invented by al - jazari in 1206, and is central to modern machinery such as the steam engine, internal combustion engine and automatic controls. the camshaft was also first described by al - jazari in 1206. early programmable machines were also invented in the muslim world. the first music sequencer, a programmable musical instrument, was an automated flute player invented by the banu musa brothers, described in their book of ingenious devices, in the 9th century. in 1206, al - jazari invented programmable automata / robots. he described four automaton musicians, including two drummers operated by a programmable drum machine, where the drummer could be made to play different rhythms and different drum patterns. the castle clock, a hydropowered mechanical astronomical clock invented by al - jazari, was an early programmable analog computer. in the ottoman empire, a practical impulse steam turbine was invented in 1551 by taqi ad - din muhammad ibn ma ' ruf in ottoman egypt. he described a method for rotating a spit by means of a jet of steam playing on rotary vanes around the periphery of a wheel. known as a steam jack, a similar device for rotating a spit was also later described by john
is opened and the dna is purified. the gene is separated by using restriction enzymes to cut the dna into fragments or polymerase chain reaction ( pcr ) to amplify up the gene segment. these segments can then be extracted through gel electrophoresis. if the chosen gene or the donor organism ' s genome has been well studied it may already be accessible from a genetic library. if the dna sequence is known, but no copies of the gene are available, it can also be artificially synthesised. once isolated the gene is ligated into a plasmid that is then inserted into a bacterium. the plasmid is replicated when the bacteria divide, ensuring unlimited copies of the gene are available. the rk2 plasmid is notable for its ability to replicate in a wide variety of single - celled organisms, which makes it suitable as a genetic engineering tool. before the gene is inserted into the target organism it must be combined with other genetic elements. these include a promoter and terminator region, which initiate and end transcription. a selectable marker gene is added, which in most cases confers antibiotic resistance, so researchers can easily determine which cells have been successfully transformed. the gene can also be modified at this stage for better expression or effectiveness. these manipulations are carried out using recombinant dna techniques, such as restriction digests, ligations and molecular cloning. = = = inserting dna into the host genome = = = there are a number of techniques used to insert genetic material into the host genome. some bacteria can naturally take up foreign dna. this ability can be induced in other bacteria via stress ( e. g. thermal or electric shock ), which increases the cell membrane ' s permeability to dna ; up - taken dna can either integrate with the genome or exist as extrachromosomal dna. dna is generally inserted into animal cells using microinjection, where it can be injected through the cell ' s nuclear envelope directly into the nucleus, or through the use of viral vectors. plant genomes can be engineered by physical methods or by use of agrobacterium for the delivery of sequences hosted in t - dna binary vectors. in plants the dna is often inserted using agrobacterium - mediated transformation, taking advantage of the agrobacteriums t - dna sequence that allows natural insertion of genetic material into plant cells. other methods include biolistics, where particles of gold or tungsten are coated with dna and then shot into
with the lambda virus. as well as inserting genes, the process can be used to remove, or " knock out ", genes. the new dna can be inserted randomly, or targeted to a specific part of the genome. an organism that is generated through genetic engineering is considered to be genetically modified ( gm ) and the resulting entity is a genetically modified organism ( gmo ). the first gmo was a bacterium generated by herbert boyer and stanley cohen in 1973. rudolf jaenisch created the first gm animal when he inserted foreign dna into a mouse in 1974. the first company to focus on genetic engineering, genentech, was founded in 1976 and started the production of human proteins. genetically engineered human insulin was produced in 1978 and insulin - producing bacteria were commercialised in 1982. genetically modified food has been sold since 1994, with the release of the flavr savr tomato. the flavr savr was engineered to have a longer shelf life, but most current gm crops are modified to increase resistance to insects and herbicides. glofish, the first gmo designed as a pet, was sold in the united states in december 2003. in 2016 salmon modified with a growth hormone were sold. genetic engineering has been applied in numerous fields including research, medicine, industrial biotechnology and agriculture. in research, gmos are used to study gene function and expression through loss of function, gain of function, tracking and expression experiments. by knocking out genes responsible for certain conditions it is possible to create animal model organisms of human diseases. as well as producing hormones, vaccines and other drugs, genetic engineering has the potential to cure genetic diseases through gene therapy. chinese hamster ovary ( cho ) cells are used in industrial genetic engineering. additionally mrna vaccines are made through genetic engineering to prevent infections by viruses such as covid - 19. the same techniques that are used to produce drugs can also have industrial applications such as producing enzymes for laundry detergent, cheeses and other products. the rise of commercialised genetically modified crops has provided economic benefit to farmers in many different countries, but has also been the source of most of the controversy surrounding the technology. this has been present since its early use ; the first field trials were destroyed by anti - gm activists. although there is a scientific consensus that currently available food derived from gm crops poses no greater risk to human health than conventional food, critics consider gm food safety a leading concern. gene flow, impact on non - target organisms, control of the food supply and intellectual property rights have also been raised as potential
temperature changes up to 1000 Β°c. = = processing steps = = the traditional ceramic process generally follows this sequence : milling β batching β mixing β forming β drying β firing β assembly. milling is the process by which materials are reduced from a large size to a smaller size. milling may involve breaking up cemented material ( in which case individual particles retain their shape ) or pulverization ( which involves grinding the particles themselves to a smaller size ). milling is generally done by mechanical means, including attrition ( which is particle - to - particle collision that results in agglomerate break up or particle shearing ), compression ( which applies a forces that results in fracturing ), and impact ( which employs a milling medium or the particles themselves to cause fracturing ). attrition milling equipment includes the wet scrubber ( also called the planetary mill or wet attrition mill ), which has paddles in water creating vortexes in which the material collides and break up. compression mills include the jaw crusher, roller crusher and cone crusher. impact mills include the ball mill, which has media that tumble and fracture the material, or the resonantacoustic mixer. shaft impactors cause particle - to particle attrition and compression. batching is the process of weighing the oxides according to recipes, and preparing them for mixing and drying. mixing occurs after batching and is performed with various machines, such as dry mixing ribbon mixers ( a type of cement mixer ), resonantacoustic mixers, mueller mixers, and pug mills. wet mixing generally involves the same equipment. forming is making the mixed material into shapes, ranging from toilet bowls to spark plug insulators. forming can involve : ( 1 ) extrusion, such as extruding " slugs " to make bricks, ( 2 ) pressing to make shaped parts, ( 3 ) slip casting, as in making toilet bowls, wash basins and ornamentals like ceramic statues. forming produces a " green " part, ready for drying. green parts are soft, pliable, and over time will lose shape. handling the green product will change its shape. for example, a green brick can be " squeezed ", and after squeezing it will stay that way. drying is removing the water or binder from the formed material. spray drying is widely used to prepare powder for pressing operations. other dryers are tunnel dryers and periodic dryers. controlled heat is applied in this two - stage process. first,
Question: Which process will separate the sugar from a mixture of sugar and water?
A) Stirring the mixture
B) Evaporating the water
C) Adding more water
D) Adding more sugar
|
B) Evaporating the water
|
Context:
are the stone - paved streets of the city - state of ur, dating to c. 4, 000 bce, and timber roads leading through the swamps of glastonbury, england, dating to around the same period. the first long - distance road, which came into use around 3, 500 bce, spanned 2, 400 km from the persian gulf to the mediterranean sea, but was not paved and was only partially maintained. in around 2, 000 bce, the minoans on the greek island of crete built a 50 km road leading from the palace of gortyn on the south side of the island, through the mountains, to the palace of knossos on the north side of the island. unlike the earlier road, the minoan road was completely paved. ancient minoan private homes had running water. a bathtub virtually identical to modern ones was unearthed at the palace of knossos. several minoan private homes also had toilets, which could be flushed by pouring water down the drain. the ancient romans had many public flush toilets, which emptied into an extensive sewage system. the primary sewer in rome was the cloaca maxima ; construction began on it in the sixth century bce and it is still in use today. the ancient romans also had a complex system of aqueducts, which were used to transport water across long distances. the first roman aqueduct was built in 312 bce. the eleventh and final ancient roman aqueduct was built in 226 ce. put together, the roman aqueducts extended over 450 km, but less than 70 km of this was above ground and supported by arches. = = = pre - modern = = = innovations continued through the middle ages with the introduction of silk production ( in asia and later europe ), the horse collar, and horseshoes. simple machines ( such as the lever, the screw, and the pulley ) were combined into more complicated tools, such as the wheelbarrow, windmills, and clocks. a system of universities developed and spread scientific ideas and practices, including oxford and cambridge. the renaissance era produced many innovations, including the introduction of the movable type printing press to europe, which facilitated the communication of knowledge. technology became increasingly influenced by science, beginning a cycle of mutual advancement. = = = modern = = = starting in the united kingdom in the 18th century, the discovery of steam power set off the industrial revolution, which saw wide - ranging technological discoveries, particularly in the areas of agriculture, manufacturing, mining, metallurgy, and transport, and the
depends on the extent of the continent in which it is situated, its position in relation to the hilly regions in which rivers generally arise and the sea into which they flow, and the distance between the source and the outlet into the sea of the river draining it. the rate of flow of rivers depends mainly upon their fall, also known as the gradient or slope. when two rivers of different sizes have the same fall, the larger river has the quicker flow, as its retardation by friction against its bed and banks is less in proportion to its volume than is the case with the smaller river. the fall available in a section of a river approximately corresponds to the slope of the country it traverses ; as rivers rise close to the highest part of their basins, generally in hilly regions, their fall is rapid near their source and gradually diminishes, with occasional irregularities, until, in traversing plains along the latter part of their course, their fall usually becomes quite gentle. accordingly, in large basins, rivers in most cases begin as torrents with a variable flow, and end as gently flowing rivers with a comparatively regular discharge. the irregular flow of rivers throughout their course forms one of the main difficulties in devising works for mitigating inundations or for increasing the navigable capabilities of rivers. in tropical countries subject to periodical rains, the rivers are in flood during the rainy season and have hardly any flow during the rest of the year, while in temperate regions, where the rainfall is more evenly distributed throughout the year, evaporation causes the available rainfall to be much less in hot summer weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform
made of steel. the shoe is generally wider than the caisson to reduce friction, and the leading edge may be supplied with pressurised bentonite slurry, which swells in water, stabilizing settlement by filling depressions and voids. an open caisson may fill with water during sinking. the material is excavated by clamshell excavator bucket on crane. the formation level subsoil may still not be suitable for excavation or bearing capacity. the water in the caisson ( due to a high water table ) balances the upthrust forces of the soft soils underneath. if dewatered, the base may " pipe " or " boil ", causing the caisson to sink. to combat this problem, piles may be driven from the surface to act as : load - bearing walls, in that they transmit loads to deeper soils. anchors, in that they resist flotation because of the friction at the interface between their surfaces and the surrounding earth into which they have been driven. h - beam sections ( typical column sections, due to resistance to bending in all axis ) may be driven at angles " raked " to rock or other firmer soils ; the h - beams are left extended above the base. a reinforced concrete plug may be placed under the water, a process known as tremie concrete placement. when the caisson is dewatered, this plug acts as a pile cap, resisting the upward forces of the subsoil. = = = monolithic = = = a monolithic caisson ( or simply a monolith ) is larger than the other types of caisson, but similar to open caissons. such caissons are often found in quay walls, where resistance to impact from ships is required. = = = pneumatic = = = shallow caissons may be open to the air, whereas pneumatic caissons ( sometimes called pressurized caissons ), which penetrate soft mud, are bottomless boxes sealed at the top and filled with compressed air to keep water and mud out at depth. an airlock allows access to the chamber. workers, called sandhogs in american english, move mud and rock debris ( called muck ) from the edge of the workspace to a water - filled pit, connected by a tube ( called the muck tube ) to the surface. a crane at the surface removes the soil with a clamshell bucket. the water pressure in the tube balances the air pressure, with excess air escaping up
from the insignificant drainage areas of streams rising on high ground near the coast and flowing straight down into the sea, up to immense tracts of continents, where rivers rising on the slopes of mountain ranges far inland have to traverse vast stretches of valleys and plains before reaching the ocean. the size of the largest river basin of any country depends on the extent of the continent in which it is situated, its position in relation to the hilly regions in which rivers generally arise and the sea into which they flow, and the distance between the source and the outlet into the sea of the river draining it. the rate of flow of rivers depends mainly upon their fall, also known as the gradient or slope. when two rivers of different sizes have the same fall, the larger river has the quicker flow, as its retardation by friction against its bed and banks is less in proportion to its volume than is the case with the smaller river. the fall available in a section of a river approximately corresponds to the slope of the country it traverses ; as rivers rise close to the highest part of their basins, generally in hilly regions, their fall is rapid near their source and gradually diminishes, with occasional irregularities, until, in traversing plains along the latter part of their course, their fall usually becomes quite gentle. accordingly, in large basins, rivers in most cases begin as torrents with a variable flow, and end as gently flowing rivers with a comparatively regular discharge. the irregular flow of rivers throughout their course forms one of the main difficulties in devising works for mitigating inundations or for increasing the navigable capabilities of rivers. in tropical countries subject to periodical rains, the rivers are in flood during the rainy season and have hardly any flow during the rest of the year, while in temperate regions, where the rainfall is more evenly distributed throughout the year, evaporation causes the available rainfall to be much less in hot summer weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their
also known as the gradient or slope. when two rivers of different sizes have the same fall, the larger river has the quicker flow, as its retardation by friction against its bed and banks is less in proportion to its volume than is the case with the smaller river. the fall available in a section of a river approximately corresponds to the slope of the country it traverses ; as rivers rise close to the highest part of their basins, generally in hilly regions, their fall is rapid near their source and gradually diminishes, with occasional irregularities, until, in traversing plains along the latter part of their course, their fall usually becomes quite gentle. accordingly, in large basins, rivers in most cases begin as torrents with a variable flow, and end as gently flowing rivers with a comparatively regular discharge. the irregular flow of rivers throughout their course forms one of the main difficulties in devising works for mitigating inundations or for increasing the navigable capabilities of rivers. in tropical countries subject to periodical rains, the rivers are in flood during the rainy season and have hardly any flow during the rest of the year, while in temperate regions, where the rainfall is more evenly distributed throughout the year, evaporation causes the available rainfall to be much less in hot summer weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in
the world is changing at an ever - increasing pace. and it has changed in a much more fundamental way than one would think, primarily because it has become more connected and interdependent than in our entire history. every new product, every new invention can be combined with those that existed before, thereby creating an explosion of complexity : structural complexity, dynamic complexity, functional complexity, and algorithmic complexity. how to respond to this challenge? and what are the costs?
##s ( sometimes called pressurized caissons ), which penetrate soft mud, are bottomless boxes sealed at the top and filled with compressed air to keep water and mud out at depth. an airlock allows access to the chamber. workers, called sandhogs in american english, move mud and rock debris ( called muck ) from the edge of the workspace to a water - filled pit, connected by a tube ( called the muck tube ) to the surface. a crane at the surface removes the soil with a clamshell bucket. the water pressure in the tube balances the air pressure, with excess air escaping up the muck tube. the pressurized air flow must be constant to ensure regular air changes for the workers and prevent excessive inflow of mud or water at the base of the caisson. when the caisson hits bedrock, the sandhogs exit through the airlock and fill the box with concrete, forming a solid foundation pier. a pneumatic ( compressed - air ) caisson has the advantage of providing dry working conditions, which is better for placing concrete. it is also well suited for foundations for which other methods might cause settlement of adjacent structures. construction workers who leave the pressurized environment of the caisson must decompress at a rate that allows symptom - free release of inert gases dissolved in the body tissues if they are to avoid decompression sickness, a condition first identified in caisson workers, and originally named " caisson disease " in recognition of the occupational hazard. construction of the brooklyn bridge, which was built with the help of pressurised caissons, resulted in numerous workers being either killed or permanently injured by caisson disease during its construction. barotrauma of the ears, sinus cavities and lungs and dysbaric osteonecrosis are other risks. = = other uses = = caissons have also been used in the installation of hydraulic elevators where a single - stage ram is installed below the ground level. caissons, codenamed phoenix, were an integral part of the mulberry harbours used during the world war ii allied invasion of normandy. = = other meanings = = boat lift caissons : the word caisson is also used as a synonym for the moving trough part of caisson locks, canal lifts and inclines in which boats and ships rest while being lifted from one canal elevation to another ; the water is retained on the inside of the caisson, or excluded from the caisson
the following purposes : allowing cell attachment and migration, delivering and retaining cells and biochemical factors, enabling diffusion of vital cell nutrients and expressed products, and exerting certain mechanical and biological influences to modify the behaviour of the cell phase. in 2009, an interdisciplinary team led by the thoracic surgeon thorsten walles implanted the first bioartificial transplant that provides an innate vascular network for post - transplant graft supply successfully into a patient awaiting tracheal reconstruction. to achieve the goal of tissue reconstruction, scaffolds must meet some specific requirements. high porosity and adequate pore size are necessary to facilitate cell seeding and diffusion throughout the whole structure of both cells and nutrients. biodegradability is often an essential factor since scaffolds should preferably be absorbed by the surrounding tissues without the necessity of surgical removal. the rate at which degradation occurs has to coincide as much as possible with the rate of tissue formation : this means that while cells are fabricating their own natural matrix structure around themselves, the scaffold is able to provide structural integrity within the body and eventually it will break down leaving the newly formed tissue which will take over the mechanical load. injectability is also important for clinical uses. recent research on organ printing is showing how crucial a good control of the 3d environment is to ensure reproducibility of experiments and offer better results. = = = materials = = = material selection is an essential aspect of producing a scaffold. the materials utilized can be natural or synthetic and can be biodegradable or non - biodegradable. additionally, they must be biocompatible, meaning that they do not cause any adverse effects to cells. silicone, for example, is a synthetic, non - biodegradable material commonly used as a drug delivery material, while gelatin is a biodegradable, natural material commonly used in cell - culture scaffolds the material needed for each application is different, and dependent on the desired mechanical properties of the material. tissue engineering of long bone defects for example, will require a rigid scaffold with a compressive strength similar to that of cortical bone ( 100 - 150 mpa ), which is much higher compared to a scaffold for skin regeneration. there are a few versatile synthetic materials used for many different scaffold applications. one of these commonly used materials is polylactic acid ( pla ), a synthetic polymer. pla β polylactic acid. this is a polyester which
a watershed ( called a " divide " in north america ) over which rainfall flows down towards the river traversing the lowest part of the valley, whereas the rain falling on the far slope of the watershed flows away to another river draining an adjacent basin. river basins vary in extent according to the configuration of the country, ranging from the insignificant drainage areas of streams rising on high ground near the coast and flowing straight down into the sea, up to immense tracts of continents, where rivers rising on the slopes of mountain ranges far inland have to traverse vast stretches of valleys and plains before reaching the ocean. the size of the largest river basin of any country depends on the extent of the continent in which it is situated, its position in relation to the hilly regions in which rivers generally arise and the sea into which they flow, and the distance between the source and the outlet into the sea of the river draining it. the rate of flow of rivers depends mainly upon their fall, also known as the gradient or slope. when two rivers of different sizes have the same fall, the larger river has the quicker flow, as its retardation by friction against its bed and banks is less in proportion to its volume than is the case with the smaller river. the fall available in a section of a river approximately corresponds to the slope of the country it traverses ; as rivers rise close to the highest part of their basins, generally in hilly regions, their fall is rapid near their source and gradually diminishes, with occasional irregularities, until, in traversing plains along the latter part of their course, their fall usually becomes quite gentle. accordingly, in large basins, rivers in most cases begin as torrents with a variable flow, and end as gently flowing rivers with a comparatively regular discharge. the irregular flow of rivers throughout their course forms one of the main difficulties in devising works for mitigating inundations or for increasing the navigable capabilities of rivers. in tropical countries subject to periodical rains, the rivers are in flood during the rainy season and have hardly any flow during the rest of the year, while in temperate regions, where the rainfall is more evenly distributed throughout the year, evaporation causes the available rainfall to be much less in hot summer weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern
muck ) from the edge of the workspace to a water - filled pit, connected by a tube ( called the muck tube ) to the surface. a crane at the surface removes the soil with a clamshell bucket. the water pressure in the tube balances the air pressure, with excess air escaping up the muck tube. the pressurized air flow must be constant to ensure regular air changes for the workers and prevent excessive inflow of mud or water at the base of the caisson. when the caisson hits bedrock, the sandhogs exit through the airlock and fill the box with concrete, forming a solid foundation pier. a pneumatic ( compressed - air ) caisson has the advantage of providing dry working conditions, which is better for placing concrete. it is also well suited for foundations for which other methods might cause settlement of adjacent structures. construction workers who leave the pressurized environment of the caisson must decompress at a rate that allows symptom - free release of inert gases dissolved in the body tissues if they are to avoid decompression sickness, a condition first identified in caisson workers, and originally named " caisson disease " in recognition of the occupational hazard. construction of the brooklyn bridge, which was built with the help of pressurised caissons, resulted in numerous workers being either killed or permanently injured by caisson disease during its construction. barotrauma of the ears, sinus cavities and lungs and dysbaric osteonecrosis are other risks. = = other uses = = caissons have also been used in the installation of hydraulic elevators where a single - stage ram is installed below the ground level. caissons, codenamed phoenix, were an integral part of the mulberry harbours used during the world war ii allied invasion of normandy. = = other meanings = = boat lift caissons : the word caisson is also used as a synonym for the moving trough part of caisson locks, canal lifts and inclines in which boats and ships rest while being lifted from one canal elevation to another ; the water is retained on the inside of the caisson, or excluded from the caisson, according to the respective operating principle. structural caissons : caisson is also sometimes used as a colloquial term for a reinforced concrete structure formed by pouring into a hollow cylindrical form, typically by placing a caisson form below grade in an open excavation and pouring once backfill is complete, or by
Question: Which movement of carbon between reservoirs takes the longest to achieve?
A) from burning coal to the atmosphere
B) from plant tissues to coal in the crust
C) from the atmosphere to the sugars in plants
D) from the respiration of fish to the ocean waters
|
B) from plant tissues to coal in the crust
|
Context:
it is hard for us humans to recognize things in nature until we have invented them ourselves. for image - forming optics, nature has made virtually every kind of lens humans have devised. but what about lensless " imaging "? recently, we showed that a bare array of sensors on a curved substrate could achieve resolution not limited by diffraction - without any lens at all provided that the objects imaged conform to our a priori assumptions. is it possible that somewhere in nature we will find this kind of vision system? we think so and provide examples that seem to make no sense whatever unless they are using something like our lensless imaging work.
##d product that is the focus of a tooling drawing. lines can also be classified by a letter classification in which each line is given a letter. type a lines show the outline of the feature of an object. they are the thickest lines on a drawing and done with a pencil softer than hb. type b lines are dimension lines and are used for dimensioning, projecting, extending, or leaders. a harder pencil should be used, such as a 2h pencil. type c lines are used for breaks when the whole object is not shown. these are freehand drawn and only for short breaks. 2h pencil type d lines are similar to type c, except these are zigzagged and only for longer breaks. 2h pencil type e lines indicate hidden outlines of internal features of an object. these are dotted lines. 2h pencil type f lines are type e lines, except these are used for drawings in electrotechnology. 2h pencil type g lines are used for centre lines. these are dotted lines, but a long line of 10 β 20 mm, then a 1 mm gap, then a small line of 2 mm. 2h pencil type h lines are the same as type g, except that every second long line is thicker. these indicate the cutting plane of an object. 2h pencil type k lines indicate the alternate positions of an object and the line taken by that object. these are drawn with a long line of 10 β 20 mm, then a small gap, then a small line of 2 mm, then a gap, then another small line. 2h pencil. = = = multiple views and projections = = = in most cases, a single view is not sufficient to show all necessary features, and several views are used. types of views include the following : = = = = multiview projection = = = = a multiview projection is a type of orthographic projection that shows the object as it looks from the front, right, left, top, bottom, or back ( e. g. the primary views ), and is typically positioned relative to each other according to the rules of either first - angle or third - angle projection. the origin and vector direction of the projectors ( also called projection lines ) differs, as explained below. in first - angle projection, the parallel projectors originate as if radiated from behind the viewer and pass through the 3d object to project a 2d image onto the orthogonal plane behind it. the 3d object is projected into 2d " paper " space as if you were looking at
are commonly referred to as " cross - hatching ". phantom β ( not shown ) are alternately long - and double short - dashed thin lines used to represent a feature or component that is not part of the specified part or assembly. e. g. billet ends that may be used for testing, or the machined product that is the focus of a tooling drawing. lines can also be classified by a letter classification in which each line is given a letter. type a lines show the outline of the feature of an object. they are the thickest lines on a drawing and done with a pencil softer than hb. type b lines are dimension lines and are used for dimensioning, projecting, extending, or leaders. a harder pencil should be used, such as a 2h pencil. type c lines are used for breaks when the whole object is not shown. these are freehand drawn and only for short breaks. 2h pencil type d lines are similar to type c, except these are zigzagged and only for longer breaks. 2h pencil type e lines indicate hidden outlines of internal features of an object. these are dotted lines. 2h pencil type f lines are type e lines, except these are used for drawings in electrotechnology. 2h pencil type g lines are used for centre lines. these are dotted lines, but a long line of 10 β 20 mm, then a 1 mm gap, then a small line of 2 mm. 2h pencil type h lines are the same as type g, except that every second long line is thicker. these indicate the cutting plane of an object. 2h pencil type k lines indicate the alternate positions of an object and the line taken by that object. these are drawn with a long line of 10 β 20 mm, then a small gap, then a small line of 2 mm, then a gap, then another small line. 2h pencil. = = = multiple views and projections = = = in most cases, a single view is not sufficient to show all necessary features, and several views are used. types of views include the following : = = = = multiview projection = = = = a multiview projection is a type of orthographic projection that shows the object as it looks from the front, right, left, top, bottom, or back ( e. g. the primary views ), and is typically positioned relative to each other according to the rules of either first - angle or third - angle projection. the origin and vector direction of the projectors (
scientists look through telescopes, study images on electronic screens, record meter readings, and so on. generally, on a basic level, they can agree on what they see, e. g., the thermometer shows 37. 9 degrees c. but, if these scientists have different ideas about the theories that have been developed to explain these basic observations, they may disagree about what they are observing. for example, before albert einstein ' s general theory of relativity, observers would have likely interpreted an image of the einstein cross as five different objects in space. in light of that theory, however, astronomers will tell you that there are actually only two objects, one in the center and four different images of a second object around the sides. alternatively, if other scientists suspect that something is wrong with the telescope and only one object is actually being observed, they are operating under yet another theory. observations that cannot be separated from theoretical interpretation are said to be theory - laden. all observation involves both perception and cognition. that is, one does not make an observation passively, but rather is actively engaged in distinguishing the phenomenon being observed from surrounding sensory data. therefore, observations are affected by one ' s underlying understanding of the way in which the world functions, and that understanding may influence what is perceived, noticed, or deemed worthy of consideration. in this sense, it can be argued that all observation is theory - laden. = = = the purpose of science = = = should science aim to determine ultimate truth, or are there questions that science cannot answer? scientific realists claim that science aims at truth and that one ought to regard scientific theories as true, approximately true, or likely true. conversely, scientific anti - realists argue that science does not aim ( or at least does not succeed ) at truth, especially truth about unobservables like electrons or other universes. instrumentalists argue that scientific theories should only be evaluated on whether they are useful. in their view, whether theories are true or not is beside the point, because the purpose of science is to make predictions and enable effective technology. realists often point to the success of recent scientific theories as evidence for the truth ( or near truth ) of current theories. antirealists point to either the many false theories in the history of science, epistemic morals, the success of false modeling assumptions, or widely termed postmodern criticisms of objectivity as evidence against scientific realism. antirealists attempt to explain the success of scientific theories without reference to truth. some antirealists claim that scientific
emitter rather than returning a diffuse signal detectable at many angles. the effect is sometimes called " glitter " after the very brief signal seen when the reflected beam passes across a detector. it can be difficult for the radar operator to distinguish between a glitter event and a digital glitch in the processing system. stealth airframes sometimes display distinctive serrations on some exposed edges, such as the engine ports. the yf - 23 has such serrations on the exhaust ports. this is another example in the parallel alignment of features, this time on the external airframe. the shaping requirements detracted greatly from the f - 117 ' s aerodynamic properties. it is inherently unstable, and cannot be flown without a fly - by - wire control system. similarly, coating the cockpit canopy with a thin film transparent conductor ( vapor - deposited gold or indium tin oxide ) helps to reduce the aircraft ' s radar profile, because radar waves would normally enter the cockpit, reflect off objects ( the inside of a cockpit has a complex shape, with a pilot helmet alone forming a sizeable return ), and possibly return to the radar, but the conductive coating creates a controlled shape that deflects the incoming radar waves away from the radar. the coating is thin enough that it has no adverse effect on pilot vision. = = = = ships = = = = ships have also adopted similar methods. though the earlier american arleigh burke - class destroyers incorporated some signature - reduction features. the norwegian skjold - class corvettes was the first coastal defence and the french la fayette - class frigates the first ocean - going stealth ships to enter service. other examples are the dutch de zeven provincien - class frigates, the taiwanese tuo chiang - class corvettes, german sachsen - class frigates, the swedish visby - class corvette, the american san antonio - class amphibious transport docks, and most modern warship designs. = = = materials = = = = = = = non - metallic airframe = = = = dielectric composite materials are more transparent to radar, whereas electrically conductive materials such as metals and carbon fibers reflect electromagnetic energy incident on the material ' s surface. composites may also contain ferrites to optimize the dielectric and magnetic properties of a material for its application. = = = = radar - absorbent material = = = = radiation - absorbent material ( ram ), often as paints, are used especially on the edges of metal surfaces. while the material and thickness of ram coatings can
, including objects we can see with our naked eyes. it is one of the oldest sciences. astronomers of early civilizations performed methodical observations of the night sky, and astronomical artifacts have been found from much earlier periods. there are two types of astronomy : observational astronomy and theoretical astronomy. observational astronomy is focused on acquiring and analyzing data, mainly using basic principles of physics. in contrast, theoretical astronomy is oriented towards developing computer or analytical models to describe astronomical objects and phenomena. this discipline is the science of celestial objects and phenomena that originate outside the earth ' s atmosphere. it is concerned with the evolution, physics, chemistry, meteorology, geology, and motion of celestial objects, as well as the formation and development of the universe. astronomy includes examining, studying, and modeling stars, planets, and comets. most of the information used by astronomers is gathered by remote observation. however, some laboratory reproduction of celestial phenomena has been performed ( such as the molecular chemistry of the interstellar medium ). there is considerable overlap with physics and in some areas of earth science. there are also interdisciplinary fields such as astrophysics, planetary sciences, and cosmology, along with allied disciplines such as space physics and astrochemistry. while the study of celestial features and phenomena can be traced back to antiquity, the scientific methodology of this field began to develop in the middle of the 17th century. a key factor was galileo ' s introduction of the telescope to examine the night sky in more detail. the mathematical treatment of astronomy began with newton ' s development of celestial mechanics and the laws of gravitation. however, it was triggered by earlier work of astronomers such as kepler. by the 19th century, astronomy had developed into formal science, with the introduction of instruments such as the spectroscope and photography, along with much - improved telescopes and the creation of professional observatories. = = interdisciplinary studies = = the distinctions between the natural science disciplines are not always sharp, and they share many cross - discipline fields. physics plays a significant role in the other natural sciences, as represented by astrophysics, geophysics, chemical physics and biophysics. likewise chemistry is represented by such fields as biochemistry, physical chemistry, geochemistry and astrochemistry. a particular example of a scientific discipline that draws upon multiple natural sciences is environmental science. this field studies the interactions of physical, chemical, geological, and biological components of the environment, with particular regard to the effect of human activities and the impact on biodiversity and sustainability. this science also draws upon expertise from other fields, such
the mean apparent magnitude of starlink mini direct - to - cell ( dtc ) satellites is 4. 62 while the mean of magnitudes adjusted to a uniform distance of 1000 km is 5. 50. dtcs average 4. 9 times brighter than other starlink mini spacecraft at a common distance. we cannot currently separate the effects of the dtc antenna itself, the different attitude modes that may be required for dtc operations and to what extent brightness mitigation procedures were in place at the times of our observations. in a best case scenario, where dtc brightness mitigation is as successful as that for other minis and the dtc antenna does not add significantly to brightness, we estimate that dtcs will be about 2. 6 times as bright as the others based upon their lower altitudes. the dtcs spend a greater fraction of their time in the earth ' s shadow than satellites at higher altitudes. that will offset some of their impact on astronomical observing.
are continuous lines used to depict edges directly visible from a particular angle. hidden β are short - dashed lines that may be used to represent edges that are not directly visible. center β are alternately long - and short - dashed lines that may be used to represent the axes of circular features. cutting plane β are thin, medium - dashed lines, or thick alternately long - and double short - dashed that may be used to define sections for section views. section β are thin lines in a pattern ( pattern determined by the material being " cut " or " sectioned " ) used to indicate surfaces in section views resulting from " cutting ". section lines are commonly referred to as " cross - hatching ". phantom β ( not shown ) are alternately long - and double short - dashed thin lines used to represent a feature or component that is not part of the specified part or assembly. e. g. billet ends that may be used for testing, or the machined product that is the focus of a tooling drawing. lines can also be classified by a letter classification in which each line is given a letter. type a lines show the outline of the feature of an object. they are the thickest lines on a drawing and done with a pencil softer than hb. type b lines are dimension lines and are used for dimensioning, projecting, extending, or leaders. a harder pencil should be used, such as a 2h pencil. type c lines are used for breaks when the whole object is not shown. these are freehand drawn and only for short breaks. 2h pencil type d lines are similar to type c, except these are zigzagged and only for longer breaks. 2h pencil type e lines indicate hidden outlines of internal features of an object. these are dotted lines. 2h pencil type f lines are type e lines, except these are used for drawings in electrotechnology. 2h pencil type g lines are used for centre lines. these are dotted lines, but a long line of 10 β 20 mm, then a 1 mm gap, then a small line of 2 mm. 2h pencil type h lines are the same as type g, except that every second long line is thicker. these indicate the cutting plane of an object. 2h pencil type k lines indicate the alternate positions of an object and the line taken by that object. these are drawn with a long line of 10 β 20 mm, then a small gap, then a small line of 2 mm, then a gap, then another small line. 2h
. most cells are very small, with diameters ranging from 1 to 100 micrometers and are therefore only visible under a light or electron microscope. there are generally two types of cells : eukaryotic cells, which contain a nucleus, and prokaryotic cells, which do not. prokaryotes are single - celled organisms such as bacteria, whereas eukaryotes can be single - celled or multicellular. in multicellular organisms, every cell in the organism ' s body is derived ultimately from a single cell in a fertilized egg. = = = cell structure = = = every cell is enclosed within a cell membrane that separates its cytoplasm from the extracellular space. a cell membrane consists of a lipid bilayer, including cholesterols that sit between phospholipids to maintain their fluidity at various temperatures. cell membranes are semipermeable, allowing small molecules such as oxygen, carbon dioxide, and water to pass through while restricting the movement of larger molecules and charged particles such as ions. cell membranes also contain membrane proteins, including integral membrane proteins that go across the membrane serving as membrane transporters, and peripheral proteins that loosely attach to the outer side of the cell membrane, acting as enzymes shaping the cell. cell membranes are involved in various cellular processes such as cell adhesion, storing electrical energy, and cell signalling and serve as the attachment surface for several extracellular structures such as a cell wall, glycocalyx, and cytoskeleton. within the cytoplasm of a cell, there are many biomolecules such as proteins and nucleic acids. in addition to biomolecules, eukaryotic cells have specialized structures called organelles that have their own lipid bilayers or are spatially units. these organelles include the cell nucleus, which contains most of the cell ' s dna, or mitochondria, which generate adenosine triphosphate ( atp ) to power cellular processes. other organelles such as endoplasmic reticulum and golgi apparatus play a role in the synthesis and packaging of proteins, respectively. biomolecules such as proteins can be engulfed by lysosomes, another specialized organelle. plant cells have additional organelles that distinguish them from animal cells such as a cell wall that provides support for the plant cell, chloroplasts that harvest sunlight energy to produce sugar, and vacuoles that provide storage and structural support
- dashed lines, or thick alternately long - and double short - dashed that may be used to define sections for section views. section β are thin lines in a pattern ( pattern determined by the material being " cut " or " sectioned " ) used to indicate surfaces in section views resulting from " cutting ". section lines are commonly referred to as " cross - hatching ". phantom β ( not shown ) are alternately long - and double short - dashed thin lines used to represent a feature or component that is not part of the specified part or assembly. e. g. billet ends that may be used for testing, or the machined product that is the focus of a tooling drawing. lines can also be classified by a letter classification in which each line is given a letter. type a lines show the outline of the feature of an object. they are the thickest lines on a drawing and done with a pencil softer than hb. type b lines are dimension lines and are used for dimensioning, projecting, extending, or leaders. a harder pencil should be used, such as a 2h pencil. type c lines are used for breaks when the whole object is not shown. these are freehand drawn and only for short breaks. 2h pencil type d lines are similar to type c, except these are zigzagged and only for longer breaks. 2h pencil type e lines indicate hidden outlines of internal features of an object. these are dotted lines. 2h pencil type f lines are type e lines, except these are used for drawings in electrotechnology. 2h pencil type g lines are used for centre lines. these are dotted lines, but a long line of 10 β 20 mm, then a 1 mm gap, then a small line of 2 mm. 2h pencil type h lines are the same as type g, except that every second long line is thicker. these indicate the cutting plane of an object. 2h pencil type k lines indicate the alternate positions of an object and the line taken by that object. these are drawn with a long line of 10 β 20 mm, then a small gap, then a small line of 2 mm, then a gap, then another small line. 2h pencil. = = = multiple views and projections = = = in most cases, a single view is not sufficient to show all necessary features, and several views are used. types of views include the following : = = = = multiview projection = = = = a multiview projection is a type of orthographic projection
Question: Which types of objects are most easily seen by the human eye?
A) objects that reflect most of the light that strikes them
B) objects that refract most of the light that strikes them
C) objects that diffract most of the light that strikes them
D) objects that absorb most of the light that strikes them
|
A) objects that reflect most of the light that strikes them
|
Context:
the best - suited crops ( e. g., those with the highest yields ) to produce enough food to support a growing population. as crops and fields became increasingly large and difficult to maintain, it was discovered that specific organisms and their by - products could effectively fertilize, restore nitrogen, and control pests. throughout the history of agriculture, farmers have inadvertently altered the genetics of their crops through introducing them to new environments and breeding them with other plants β one of the first forms of biotechnology. these processes also were included in early fermentation of beer. these processes were introduced in early mesopotamia, egypt, china and india, and still use the same basic biological methods. in brewing, malted grains ( containing enzymes ) convert starch from grains into sugar and then adding specific yeasts to produce beer. in this process, carbohydrates in the grains broke down into alcohols, such as ethanol. later, other cultures produced the process of lactic acid fermentation, which produced other preserved foods, such as soy sauce. fermentation was also used in this time period to produce leavened bread. although the process of fermentation was not fully understood until louis pasteur ' s work in 1857, it is still the first use of biotechnology to convert a food source into another form. before the time of charles darwin ' s work and life, animal and plant scientists had already used selective breeding. darwin added to that body of work with his scientific observations about the ability of science to change species. these accounts contributed to darwin ' s theory of natural selection. for thousands of years, humans have used selective breeding to improve the production of crops and livestock to use them for food. in selective breeding, organisms with desirable characteristics are mated to produce offspring with the same characteristics. for example, this technique was used with corn to produce the largest and sweetest crops. in the early twentieth century scientists gained a greater understanding of microbiology and explored ways of manufacturing specific products. in 1917, chaim weizmann first used a pure microbiological culture in an industrial process, that of manufacturing corn starch using clostridium acetobutylicum, to produce acetone, which the united kingdom desperately needed to manufacture explosives during world war i. biotechnology has also led to the development of antibiotics. in 1928, alexander fleming discovered the mold penicillium. his work led to the purification of the antibiotic formed by the mold by howard florey, ernst boris chain and norman heatley β to form
best - known and controversial applications of genetic engineering is the creation and use of genetically modified crops or genetically modified livestock to produce genetically modified food. crops have been developed to increase production, increase tolerance to abiotic stresses, alter the composition of the food, or to produce novel products. the first crops to be released commercially on a large scale provided protection from insect pests or tolerance to herbicides. fungal and virus resistant crops have also been developed or are in development. this makes the insect and weed management of crops easier and can indirectly increase crop yield. gm crops that directly improve yield by accelerating growth or making the plant more hardy ( by improving salt, cold or drought tolerance ) are also under development. in 2016 salmon have been genetically modified with growth hormones to reach normal adult size much faster. gmos have been developed that modify the quality of produce by increasing the nutritional value or providing more industrially useful qualities or quantities. the amflora potato produces a more industrially useful blend of starches. soybeans and canola have been genetically modified to produce more healthy oils. the first commercialised gm food was a tomato that had delayed ripening, increasing its shelf life. plants and animals have been engineered to produce materials they do not normally make. pharming uses crops and animals as bioreactors to produce vaccines, drug intermediates, or the drugs themselves ; the useful product is purified from the harvest and then used in the standard pharmaceutical production process. cows and goats have been engineered to express drugs and other proteins in their milk, and in 2009 the fda approved a drug produced in goat milk. = = = other applications = = = genetic engineering has potential applications in conservation and natural area management. gene transfer through viral vectors has been proposed as a means of controlling invasive species as well as vaccinating threatened fauna from disease. transgenic trees have been suggested as a way to confer resistance to pathogens in wild populations. with the increasing risks of maladaptation in organisms as a result of climate change and other perturbations, facilitated adaptation through gene tweaking could be one solution to reducing extinction risks. applications of genetic engineering in conservation are thus far mostly theoretical and have yet to be put into practice. genetic engineering is also being used to create microbial art. some bacteria have been genetically engineered to create black and white photographs. novelty items such as lavender - colored carnations, blue roses, and glowing fish, have also been produced through genetic engineering. = = regulation = = the regulation of genetic engineering
naturally take up foreign dna. this ability can be induced in other bacteria via stress ( e. g. thermal or electric shock ), which increases the cell membrane ' s permeability to dna ; up - taken dna can either integrate with the genome or exist as extrachromosomal dna. dna is generally inserted into animal cells using microinjection, where it can be injected through the cell ' s nuclear envelope directly into the nucleus, or through the use of viral vectors. plant genomes can be engineered by physical methods or by use of agrobacterium for the delivery of sequences hosted in t - dna binary vectors. in plants the dna is often inserted using agrobacterium - mediated transformation, taking advantage of the agrobacteriums t - dna sequence that allows natural insertion of genetic material into plant cells. other methods include biolistics, where particles of gold or tungsten are coated with dna and then shot into young plant cells, and electroporation, which involves using an electric shock to make the cell membrane permeable to plasmid dna. as only a single cell is transformed with genetic material, the organism must be regenerated from that single cell. in plants this is accomplished through the use of tissue culture. in animals it is necessary to ensure that the inserted dna is present in the embryonic stem cells. bacteria consist of a single cell and reproduce clonally so regeneration is not necessary. selectable markers are used to easily differentiate transformed from untransformed cells. these markers are usually present in the transgenic organism, although a number of strategies have been developed that can remove the selectable marker from the mature transgenic plant. further testing using pcr, southern hybridization, and dna sequencing is conducted to confirm that an organism contains the new gene. these tests can also confirm the chromosomal location and copy number of the inserted gene. the presence of the gene does not guarantee it will be expressed at appropriate levels in the target tissue so methods that look for and measure the gene products ( rna and protein ) are also used. these include northern hybridisation, quantitative rt - pcr, western blot, immunofluorescence, elisa and phenotypic analysis. the new genetic material can be inserted randomly within the host genome or targeted to a specific location. the technique of gene targeting uses homologous recombination to make desired changes to a specific endogenous gene. this tends to occur at a relatively low frequency in plants and animals and generally
inherited traits such as shape in pisum sativum ( peas ). what mendel learned from studying plants has had far - reaching benefits outside of botany. similarly, " jumping genes " were discovered by barbara mcclintock while she was studying maize. nevertheless, there are some distinctive genetic differences between plants and other organisms. species boundaries in plants may be weaker than in animals, and cross species hybrids are often possible. a familiar example is peppermint, mentha Γ piperita, a sterile hybrid between mentha aquatica and spearmint, mentha spicata. the many cultivated varieties of wheat are the result of multiple inter - and intra - specific crosses between wild species and their hybrids. angiosperms with monoecious flowers often have self - incompatibility mechanisms that operate between the pollen and stigma so that the pollen either fails to reach the stigma or fails to germinate and produce male gametes. this is one of several methods used by plants to promote outcrossing. in many land plants the male and female gametes are produced by separate individuals. these species are said to be dioecious when referring to vascular plant sporophytes and dioicous when referring to bryophyte gametophytes. charles darwin in his 1878 book the effects of cross and self - fertilization in the vegetable kingdom at the start of chapter xii noted " the first and most important of the conclusions which may be drawn from the observations given in this volume, is that generally cross - fertilisation is beneficial and self - fertilisation often injurious, at least with the plants on which i experimented. " an important adaptive benefit of outcrossing is that it allows the masking of deleterious mutations in the genome of progeny. this beneficial effect is also known as hybrid vigor or heterosis. once outcrossing is established, subsequent switching to inbreeding becomes disadvantageous since it allows expression of the previously masked deleterious recessive mutations, commonly referred to as inbreeding depression. unlike in higher animals, where parthenogenesis is rare, asexual reproduction may occur in plants by several different mechanisms. the formation of stem tubers in potato is one example. particularly in arctic or alpine habitats, where opportunities for fertilisation of flowers by animals are rare, plantlets or bulbs, may develop instead of flowers, replacing sexual reproduction with asexual reproduction and giving rise to clonal populations genetically identical to the parent. this is one
generation times. corn has been used to study mechanisms of photosynthesis and phloem loading of sugar in c4 plants. the single celled green alga chlamydomonas reinhardtii, while not an embryophyte itself, contains a green - pigmented chloroplast related to that of land plants, making it useful for study. a red alga cyanidioschyzon merolae has also been used to study some basic chloroplast functions. spinach, peas, soybeans and a moss physcomitrella patens are commonly used to study plant cell biology. agrobacterium tumefaciens, a soil rhizosphere bacterium, can attach to plant cells and infect them with a callus - inducing ti plasmid by horizontal gene transfer, causing a callus infection called crown gall disease. schell and van montagu ( 1977 ) hypothesised that the ti plasmid could be a natural vector for introducing the nif gene responsible for nitrogen fixation in the root nodules of legumes and other plant species. today, genetic modification of the ti plasmid is one of the main techniques for introduction of transgenes to plants and the creation of genetically modified crops. = = = epigenetics = = = epigenetics is the study of heritable changes in gene function that cannot be explained by changes in the underlying dna sequence but cause the organism ' s genes to behave ( or " express themselves " ) differently. one example of epigenetic change is the marking of the genes by dna methylation which determines whether they will be expressed or not. gene expression can also be controlled by repressor proteins that attach to silencer regions of the dna and prevent that region of the dna code from being expressed. epigenetic marks may be added or removed from the dna during programmed stages of development of the plant, and are responsible, for example, for the differences between anthers, petals and normal leaves, despite the fact that they all have the same underlying genetic code. epigenetic changes may be temporary or may remain through successive cell divisions for the remainder of the cell ' s life. some epigenetic changes have been shown to be heritable, while others are reset in the germ cells. epigenetic changes in eukaryotic biology serve to regulate the process of cellular differentiation. during morphogenesis, totipotent stem cells become the various
the reply to g. w. bruhn is added.
cytokinesis ). the natural cytokinin zeatin was discovered in corn, zea mays, and is a derivative of the purine adenine. zeatin is produced in roots and transported to shoots in the xylem where it promotes cell division, bud development, and the greening of chloroplasts. the gibberelins, such as gibberelic acid are diterpenes synthesised from acetyl coa via the mevalonate pathway. they are involved in the promotion of germination and dormancy - breaking in seeds, in regulation of plant height by controlling stem elongation and the control of flowering. abscisic acid ( aba ) occurs in all land plants except liverworts, and is synthesised from carotenoids in the chloroplasts and other plastids. it inhibits cell division, promotes seed maturation, and dormancy, and promotes stomatal closure. it was so named because it was originally thought to control abscission. ethylene is a gaseous hormone that is produced in all higher plant tissues from methionine. it is now known to be the hormone that stimulates or regulates fruit ripening and abscission, and it, or the synthetic growth regulator ethephon which is rapidly metabolised to produce ethylene, are used on industrial scale to promote ripening of cotton, pineapples and other climacteric crops. another class of phytohormones is the jasmonates, first isolated from the oil of jasminum grandiflorum which regulates wound responses in plants by unblocking the expression of genes required in the systemic acquired resistance response to pathogen attack. in addition to being the primary energy source for plants, light functions as a signalling device, providing information to the plant, such as how much sunlight the plant receives each day. this can result in adaptive changes in a process known as photomorphogenesis. phytochromes are the photoreceptors in a plant that are sensitive to light. = = plant anatomy and morphology = = plant anatomy is the study of the structure of plant cells and tissues, whereas plant morphology is the study of their external form. all plants are multicellular eukaryotes, their dna stored in nuclei. the characteristic features of plant cells that distinguish them from those of animals and fungi include a primary cell wall composed of the polysaccharides cellulose,
the broad definition of " utilizing a biotechnological system to make products ". indeed, the cultivation of plants may be viewed as the earliest biotechnological enterprise. agriculture has been theorized to have become the dominant way of producing food since the neolithic revolution. through early biotechnology, the earliest farmers selected and bred the best - suited crops ( e. g., those with the highest yields ) to produce enough food to support a growing population. as crops and fields became increasingly large and difficult to maintain, it was discovered that specific organisms and their by - products could effectively fertilize, restore nitrogen, and control pests. throughout the history of agriculture, farmers have inadvertently altered the genetics of their crops through introducing them to new environments and breeding them with other plants β one of the first forms of biotechnology. these processes also were included in early fermentation of beer. these processes were introduced in early mesopotamia, egypt, china and india, and still use the same basic biological methods. in brewing, malted grains ( containing enzymes ) convert starch from grains into sugar and then adding specific yeasts to produce beer. in this process, carbohydrates in the grains broke down into alcohols, such as ethanol. later, other cultures produced the process of lactic acid fermentation, which produced other preserved foods, such as soy sauce. fermentation was also used in this time period to produce leavened bread. although the process of fermentation was not fully understood until louis pasteur ' s work in 1857, it is still the first use of biotechnology to convert a food source into another form. before the time of charles darwin ' s work and life, animal and plant scientists had already used selective breeding. darwin added to that body of work with his scientific observations about the ability of science to change species. these accounts contributed to darwin ' s theory of natural selection. for thousands of years, humans have used selective breeding to improve the production of crops and livestock to use them for food. in selective breeding, organisms with desirable characteristics are mated to produce offspring with the same characteristics. for example, this technique was used with corn to produce the largest and sweetest crops. in the early twentieth century scientists gained a greater understanding of microbiology and explored ways of manufacturing specific products. in 1917, chaim weizmann first used a pure microbiological culture in an industrial process, that of manufacturing corn starch using clostridium acetobutylicum, to produce acetone, which the united
cellular and molecular biology of cereals, grasses and monocots generally. model plants such as arabidopsis thaliana are used for studying the molecular biology of plant cells and the chloroplast. ideally, these organisms have small genomes that are well known or completely sequenced, small stature and short generation times. corn has been used to study mechanisms of photosynthesis and phloem loading of sugar in c4 plants. the single celled green alga chlamydomonas reinhardtii, while not an embryophyte itself, contains a green - pigmented chloroplast related to that of land plants, making it useful for study. a red alga cyanidioschyzon merolae has also been used to study some basic chloroplast functions. spinach, peas, soybeans and a moss physcomitrella patens are commonly used to study plant cell biology. agrobacterium tumefaciens, a soil rhizosphere bacterium, can attach to plant cells and infect them with a callus - inducing ti plasmid by horizontal gene transfer, causing a callus infection called crown gall disease. schell and van montagu ( 1977 ) hypothesised that the ti plasmid could be a natural vector for introducing the nif gene responsible for nitrogen fixation in the root nodules of legumes and other plant species. today, genetic modification of the ti plasmid is one of the main techniques for introduction of transgenes to plants and the creation of genetically modified crops. = = = epigenetics = = = epigenetics is the study of heritable changes in gene function that cannot be explained by changes in the underlying dna sequence but cause the organism ' s genes to behave ( or " express themselves " ) differently. one example of epigenetic change is the marking of the genes by dna methylation which determines whether they will be expressed or not. gene expression can also be controlled by repressor proteins that attach to silencer regions of the dna and prevent that region of the dna code from being expressed. epigenetic marks may be added or removed from the dna during programmed stages of development of the plant, and are responsible, for example, for the differences between anthers, petals and normal leaves, despite the fact that they all have the same underlying genetic code. epigenetic changes may be temporary or may remain through successive cell divisions for the remainder of
##aggeration to say that the tip of the radicle.. acts like the brain of one of the lower animals.. directing the several movements ". about the same time, the role of auxins ( from the greek auxein, to grow ) in control of plant growth was first outlined by the dutch scientist frits went. the first known auxin, indole - 3 - acetic acid ( iaa ), which promotes cell growth, was only isolated from plants about 50 years later. this compound mediates the tropic responses of shoots and roots towards light and gravity. the finding in 1939 that plant callus could be maintained in culture containing iaa, followed by the observation in 1947 that it could be induced to form roots and shoots by controlling the concentration of growth hormones were key steps in the development of plant biotechnology and genetic modification. cytokinins are a class of plant hormones named for their control of cell division ( especially cytokinesis ). the natural cytokinin zeatin was discovered in corn, zea mays, and is a derivative of the purine adenine. zeatin is produced in roots and transported to shoots in the xylem where it promotes cell division, bud development, and the greening of chloroplasts. the gibberelins, such as gibberelic acid are diterpenes synthesised from acetyl coa via the mevalonate pathway. they are involved in the promotion of germination and dormancy - breaking in seeds, in regulation of plant height by controlling stem elongation and the control of flowering. abscisic acid ( aba ) occurs in all land plants except liverworts, and is synthesised from carotenoids in the chloroplasts and other plastids. it inhibits cell division, promotes seed maturation, and dormancy, and promotes stomatal closure. it was so named because it was originally thought to control abscission. ethylene is a gaseous hormone that is produced in all higher plant tissues from methionine. it is now known to be the hormone that stimulates or regulates fruit ripening and abscission, and it, or the synthetic growth regulator ethephon which is rapidly metabolised to produce ethylene, are used on industrial scale to promote ripening of cotton, pineapples and other climacteric crops. another class of phytohormones is the jasmonates, first isolated
Question: Planting legumes in order to add nitrogen to soil most likely causes
A) a decrease in the amount of acid rain.
B) an enrichment in the fertility of the soil.
C) an increase in the ratio of nitrogen to oxygen in the system.
D) a reduction in the soil's ability to manufacture its own nitrogen.
|
B) an enrichment in the fertility of the soil.
|
Context:
and measuring radiation levels. the surveyor program conducted uncrewed lunar landings and takeoffs, as well as taking surface and regolith observations. despite the setback caused by the apollo 1 fire, which killed three astronauts, the program proceeded. apollo 8 was the first crewed spacecraft to leave low earth orbit and the first human spaceflight to reach the moon. the crew orbited the moon ten times on december 24 and 25, 1968, and then traveled safely back to earth. the three apollo 8 astronauts β frank borman, james lovell, and william anders β were the first humans to see the earth as a globe in space, the first to witness an earthrise, and the first to see and manually photograph the far side of the moon. the first lunar landing was conducted by apollo 11. commanded by neil armstrong with astronauts buzz aldrin and michael collins, apollo 11 was one of the most significant missions in nasa ' s history, marking the end of the space race when the soviet union gave up its lunar ambitions. as the first human to step on the surface of the moon, neil armstrong uttered the now famous words : that ' s one small step for man, one giant leap for mankind. nasa would conduct six total lunar landings as part of the apollo program, with apollo 17 concluding the program in 1972. = = = = end of apollo = = = = wernher von braun had advocated for nasa to develop a space station since the agency was created. in 1973, following the end of the apollo lunar missions, nasa launched its first space station, skylab, on the final launch of the saturn v. skylab reused a significant amount of apollo and saturn hardware, with a repurposed saturn v third stage serving as the primary module for the space station. damage to skylab during its launch required spacewalks to be performed by the first crew to make it habitable and operational. skylab hosted nine missions and was decommissioned in 1974 and deorbited in 1979, two years prior to the first launch of the space shuttle and any possibility of boosting its orbit. in 1975, the apollo β soyuz mission was the first ever international spaceflight and a major diplomatic accomplishment between the cold war rivals, which also marked the last flight of the apollo capsule. flown in 1975, a us apollo spacecraft docked with a soviet soyuz capsule. = = = interplanetary exploration and space science = = = during the 1960s, nasa started its space science and interplanetary probe program. the mariner program was its flagship
fuel cells instead of batteries, and conducted the first american spacewalks and rendezvous operations. the ranger program was started in the 1950s as a response to soviet lunar exploration, however most missions ended in failure. the lunar orbiter program had greater success, mapping the surface in preparation for apollo landings, conducting meteoroid detection, and measuring radiation levels. the surveyor program conducted uncrewed lunar landings and takeoffs, as well as taking surface and regolith observations. despite the setback caused by the apollo 1 fire, which killed three astronauts, the program proceeded. apollo 8 was the first crewed spacecraft to leave low earth orbit and the first human spaceflight to reach the moon. the crew orbited the moon ten times on december 24 and 25, 1968, and then traveled safely back to earth. the three apollo 8 astronauts β frank borman, james lovell, and william anders β were the first humans to see the earth as a globe in space, the first to witness an earthrise, and the first to see and manually photograph the far side of the moon. the first lunar landing was conducted by apollo 11. commanded by neil armstrong with astronauts buzz aldrin and michael collins, apollo 11 was one of the most significant missions in nasa ' s history, marking the end of the space race when the soviet union gave up its lunar ambitions. as the first human to step on the surface of the moon, neil armstrong uttered the now famous words : that ' s one small step for man, one giant leap for mankind. nasa would conduct six total lunar landings as part of the apollo program, with apollo 17 concluding the program in 1972. = = = = end of apollo = = = = wernher von braun had advocated for nasa to develop a space station since the agency was created. in 1973, following the end of the apollo lunar missions, nasa launched its first space station, skylab, on the final launch of the saturn v. skylab reused a significant amount of apollo and saturn hardware, with a repurposed saturn v third stage serving as the primary module for the space station. damage to skylab during its launch required spacewalks to be performed by the first crew to make it habitable and operational. skylab hosted nine missions and was decommissioned in 1974 and deorbited in 1979, two years prior to the first launch of the space shuttle and any possibility of boosting its orbit. in 1975, the apollo β soyuz mission was the first ever international spaceflight and a major diplomatic accomplishment between the cold war
aviation, while the apollo lunar module was designed and built by grumman. to develop the spaceflight skills and equipment required for a lunar mission, nasa initiated project gemini. using a modified air force titan ii launch vehicle, the gemini capsule could hold two astronauts for flights of over two weeks. gemini pioneered the use of fuel cells instead of batteries, and conducted the first american spacewalks and rendezvous operations. the ranger program was started in the 1950s as a response to soviet lunar exploration, however most missions ended in failure. the lunar orbiter program had greater success, mapping the surface in preparation for apollo landings, conducting meteoroid detection, and measuring radiation levels. the surveyor program conducted uncrewed lunar landings and takeoffs, as well as taking surface and regolith observations. despite the setback caused by the apollo 1 fire, which killed three astronauts, the program proceeded. apollo 8 was the first crewed spacecraft to leave low earth orbit and the first human spaceflight to reach the moon. the crew orbited the moon ten times on december 24 and 25, 1968, and then traveled safely back to earth. the three apollo 8 astronauts β frank borman, james lovell, and william anders β were the first humans to see the earth as a globe in space, the first to witness an earthrise, and the first to see and manually photograph the far side of the moon. the first lunar landing was conducted by apollo 11. commanded by neil armstrong with astronauts buzz aldrin and michael collins, apollo 11 was one of the most significant missions in nasa ' s history, marking the end of the space race when the soviet union gave up its lunar ambitions. as the first human to step on the surface of the moon, neil armstrong uttered the now famous words : that ' s one small step for man, one giant leap for mankind. nasa would conduct six total lunar landings as part of the apollo program, with apollo 17 concluding the program in 1972. = = = = end of apollo = = = = wernher von braun had advocated for nasa to develop a space station since the agency was created. in 1973, following the end of the apollo lunar missions, nasa launched its first space station, skylab, on the final launch of the saturn v. skylab reused a significant amount of apollo and saturn hardware, with a repurposed saturn v third stage serving as the primary module for the space station. damage to skylab during its launch required spacewalks to be performed by the first crew to make it habitable and
the origin of the martian moons deimos and phobos is controversial. one hypothesis for their origin is that they are captured asteroids, but the mechanism requires an extremely dense martian atmosphere, and the mechanism by which an asteroid in solar orbit could shed sufficient orbital energy to be captured into mars orbit has not been well elucidated. since the discovery by the space probe galileo that the asteroid ida has a moon " dactyl ", a significant number of asteroids have been discovered to have smaller asteroids in orbit about them. the existence of asteroid moons provides a mechanism for the capture of the martian moons ( and the small moons of the outer planets ). when a binary asteroid makes a close approach to a planet, tidal forces can strip the moon from the asteroid. depending on the phasing, the asteroid can then be captured. clearly, the same process can be used to explain the origin of any of the small moons in the solar system.
the lunar university network for astrophysics research ( lunar ) is a team of researchers and students at leading universities, nasa centers, and federal research laboratories undertaking investigations aimed at using the moon as a platform for space science. lunar research includes lunar interior physics & gravitation using lunar laser ranging ( llr ), low frequency cosmology and astrophysics ( lfca ), planetary science and the lunar ionosphere, radio heliophysics, and exploration science. the lunar team is exploring technologies that are likely to have a dual purpose, serving both exploration and science. there is a certain degree of commonality in much of lunar ' s research. specifically, the technology development for a lunar radio telescope involves elements from lfca, heliophysics, exploration science, and planetary science ; similarly the drilling technology developed for llr applies broadly to both exploration and lunar science.
recent surveys have revealed a lack of close - in planets around evolved stars more massive than 1. 2 msun. such planets are common around solar - mass stars. we have calculated the orbital evolution of planets around stars with a range of initial masses, and have shown how planetary orbits are affected by the evolution of the stars all the way to the tip of the red giant branch ( rgb ). we find that tidal interaction can lead to the engulfment of close - in planets by evolved stars. the engulfment is more efficient for more - massive planets and less - massive stars. these results may explain the observed semi - major axis distribution of planets around evolved stars with masses larger than 1. 5 msun. our results also suggest that massive planets may form more efficiently around intermediate - mass stars.
large scale manned space flight within the solar system is still confronted with the solution of two problems : 1. a propulsion system to transport large payloads with short transit times between different planetary orbits. 2. a cost effective lifting of large payloads into earth orbit. for the solution of the first problem a deuterium fusion bomb propulsion system is proposed where a thermonuclear detonation wave is ignited in a small cylindrical assembly of deuterium with a gigavolt - multimegampere proton beam, drawn from the magnetically insulated spacecraft acting in the ultrahigh vacuum of space as a gigavolt capacitor. for the solution of the second problem, the ignition is done by argon ion lasers driven by high explosives, with the lasers destroyed in the fusion explosion and becoming part of the exhaust.
the hun tian theory ), or as being without substance while the heavenly bodies float freely ( the hsuan yeh theory ), the earth was at all times flat, although perhaps bulging up slightly. the model of an egg was often used by chinese astronomers such as zhang heng ( 78 β 139 ad ) to describe the heavens as spherical : the heavens are like a hen ' s egg and as round as a crossbow bullet ; the earth is like the yolk of the egg, and lies in the centre. this analogy with a curved egg led some modern historians, notably joseph needham, to conjecture that chinese astronomers were, after all, aware of the earth ' s sphericity. the egg reference, however, was rather meant to clarify the relative position of the flat earth to the heavens : in a passage of zhang heng ' s cosmogony not translated by needham, zhang himself says : " heaven takes its body from the yang, so it is round and in motion. earth takes its body from the yin, so it is flat and quiescent ". the point of the egg analogy is simply to stress that the earth is completely enclosed by heaven, rather than merely covered from above as the kai tian describes. chinese astronomers, many of them brilliant men by any standards, continued to think in flat - earth terms until the seventeenth century ; this surprising fact might be the starting - point for a re - examination of the apparent facility with which the idea of a spherical earth found acceptance in fifth - century bc greece. further examples cited by needham supposed to demonstrate dissenting voices from the ancient chinese consensus actually refer without exception to the earth being square, not to it being flat. accordingly, the 13th - century scholar li ye, who argued that the movements of the round heaven would be hindered by a square earth, did not advocate a spherical earth, but rather that its edge should be rounded off so as to be circular. however, needham disagrees, affirming that li ye believed the earth to be spherical, similar in shape to the heavens but much smaller. this was preconceived by the 4th - century scholar yu xi, who argued for the infinity of outer space surrounding the earth and that the latter could be either square or round, in accordance to the shape of the heavens. when chinese geographers of the 17th century, influenced by european cartography and astronomy, showed the earth as a sphere that could be circumnavigated by sailing around the globe, they
the motion of celestial bodies through a higher power such as god. aristotle did not have the technological advancements that would have explained the motion of celestial bodies. in addition, aristotle had many views on the elements. he believed that everything was derived of the elements earth, water, air, fire, and lastly the aether. the aether was a celestial element, and therefore made up the matter of the celestial bodies. the elements of earth, water, air and fire were derived of a combination of two of the characteristics of hot, wet, cold, and dry, and all had their inevitable place and motion. the motion of these elements begins with earth being the closest to " the earth, " then water, air, fire, and finally aether. in addition to the makeup of all things, aristotle came up with theories as to why things did not return to their natural motion. he understood that water sits above earth, air above water, and fire above air in their natural state. he explained that although all elements must return to their natural state, the human body and other living things have a constraint on the elements β thus not allowing the elements making one who they are to return to their natural state. the important legacy of this period included substantial advances in factual knowledge, especially in anatomy, zoology, botany, mineralogy, geography, mathematics and astronomy ; an awareness of the importance of certain scientific problems, especially those related to the problem of change and its causes ; and a recognition of the methodological importance of applying mathematics to natural phenomena and of undertaking empirical research. in the hellenistic age scholars frequently employed the principles developed in earlier greek thought : the application of mathematics and deliberate empirical research, in their scientific investigations. thus, clear unbroken lines of influence lead from ancient greek and hellenistic philosophers, to medieval muslim philosophers and scientists, to the european renaissance and enlightenment, to the secular sciences of the modern day. neither reason nor inquiry began with the ancient greeks, but the socratic method did, along with the idea of forms, give great advances in geometry, logic, and the natural sciences. according to benjamin farrington, former professor of classics at swansea university : " men were weighing for thousands of years before archimedes worked out the laws of equilibrium ; they must have had practical and intuitional knowledge of the principals involved. what archimedes did was to sort out the theoretical implications of this practical knowledge and present the resulting body of knowledge as a logically coherent system. " and again : " with astonishment we find ourselves on the threshold of modern science
ultra high energy particles arrive at earth constantly. they provide a beam at energies higher than any man - made accelerator, but at a very low rate. two large experiments, the pierre auger observatory and the telescope array experiment, have been taking data for several years now covering together the whole sky. i summarize the most recent measurements from both experiments, i compare their results and, for a change, i highlight their agreements.
Question: Earth is about 81 times more massive than the Moon. How will the mass and weight of a rocket on the Moon compare to its mass and weight on Earth?
A) The mass and weight will be less on the Moon.
B) The mass and weight will be greater on the Moon.
C) The mass will be the same, but its weight will be less on the Moon.
D) The mass will be the same, but its weight will be greater on the Moon.
|
C) The mass will be the same, but its weight will be less on the Moon.
|
Context:
do not survive or become incapable of procreation. plants cannot continue the natural ripening or aging process. all these effects are beneficial to the consumer and the food industry, likewise. the amount of energy imparted for effective food irradiation is low compared to cooking the same ; even at a typical dose of 10 kgy most food, which is ( with regard to warming ) physically equivalent to water, would warm by only about 2. 5 Β°c ( 4. 5 Β°f ). the specialty of processing food by ionizing radiation is the fact, that the energy density per atomic transition is very high, it can cleave molecules and induce ionization ( hence the name ) which cannot be achieved by mere heating. this is the reason for new beneficial effects, however at the same time, for new concerns. the treatment of solid food by ionizing radiation can provide an effect similar to heat pasteurization of liquids, such as milk. however, the use of the term, cold pasteurization, to describe irradiated foods is controversial, because pasteurization and irradiation are fundamentally different processes, although the intended end results can in some cases be similar. detractors of food irradiation have concerns about the health hazards of induced radioactivity. a report for the industry advocacy group american council on science and health entitled " irradiated foods " states : " the types of radiation sources approved for the treatment of foods have specific energy levels well below that which would cause any element in food to become radioactive. food undergoing irradiation does not become any more radioactive than luggage passing through an airport x - ray scanner or teeth that have been x - rayed. " food irradiation is currently permitted by over 40 countries and volumes are estimated to exceed 500, 000 metric tons ( 490, 000 long tons ; 550, 000 short tons ) annually worldwide. food irradiation is essentially a non - nuclear technology ; it relies on the use of ionizing radiation which may be generated by accelerators for electrons and conversion into bremsstrahlung, but which may use also gamma - rays from nuclear decay. there is a worldwide industry for processing by ionizing radiation, the majority by number and by processing power using accelerators. food irradiation is only a niche application compared to medical supplies, plastic materials, raw materials, gemstones, cables and wires, etc. = = accidents = = nuclear accidents, because of the powerful forces involved, are often very dangerous. historically, the first incidents involved fatal
##tion, and pasteurization in order to become products that can be sold. there are three levels of food processing : primary, secondary, and tertiary. primary food processing involves turning agricultural products into other products that can be turned into food, secondary food processing is the making of food from readily available ingredients, and tertiary food processing is commercial production of ready - to eat or heat - and - serve foods. drying, pickling, salting, and fermenting foods were some of the oldest food processing techniques used to preserve food by preventing yeasts, molds, and bacteria to cause spoiling. methods for preserving food have evolved to meet current standards of food safety but still use the same processes as the past. biochemical engineers also work to improve the nutritional value of food products, such as in golden rice, which was developed to prevent vitamin a deficiency in certain areas where this was an issue. efforts to advance preserving technologies can also ensure lasting retention of nutrients as foods are stored. packaging plays a key role in preserving as well as ensuring the safety of the food by protecting the product from contamination, physical damage, and tampering. packaging can also make it easier to transport and serve food. a common job for biochemical engineers working in the food industry is to design ways to perform all these processes on a large scale in order to meet the demands of the population. responsibilities for this career path include designing and performing experiments, optimizing processes, consulting with groups to develop new technologies, and preparing project plans for equipment and facilities. = = = pharmaceuticals = = = in the pharmaceutical industry, bioprocess engineering plays a crucial role in the large - scale production of biopharmaceuticals, such as monoclonal antibodies, vaccines, and therapeutic proteins. the development and optimization of bioreactors and fermentation systems are essential for the mass production of these products, ensuring consistent quality and high yields. for example, recombinant proteins like insulin and erythropoietin are produced through cell culture systems using genetically modified cells. the bioprocess engineer β s role is to optimize variables like temperature, ph, nutrient availability, and oxygen levels to maximize the efficiency of these systems. the growing field of gene therapy also relies on bioprocessing techniques to produce viral vectors, which are used to deliver therapeutic genes to patients. this involves scaling up processes from laboratory to industrial scale while maintaining safety and regulatory compliance. as the demand for biopharmaceutical products increases, advancements
10 kgy most food, which is ( with regard to warming ) physically equivalent to water, would warm by only about 2. 5 Β°c ( 4. 5 Β°f ). the specialty of processing food by ionizing radiation is the fact, that the energy density per atomic transition is very high, it can cleave molecules and induce ionization ( hence the name ) which cannot be achieved by mere heating. this is the reason for new beneficial effects, however at the same time, for new concerns. the treatment of solid food by ionizing radiation can provide an effect similar to heat pasteurization of liquids, such as milk. however, the use of the term, cold pasteurization, to describe irradiated foods is controversial, because pasteurization and irradiation are fundamentally different processes, although the intended end results can in some cases be similar. detractors of food irradiation have concerns about the health hazards of induced radioactivity. a report for the industry advocacy group american council on science and health entitled " irradiated foods " states : " the types of radiation sources approved for the treatment of foods have specific energy levels well below that which would cause any element in food to become radioactive. food undergoing irradiation does not become any more radioactive than luggage passing through an airport x - ray scanner or teeth that have been x - rayed. " food irradiation is currently permitted by over 40 countries and volumes are estimated to exceed 500, 000 metric tons ( 490, 000 long tons ; 550, 000 short tons ) annually worldwide. food irradiation is essentially a non - nuclear technology ; it relies on the use of ionizing radiation which may be generated by accelerators for electrons and conversion into bremsstrahlung, but which may use also gamma - rays from nuclear decay. there is a worldwide industry for processing by ionizing radiation, the majority by number and by processing power using accelerators. food irradiation is only a niche application compared to medical supplies, plastic materials, raw materials, gemstones, cables and wires, etc. = = accidents = = nuclear accidents, because of the powerful forces involved, are often very dangerous. historically, the first incidents involved fatal radiation exposure. marie curie died from aplastic anemia which resulted from her high levels of exposure. two scientists, an american and canadian respectively, harry daghlian and louis slotin, died after mishandling the same plutonium mass. unlike conventional weapons, the intense light, heat, and explosive force is
process of lactic acid fermentation, which produced other preserved foods, such as soy sauce. fermentation was also used in this time period to produce leavened bread. although the process of fermentation was not fully understood until louis pasteur ' s work in 1857, it is still the first use of biotechnology to convert a food source into another form. before the time of charles darwin ' s work and life, animal and plant scientists had already used selective breeding. darwin added to that body of work with his scientific observations about the ability of science to change species. these accounts contributed to darwin ' s theory of natural selection. for thousands of years, humans have used selective breeding to improve the production of crops and livestock to use them for food. in selective breeding, organisms with desirable characteristics are mated to produce offspring with the same characteristics. for example, this technique was used with corn to produce the largest and sweetest crops. in the early twentieth century scientists gained a greater understanding of microbiology and explored ways of manufacturing specific products. in 1917, chaim weizmann first used a pure microbiological culture in an industrial process, that of manufacturing corn starch using clostridium acetobutylicum, to produce acetone, which the united kingdom desperately needed to manufacture explosives during world war i. biotechnology has also led to the development of antibiotics. in 1928, alexander fleming discovered the mold penicillium. his work led to the purification of the antibiotic formed by the mold by howard florey, ernst boris chain and norman heatley β to form what we today know as penicillin. in 1940, penicillin became available for medicinal use to treat bacterial infections in humans. the field of modern biotechnology is generally thought of as having been born in 1971 when paul berg ' s ( stanford ) experiments in gene splicing had early success. herbert w. boyer ( univ. calif. at san francisco ) and stanley n. cohen ( stanford ) significantly advanced the new technology in 1972 by transferring genetic material into a bacterium, such that the imported material would be reproduced. the commercial viability of a biotechnology industry was significantly expanded on june 16, 1980, when the united states supreme court ruled that a genetically modified microorganism could be patented in the case of diamond v. chakrabarty. indian - born ananda chakrabarty, working for general electric, had modified a bacterium ( of the genus pseudomonas ) capable of breaking down crude oil, which he proposed to
molecules and induce ionization ( hence the name ) which cannot be achieved by mere heating. this is the reason for new beneficial effects, however at the same time, for new concerns. the treatment of solid food by ionizing radiation can provide an effect similar to heat pasteurization of liquids, such as milk. however, the use of the term, cold pasteurization, to describe irradiated foods is controversial, because pasteurization and irradiation are fundamentally different processes, although the intended end results can in some cases be similar. detractors of food irradiation have concerns about the health hazards of induced radioactivity. a report for the industry advocacy group american council on science and health entitled " irradiated foods " states : " the types of radiation sources approved for the treatment of foods have specific energy levels well below that which would cause any element in food to become radioactive. food undergoing irradiation does not become any more radioactive than luggage passing through an airport x - ray scanner or teeth that have been x - rayed. " food irradiation is currently permitted by over 40 countries and volumes are estimated to exceed 500, 000 metric tons ( 490, 000 long tons ; 550, 000 short tons ) annually worldwide. food irradiation is essentially a non - nuclear technology ; it relies on the use of ionizing radiation which may be generated by accelerators for electrons and conversion into bremsstrahlung, but which may use also gamma - rays from nuclear decay. there is a worldwide industry for processing by ionizing radiation, the majority by number and by processing power using accelerators. food irradiation is only a niche application compared to medical supplies, plastic materials, raw materials, gemstones, cables and wires, etc. = = accidents = = nuclear accidents, because of the powerful forces involved, are often very dangerous. historically, the first incidents involved fatal radiation exposure. marie curie died from aplastic anemia which resulted from her high levels of exposure. two scientists, an american and canadian respectively, harry daghlian and louis slotin, died after mishandling the same plutonium mass. unlike conventional weapons, the intense light, heat, and explosive force is not the only deadly component to a nuclear weapon. approximately half of the deaths from hiroshima and nagasaki died two to five years afterward from radiation exposure. civilian nuclear and radiological accidents primarily involve nuclear power plants. most common are nuclear leaks that expose workers to hazardous material. a nuclear meltdown refers to the more serious hazard of
no offspring, to reduce the population. in industrial and food applications, radiation is used for sterilization of tools and equipment. an advantage is that the object may be sealed in plastic before sterilization. an emerging use in food production is the sterilization of food using food irradiation. food irradiation is the process of exposing food to ionizing radiation in order to destroy microorganisms, bacteria, viruses, or insects that might be present in the food. the radiation sources used include radioisotope gamma ray sources, x - ray generators and electron accelerators. further applications include sprout inhibition, delay of ripening, increase of juice yield, and improvement of re - hydration. irradiation is a more general term of deliberate exposure of materials to radiation to achieve a technical goal ( in this context ' ionizing radiation ' is implied ). as such it is also used on non - food items, such as medical hardware, plastics, tubes for gas - pipelines, hoses for floor - heating, shrink - foils for food packaging, automobile parts, wires and cables ( isolation ), tires, and even gemstones. compared to the amount of food irradiated, the volume of those every - day applications is huge but not noticed by the consumer. the genuine effect of processing food by ionizing radiation relates to damages to the dna, the basic genetic information for life. microorganisms can no longer proliferate and continue their malignant or pathogenic activities. spoilage causing micro - organisms cannot continue their activities. insects do not survive or become incapable of procreation. plants cannot continue the natural ripening or aging process. all these effects are beneficial to the consumer and the food industry, likewise. the amount of energy imparted for effective food irradiation is low compared to cooking the same ; even at a typical dose of 10 kgy most food, which is ( with regard to warming ) physically equivalent to water, would warm by only about 2. 5 Β°c ( 4. 5 Β°f ). the specialty of processing food by ionizing radiation is the fact, that the energy density per atomic transition is very high, it can cleave molecules and induce ionization ( hence the name ) which cannot be achieved by mere heating. this is the reason for new beneficial effects, however at the same time, for new concerns. the treatment of solid food by ionizing radiation can provide an effect similar to heat pasteurization of liquids, such as milk. however
as medical hardware, plastics, tubes for gas - pipelines, hoses for floor - heating, shrink - foils for food packaging, automobile parts, wires and cables ( isolation ), tires, and even gemstones. compared to the amount of food irradiated, the volume of those every - day applications is huge but not noticed by the consumer. the genuine effect of processing food by ionizing radiation relates to damages to the dna, the basic genetic information for life. microorganisms can no longer proliferate and continue their malignant or pathogenic activities. spoilage causing micro - organisms cannot continue their activities. insects do not survive or become incapable of procreation. plants cannot continue the natural ripening or aging process. all these effects are beneficial to the consumer and the food industry, likewise. the amount of energy imparted for effective food irradiation is low compared to cooking the same ; even at a typical dose of 10 kgy most food, which is ( with regard to warming ) physically equivalent to water, would warm by only about 2. 5 Β°c ( 4. 5 Β°f ). the specialty of processing food by ionizing radiation is the fact, that the energy density per atomic transition is very high, it can cleave molecules and induce ionization ( hence the name ) which cannot be achieved by mere heating. this is the reason for new beneficial effects, however at the same time, for new concerns. the treatment of solid food by ionizing radiation can provide an effect similar to heat pasteurization of liquids, such as milk. however, the use of the term, cold pasteurization, to describe irradiated foods is controversial, because pasteurization and irradiation are fundamentally different processes, although the intended end results can in some cases be similar. detractors of food irradiation have concerns about the health hazards of induced radioactivity. a report for the industry advocacy group american council on science and health entitled " irradiated foods " states : " the types of radiation sources approved for the treatment of foods have specific energy levels well below that which would cause any element in food to become radioactive. food undergoing irradiation does not become any more radioactive than luggage passing through an airport x - ray scanner or teeth that have been x - rayed. " food irradiation is currently permitted by over 40 countries and volumes are estimated to exceed 500, 000 metric tons ( 490, 000 long tons ; 550, 000 short tons ) annually worldwide. food irradiation
; however, a successful large - scale industrial application of the process was the development of continuous freeze drying of coffee. high - temperature short time processing β these processes, for the most part, are characterized by rapid heating and cooling, holding for a short time at a relatively high temperature and filling aseptically into sterile containers. decaffeination of coffee and tea β decaffeinated coffee and tea was first developed on a commercial basis in europe around 1900. the process is described in u. s. patent 897, 763. green coffee beans are treated with water, heat and solvents to remove the caffeine from the beans. process optimization β food technology now allows production of foods to be more efficient, oil saving technologies are now available on different forms. production methods and methodology have also become increasingly sophisticated. aseptic packaging β the process of filling a commercially sterile product into a sterile container and hermetically sealing the containers so that re - infection is prevented. thus, this results into a shelf stable product at ambient conditions. food irradiation β the process of exposing food and food packaging to ionizing radiation can effectively destroy organisms responsible for spoilage and foodborne illness and inhibit sprouting, extending shelf life. commercial fruit ripening rooms using ethylene as a plant hormone. food delivery β an order is typically made either through a restaurant or grocer ' s website or mobile app, or through a food ordering company. the ordered food is typically delivered in boxes or bags to the customer ' s doorsteps. = = categories = = technology has innovated these categories from the food industry : agricultural technology β or agtech, it is the use of technology in agriculture, horticulture, and aquaculture with the aim of improving yield, efficiency, and profitability. agricultural technology can be products, services or applications derived from agriculture that improve various input / output processes. food science β technology in this sector focuses on the development of new functional ingredients and alternative proteins. foodservice β technology innovated the way establishments prepare, supply, and serve food outside the home. there ' s a tendency to create the conditions for the restaurant of the future with robotics and cloudkitchens. consumer tech β technology allows what we call consumer electronics, which is the equipment of consumers with devices that facilitates the cooking process. food delivery β as the food delivery market is growing, companies and startups are rapidly revolutionizing the communication process between consumers and food establishments, with platform - to - consumer delivery as the
huge but not noticed by the consumer. the genuine effect of processing food by ionizing radiation relates to damages to the dna, the basic genetic information for life. microorganisms can no longer proliferate and continue their malignant or pathogenic activities. spoilage causing micro - organisms cannot continue their activities. insects do not survive or become incapable of procreation. plants cannot continue the natural ripening or aging process. all these effects are beneficial to the consumer and the food industry, likewise. the amount of energy imparted for effective food irradiation is low compared to cooking the same ; even at a typical dose of 10 kgy most food, which is ( with regard to warming ) physically equivalent to water, would warm by only about 2. 5 Β°c ( 4. 5 Β°f ). the specialty of processing food by ionizing radiation is the fact, that the energy density per atomic transition is very high, it can cleave molecules and induce ionization ( hence the name ) which cannot be achieved by mere heating. this is the reason for new beneficial effects, however at the same time, for new concerns. the treatment of solid food by ionizing radiation can provide an effect similar to heat pasteurization of liquids, such as milk. however, the use of the term, cold pasteurization, to describe irradiated foods is controversial, because pasteurization and irradiation are fundamentally different processes, although the intended end results can in some cases be similar. detractors of food irradiation have concerns about the health hazards of induced radioactivity. a report for the industry advocacy group american council on science and health entitled " irradiated foods " states : " the types of radiation sources approved for the treatment of foods have specific energy levels well below that which would cause any element in food to become radioactive. food undergoing irradiation does not become any more radioactive than luggage passing through an airport x - ray scanner or teeth that have been x - rayed. " food irradiation is currently permitted by over 40 countries and volumes are estimated to exceed 500, 000 metric tons ( 490, 000 long tons ; 550, 000 short tons ) annually worldwide. food irradiation is essentially a non - nuclear technology ; it relies on the use of ionizing radiation which may be generated by accelerators for electrons and conversion into bremsstrahlung, but which may use also gamma - rays from nuclear decay. there is a worldwide industry for processing by ionizing radiation, the majority by number
the celebrated franck - hertz experiment is reinterpreted by analogy with the glimmentladung experiment, formerly performed by heinrich hertz.
Question: Louis Pasteur discovered a process that allows milk to be stored in sealed containers for long periods of time without spoiling. The process, called pasteurization, involves heating the milk before placing it in the sealed containers.Which is the best explanation of how pasteurization works?
A) Heat strengthens molecules that prevent spoiling.
B) The chemical structure of the food changes with the heat.
C) Heat killed microorganisms that could cause spoiling.
D) The food particles are better mixed after heating.
|
C) Heat killed microorganisms that could cause spoiling.
|
Context:
weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under ordinary conditions, most of the materials brought down from the high lands by torrential water courses are carried forward by the main river to the sea, or partially strewn over flat alluvial plains during floods ; the size of the materials forming the bed of the river or borne along by the stream is gradually reduced on proceeding seawards, so that in the po river in italy, for instance, pebbles and gravel are found for about 140 miles below turin, sand along the next 100 miles, and silt and mud in the last 110 miles ( 176 km ). = = channelization = = the removal of obstructions, natural or artificial
and their competitive or mutualistic interactions with other species. some ecologists even rely on empirical data from indigenous people that is gathered by ethnobotanists. this information can relay a great deal of information on how the land once was thousands of years ago and how it has changed over that time. the goals of plant ecology are to understand the causes of their distribution patterns, productivity, environmental impact, evolution, and responses to environmental change. plants depend on certain edaphic ( soil ) and climatic factors in their environment but can modify these factors too. for example, they can change their environment ' s albedo, increase runoff interception, stabilise mineral soils and develop their organic content, and affect local temperature. plants compete with other organisms in their ecosystem for resources. they interact with their neighbours at a variety of spatial scales in groups, populations and communities that collectively constitute vegetation. regions with characteristic vegetation types and dominant plants as well as similar abiotic and biotic factors, climate, and geography make up biomes like tundra or tropical rainforest. herbivores eat plants, but plants can defend themselves and some species are parasitic or even carnivorous. other organisms form mutually beneficial relationships with plants. for example, mycorrhizal fungi and rhizobia provide plants with nutrients in exchange for food, ants are recruited by ant plants to provide protection, honey bees, bats and other animals pollinate flowers and humans and other animals act as dispersal vectors to spread spores and seeds. = = = plants, climate and environmental change = = = plant responses to climate and other environmental changes can inform our understanding of how these changes affect ecosystem function and productivity. for example, plant phenology can be a useful proxy for temperature in historical climatology, and the biological impact of climate change and global warming. palynology, the analysis of fossil pollen deposits in sediments from thousands or millions of years ago allows the reconstruction of past climates. estimates of atmospheric co2 concentrations since the palaeozoic have been obtained from stomatal densities and the leaf shapes and sizes of ancient land plants. ozone depletion can expose plants to higher levels of ultraviolet radiation - b ( uv - b ), resulting in lower growth rates. moreover, information from studies of community ecology, plant systematics, and taxonomy is essential to understanding vegetation change, habitat destruction and species extinction. = = genetics = = inheritance in plants follows the same fundamental principles of genetics as in other multicellular organisms. gregor mendel discovered the genetic laws of inheritance by studying
shuttle from the heat of re - entry into the earth ' s atmosphere. one example is reinforced carbon - carbon ( rcc ), the light gray material, which withstands re - entry temperatures up to 1, 510 Β°c ( 2, 750 Β°f ) and protects the space shuttle ' s wing leading edges and nose cap. rcc is a laminated composite material made from graphite rayon cloth and impregnated with a phenolic resin. after curing at high temperature in an autoclave, the laminate is pyrolized to convert the resin to carbon, impregnated with furfuryl alcohol in a vacuum chamber, and cured - pyrolized to convert the furfuryl alcohol to carbon. to provide oxidation resistance for reusability, the outer layers of the rcc are converted to silicon carbide. other examples can be seen in the " plastic " casings of television sets, cell - phones and so on. these plastic casings are usually a composite material made up of a thermoplastic matrix such as acrylonitrile butadiene styrene ( abs ) in which calcium carbonate chalk, talc, glass fibers or carbon fibers have been added for added strength, bulk, or electrostatic dispersion. these additions may be termed reinforcing fibers, or dispersants, depending on their purpose. = = = polymers = = = polymers are chemical compounds made up of a large number of identical components linked together like chains. polymers are the raw materials ( the resins ) used to make what are commonly called plastics and rubber. plastics and rubber are the final product, created after one or more polymers or additives have been added to a resin during processing, which is then shaped into a final form. plastics in former and in current widespread use include polyethylene, polypropylene, polyvinyl chloride ( pvc ), polystyrene, nylons, polyesters, acrylics, polyurethanes, and polycarbonates. rubbers include natural rubber, styrene - butadiene rubber, chloroprene, and butadiene rubber. plastics are generally classified as commodity, specialty and engineering plastics. polyvinyl chloride ( pvc ) is widely used, inexpensive, and annual production quantities are large. it lends itself to a vast array of applications, from artificial leather to electrical insulation and cabling, packaging, and containers. its fabrication and processing are simple and well - established.
navigable capabilities of rivers. in tropical countries subject to periodical rains, the rivers are in flood during the rainy season and have hardly any flow during the rest of the year, while in temperate regions, where the rainfall is more evenly distributed throughout the year, evaporation causes the available rainfall to be much less in hot summer weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under ordinary conditions, most of the materials brought down from the high lands by torrential water courses are carried forward by the main river to the sea, or partially strewn over flat alluvial plains during floods ; the size of the materials forming the bed of the river or borne along by the stream is gradually reduced on proceeding sea
thesis submitted for the degree of phd, queens university belfast, uk
ambient air ( see lockheed f - 117 nighthawk, rectangular nozzles on the lockheed martin f - 22 raptor, and serrated nozzle flaps on the lockheed martin f - 35 lightning ). often, cool air is deliberately injected into the exhaust flow to boost this process ( see ryan aqm - 91 firefly and northrop b - 2 spirit ). the stefan β boltzmann law shows how this results in less energy ( thermal radiation in infrared spectrum ) being released and thus reduces the heat signature. in some aircraft, the jet exhaust is vented above the wing surface to shield it from observers below, as in the lockheed f - 117 nighthawk, and the unstealthy fairchild republic a - 10 thunderbolt ii. to achieve infrared stealth, the exhaust gas is cooled to the temperatures where the brightest wavelengths it radiates are absorbed by atmospheric carbon dioxide and water vapor, greatly reducing the infrared visibility of the exhaust plume. another way to reduce the exhaust temperature is to circulate coolant fluids such as fuel inside the exhaust pipe, where the fuel tanks serve as heat sinks cooled by the flow of air along the wings. ground combat includes the use of both active and passive infrared sensors. thus, the united states marine corps ( usmc ) ground combat uniform requirements document specifies infrared reflective quality standards. = = reducing radio frequency ( rf ) emissions = = in addition to reducing infrared and acoustic emissions, a stealth vehicle must avoid radiating any other detectable energy, such as from onboard radars, communications systems, or rf leakage from electronics enclosures. the f - 117 uses passive infrared and low light level television sensor systems to aim its weapons and the f - 22 raptor has an advanced lpi radar which can illuminate enemy aircraft without triggering a radar warning receiver response. = = measuring = = the size of a target ' s image on radar is measured by the rcs, often represented by the symbol Ο and expressed in square meters. this does not equal geometric area. a perfectly conducting sphere of projected cross sectional area 1 m2 ( i. e. a diameter of 1. 13 m ) will have an rcs of 1 m2. note that for radar wavelengths much less than the diameter of the sphere, rcs is independent of frequency. conversely, a square flat plate of area 1 m2 will have an rcs of Ο = 4Ο a2 / Ξ»2 ( where a = area, Ξ» = wavelength ), or 13, 982 m2 at 10 ghz if the radar is perpendicular to the flat
uv ice photodesorption is an important non - thermal desorption pathway in many interstellar environments that has been invoked to explain observations of cold molecules in disks, clouds and cloud cores. systematic laboratory studies of the photodesorption rates, between 7 and 14 ev, from co : n2 binary ices, have been performed at the desirs vacuum uv beamline of the synchrotron facility soleil. the photodesorption spectral analysis demonstrates that the photodesorption process is indirect, i. e. the desorption is induced by a photon absorption in sub - surface molecular layers, while only surface molecules are actually desorbing. the photodesorption spectra of co and n2 in binary ices therefore depend on the absorption spectra of the dominant species in the subsurface ice layer, which implies that the photodesorption efficiency and energy dependence are dramatically different for mixed and layered ices compared to pure ices. in particular, a thin ( 1 - 2 ml ) n2 ice layer on top of co will effectively quench co photodesorption, while enhancing n2 photodesorption by a factors of a few ( compared to the pure ices ) when the ice is exposed to a typical dark cloud uv field, which may help to explain the different distributions of co and n2h + in molecular cloud cores. this indirect photodesorption mechanism may also explain observations of small amounts of complex organics in cold interstellar environments.
the gas giant planets in the solar system have a retinue of icy moons, and we expect giant exoplanets to have similar satellite systems. if a jupiter - like planet were to migrate toward its parent star the icy moons orbiting it would evaporate, creating atmospheres and possible habitable surface oceans. here, we examine how long the surface ice and possible oceans would last before being hydrodynamically lost to space. the hydrodynamic loss rate from the moons is determined, in large part, by the stellar flux available for absorption, which increases as the giant planet and icy moons migrate closer to the star. at some planet - star distance the stellar flux incident on the icy moons becomes so great that they enter a runaway greenhouse state. this runaway greenhouse state rapidly transfers all available surface water to the atmosphere as vapor, where it is easily lost from the small moons. however, for icy moons of ganymede ' s size around a sun - like star we found that surface water ( either ice or liquid ) can persist indefinitely outside the runaway greenhouse orbital distance. in contrast, the surface water on smaller moons of europa ' s size will only persist on timescales greater than 1 gyr at distances ranging 1. 49 to 0. 74 au around a sun - like star for bond albedos of 0. 2 and 0. 8, where the lower albedo becomes relevant if ice melts. consequently, small moons can lose their icy shells, which would create a torus of h atoms around their host planet that might be detectable in future observations.
industry is making composite materials. these are structured materials composed of two or more macroscopic phases. applications range from structural elements such as steel - reinforced concrete, to the thermal insulating tiles, which play a key and integral role in nasa ' s space shuttle thermal protection system, which is used to protect the surface of the shuttle from the heat of re - entry into the earth ' s atmosphere. one example is reinforced carbon - carbon ( rcc ), the light gray material, which withstands re - entry temperatures up to 1, 510 Β°c ( 2, 750 Β°f ) and protects the space shuttle ' s wing leading edges and nose cap. rcc is a laminated composite material made from graphite rayon cloth and impregnated with a phenolic resin. after curing at high temperature in an autoclave, the laminate is pyrolized to convert the resin to carbon, impregnated with furfuryl alcohol in a vacuum chamber, and cured - pyrolized to convert the furfuryl alcohol to carbon. to provide oxidation resistance for reusability, the outer layers of the rcc are converted to silicon carbide. other examples can be seen in the " plastic " casings of television sets, cell - phones and so on. these plastic casings are usually a composite material made up of a thermoplastic matrix such as acrylonitrile butadiene styrene ( abs ) in which calcium carbonate chalk, talc, glass fibers or carbon fibers have been added for added strength, bulk, or electrostatic dispersion. these additions may be termed reinforcing fibers, or dispersants, depending on their purpose. = = = polymers = = = polymers are chemical compounds made up of a large number of identical components linked together like chains. polymers are the raw materials ( the resins ) used to make what are commonly called plastics and rubber. plastics and rubber are the final product, created after one or more polymers or additives have been added to a resin during processing, which is then shaped into a final form. plastics in former and in current widespread use include polyethylene, polypropylene, polyvinyl chloride ( pvc ), polystyrene, nylons, polyesters, acrylics, polyurethanes, and polycarbonates. rubbers include natural rubber, styrene - butadiene rubber, chloroprene, and butadiene rubber. plastics are generally classified as commodity
or magnitude. magnitudes are always non - negative real numbers, and to any non - zero number there belongs a positive real number, its absolute value. for example, the absolute value of β3 and the absolute value of 3 are both equal to 3. this is written in symbols as | β3 | = 3 and | 3 | = 3. in general, any arbitrary real value can be specified by its magnitude and its sign. using the standard encoding, any real value is given by the product of the magnitude and the sign in standard encoding. this relation can be generalized to define a sign for complex numbers. since the real and complex numbers both form a field and contain the positive reals, they also contain the reciprocals of the magnitudes of all non - zero numbers. this means that any non - zero number may be multiplied with the reciprocal of its magnitude, that is, divided by its magnitude. it is immediate that the quotient of any non - zero real number by its magnitude yields exactly its sign. by analogy, the sign of a complex number z can be defined as the quotient of z and its magnitude | z |. the sign of a complex number is the exponential of the product of its argument with the imaginary unit. represents in some sense its complex argument. this is to be compared to the sign of real numbers, except with e i Ο = β 1. { \ displaystyle e ^ { i \ pi } = - 1. } for the definition of a complex sign - function. see Β§ complex sign function below. = = = sign functions = = = when dealing with numbers, it is often convenient to have their sign available as a number. this is accomplished by functions that extract the sign of any number, and map it to a predefined value before making it available for further calculations. for example, it might be advantageous to formulate an intricate algorithm for positive values only, and take care of the sign only afterwards. = = = = real sign function = = = = the sign function or signum function extracts the sign of a real number, by mapping the set of real numbers to the set of the three reals { β 1, 0, 1 }. { \ displaystyle \ { - 1, \ ; 0, \ ; 1 \ }. } it can be defined as follows : sgn : r β { β 1, 0, 1 } x β¦ sgn ( x ) = { β 1 if x < 0, 0 if x = 0
Question: Which characteristic would best help a land animal survive in a cold Arctic climate?
A) thick fur
B) large ears
C) smooth feet
D) slender body
|
A) thick fur
|
Context:
three of what is called the six simple machines, from which all machines are based. these machines are the inclined plane, the wedge, and the lever, which allowed the ancient egyptians to move millions of limestone blocks which weighed approximately 3. 5 tons ( 7, 000 lbs. ) each into place to create structures like the great pyramid of giza, which is 481 feet ( 147 meters ) high. they also made writing medium similar to paper from papyrus, which joshua mark states is the foundation for modern paper. papyrus is a plant ( cyperus papyrus ) which grew in plentiful amounts in the egyptian delta and throughout the nile river valley during ancient times. the papyrus was harvested by field workers and brought to processing centers where it was cut into thin strips. the strips were then laid - out side by side and covered in plant resin. the second layer of strips was laid on perpendicularly, then both pressed together until the sheet was dry. the sheets were then joined to form a roll and later used for writing. egyptian society made several significant advances during dynastic periods in many areas of technology. according to hossam elanzeery, they were the first civilization to use timekeeping devices such as sundials, shadow clocks, and obelisks and successfully leveraged their knowledge of astronomy to create a calendar model that society still uses today. they developed shipbuilding technology that saw them progress from papyrus reed vessels to cedar wood ships while also pioneering the use of rope trusses and stem - mounted rudders. the egyptians also used their knowledge of anatomy to lay the foundation for many modern medical techniques and practiced the earliest known version of neuroscience. elanzeery also states that they used and furthered mathematical science, as evidenced in the building of the pyramids. ancient egyptians also invented and pioneered many food technologies that have become the basis of modern food technology processes. based on paintings and reliefs found in tombs, as well as archaeological artifacts, scholars like paul t nicholson believe that the ancient egyptians established systematic farming practices, engaged in cereal processing, brewed beer and baked bread, processed meat, practiced viticulture and created the basis for modern wine production, and created condiments to complement, preserve and mask the flavors of their food. = = = = indus valley = = = = the indus valley civilization, situated in a resource - rich area ( in modern pakistan and northwestern india ), is notable for its early application of city planning, sanitation technologies, and plumbing. indus valley construction and architecture, called ' vaastu
as subjects perceive the sensory world, different stimuli elicit a number of neural representations. here, a subjective distance between stimuli is defined, measuring the degree of similarity between the underlying representations. as an example, the subjective distance between different locations in space is calculated from the activity of rodent hippocampal place cells, and lateral septal cells. such a distance is compared to the real distance, between locations. as the number of sampled neurons increases, the subjective distance shows a tendency to resemble the metrics of real space.
##ent governmental regulations. some of these requirements include : seat belt and air bag functionality testing, front and side - impact testing, and tests of rollover resistance. assessments are done with various methods and tools, including computer crash simulation ( typically finite element analysis ), crash - test dummy, and partial system sled and full vehicle crashes. fuel economy / emissions : fuel economy is the measured fuel efficiency of the vehicle in miles per gallon or kilometers per liter. emissions - testing covers the measurement of vehicle emissions, including hydrocarbons, nitrogen oxides ( nox ), carbon monoxide ( co ), carbon dioxide ( co2 ), and evaporative emissions. nvh engineering ( noise, vibration, and harshness ) : nvh involves customer feedback ( both tactile [ felt ] and audible [ heard ] ) concerning a vehicle. while sound can be interpreted as a rattle, squeal, or hot, a tactile response can be seat vibration or a buzz in the steering wheel. this feedback is generated by components either rubbing, vibrating, or rotating. nvh response can be classified in various ways : powertrain nvh, road noise, wind noise, component noise, and squeak and rattle. note, there are both good and bad nvh qualities. the nvh engineer works to either eliminate bad nvh or change the " bad nvh " to good ( i. e., exhaust tones ). vehicle electronics : automotive electronics is an increasingly important aspect of automotive engineering. modern vehicles employ dozens of electronic systems. these systems are responsible for operational controls such as the throttle, brake and steering controls ; as well as many comfort - and - convenience systems such as the hvac, infotainment, and lighting systems. it would not be possible for automobiles to meet modern safety and fuel - economy requirements without electronic controls. performance : performance is a measurable and testable value of a vehicle ' s ability to perform in various conditions. performance can be considered in a wide variety of tasks, but it generally considers how quickly a car can accelerate ( e. g. standing start 1 / 4 mile elapsed time, 0 β 60 mph, etc. ), its top speed, how short and quickly a car can come to a complete stop from a set speed ( e. g. 70 - 0 mph ), how much g - force a car can generate without losing grip, recorded lap - times, cornering speed, brake fade, etc. performance can also reflect the
the thickness and the density of the material to be measured. the method is used for containers of liquids or of grainy substances thickness gauges : if the material is of constant density, the signal measured by the radiation detector depends on the thickness of the material. this is useful for continuous production, like of paper, rubber, etc. electrostatic control - to avoid the build - up of static electricity in production of paper, plastics, synthetic textiles, etc., a ribbon - shaped source of the alpha emitter 241am can be placed close to the material at the end of the production line. the source ionizes the air to remove electric charges on the material. radioactive tracers - since radioactive isotopes behave, chemically, mostly like the inactive element, the behavior of a certain chemical substance can be followed by tracing the radioactivity. examples : adding a gamma tracer to a gas or liquid in a closed system makes it possible to find a hole in a tube. adding a tracer to the surface of the component of a motor makes it possible to measure wear by measuring the activity of the lubricating oil. oil and gas exploration - nuclear well logging is used to help predict the commercial viability of new or existing wells. the technology involves the use of a neutron or gamma - ray source and a radiation detector which are lowered into boreholes to determine the properties of the surrounding rock such as porosity and lithography. [ 1 ] road construction - nuclear moisture / density gauges are used to determine the density of soils, asphalt, and concrete. typically a cesium - 137 source is used. = = = commercial applications = = = radioluminescence tritium illumination : tritium is used with phosphor in rifle sights to increase nighttime firing accuracy. some runway markers and building exit signs use the same technology, to remain illuminated during blackouts. betavoltaics. smoke detector : an ionization smoke detector includes a tiny mass of radioactive americium - 241, which is a source of alpha radiation. two ionisation chambers are placed next to each other. both contain a small source of 241am that gives rise to a small constant current. one is closed and serves for comparison, the other is open to ambient air ; it has a gridded electrode. when smoke enters the open chamber, the current is disrupted as the smoke particles attach to the charged ions and restore them to a neutral electrical state. this reduces the current in the open chamber. when the current drops below a certain threshold, the
time estimates range from 5, 500 to 3, 000 bce with most experts putting it closer to 4, 000 bce. the oldest artifacts with drawings depicting wheeled carts date from about 3, 500 bce. more recently, the oldest - known wooden wheel in the world as of 2024 was found in the ljubljana marsh of slovenia ; austrian experts have established that the wheel is between 5, 100 and 5, 350 years old. the invention of the wheel revolutionized trade and war. it did not take long to discover that wheeled wagons could be used to carry heavy loads. the ancient sumerians used a potter ' s wheel and may have invented it. a stone pottery wheel found in the city - state of ur dates to around 3, 429 bce, and even older fragments of wheel - thrown pottery have been found in the same area. fast ( rotary ) potters ' wheels enabled early mass production of pottery, but it was the use of the wheel as a transformer of energy ( through water wheels, windmills, and even treadmills ) that revolutionized the application of nonhuman power sources. the first two - wheeled carts were derived from travois and were first used in mesopotamia and iran in around 3, 000 bce. the oldest known constructed roadways are the stone - paved streets of the city - state of ur, dating to c. 4, 000 bce, and timber roads leading through the swamps of glastonbury, england, dating to around the same period. the first long - distance road, which came into use around 3, 500 bce, spanned 2, 400 km from the persian gulf to the mediterranean sea, but was not paved and was only partially maintained. in around 2, 000 bce, the minoans on the greek island of crete built a 50 km road leading from the palace of gortyn on the south side of the island, through the mountains, to the palace of knossos on the north side of the island. unlike the earlier road, the minoan road was completely paved. ancient minoan private homes had running water. a bathtub virtually identical to modern ones was unearthed at the palace of knossos. several minoan private homes also had toilets, which could be flushed by pouring water down the drain. the ancient romans had many public flush toilets, which emptied into an extensive sewage system. the primary sewer in rome was the cloaca maxima ; construction began on it in the sixth century bce and it is still in use today. the ancient romans
; austrian experts have established that the wheel is between 5, 100 and 5, 350 years old. the invention of the wheel revolutionized trade and war. it did not take long to discover that wheeled wagons could be used to carry heavy loads. the ancient sumerians used a potter ' s wheel and may have invented it. a stone pottery wheel found in the city - state of ur dates to around 3, 429 bce, and even older fragments of wheel - thrown pottery have been found in the same area. fast ( rotary ) potters ' wheels enabled early mass production of pottery, but it was the use of the wheel as a transformer of energy ( through water wheels, windmills, and even treadmills ) that revolutionized the application of nonhuman power sources. the first two - wheeled carts were derived from travois and were first used in mesopotamia and iran in around 3, 000 bce. the oldest known constructed roadways are the stone - paved streets of the city - state of ur, dating to c. 4, 000 bce, and timber roads leading through the swamps of glastonbury, england, dating to around the same period. the first long - distance road, which came into use around 3, 500 bce, spanned 2, 400 km from the persian gulf to the mediterranean sea, but was not paved and was only partially maintained. in around 2, 000 bce, the minoans on the greek island of crete built a 50 km road leading from the palace of gortyn on the south side of the island, through the mountains, to the palace of knossos on the north side of the island. unlike the earlier road, the minoan road was completely paved. ancient minoan private homes had running water. a bathtub virtually identical to modern ones was unearthed at the palace of knossos. several minoan private homes also had toilets, which could be flushed by pouring water down the drain. the ancient romans had many public flush toilets, which emptied into an extensive sewage system. the primary sewer in rome was the cloaca maxima ; construction began on it in the sixth century bce and it is still in use today. the ancient romans also had a complex system of aqueducts, which were used to transport water across long distances. the first roman aqueduct was built in 312 bce. the eleventh and final ancient roman aqueduct was built in 226 ce. put together, the roman aqueducts extended over 450 km, but less than 70 km of this was above ground
the large scale pattern in the arrival directions of extragalactic cosmic rays that reach the earth is different from that of the flux arriving to the halo of the galaxy as a result of the propagation through the galactic magnetic field. two different effects are relevant in this process : deflections of trajectories and ( de ) acceleration by the electric field component due to the galactic rotation. the deflection of the cosmic ray trajectories makes the flux intensity arriving to the halo from some direction to appear reaching the earth from another direction. this applies to any intrinsic anisotropy in the extragalactic distribution or, even in the absence of intrinsic anisotropies, to the dipolar compton - getting anisotropy induced when the observer is moving with respect to the cosmic rays rest frame. for an observer moving with the solar system, cosmic rays traveling through far away regions of the galaxy also experience an electric force coming from the relative motion ( due to the rotation of the galaxy ) of the local system in which the field can be considered as being purely magnetic. this produces small changes in the particles momentum that can originate large scale anisotropies even for an isotropic extragalactic flux.
strongly. but the history of thought shows us that many people were totally committed to absurd beliefs. if the strengths of beliefs were a hallmark of knowledge, we should have to rank some tales about demons, angels, devils, and of heaven and hell as knowledge. scientists, on the other hand, are very sceptical even of their best theories. newton ' s is the most powerful theory science has yet produced, but newton himself never believed that bodies attract each other at a distance. so no degree of commitment to beliefs makes them knowledge. indeed, the hallmark of scientific behaviour is a certain scepticism even towards one ' s most cherished theories. blind commitment to a theory is not an intellectual virtue : it is an intellectual crime. thus a statement may be pseudoscientific even if it is eminently ' plausible ' and everybody believes in it, and it may be scientifically valuable even if it is unbelievable and nobody believes in it. a theory may even be of supreme scientific value even if no one understands it, let alone believes in it. the boundary between science and pseudoscience is disputed and difficult to determine analytically, even after more than a century of study by philosophers of science and scientists, and despite some basic agreements on the fundamentals of the scientific method. the concept of pseudoscience rests on an understanding that the scientific method has been misrepresented or misapplied with respect to a given theory, but many philosophers of science maintain that different kinds of methods are held as appropriate across different fields and different eras of human history. according to lakatos, the typical descriptive unit of great scientific achievements is not an isolated hypothesis but " a powerful problem - solving machinery, which, with the help of sophisticated mathematical techniques, digests anomalies and even turns them into positive evidence ". to popper, pseudoscience uses induction to generate theories, and only performs experiments to seek to verify them. to popper, falsifiability is what determines the scientific status of a theory. taking a historical approach, kuhn observed that scientists did not follow popper ' s rule, and might ignore falsifying data, unless overwhelming. to kuhn, puzzle - solving within a paradigm is science. lakatos attempted to resolve this debate, by suggesting history shows that science occurs in research programmes, competing according to how progressive they are. the leading idea of a programme could evolve, driven by its heuristic to make predictions that can be supported by evidence. feyerabend claimed that
5 to 20 meters. north america and japan use a frequency of 315 mhz, while europe uses 433. 92 and 868 mhz. some models can also remotely start the engine, to warm up the car. a security concern with all keyless entry systems is a replay attack, in which a thief uses a special receiver ( " code grabber " ) to record the radio signal during opening, which can later be replayed to open the door. to prevent this, keyless systems use a rolling code system in which a pseudorandom number generator in the remote control generates a different random key each time it is used. to prevent thieves from simulating the pseudorandom generator to calculate the next key, the radio signal is also encrypted. garage door opener β a short - range handheld transmitter which can open or close a building ' s electrically operated garage door from outside, so the owner can open the door upon arrival, and close it after departure. when a button is pressed the control transmits a coded fsk radio signal to a receiver in the opener, raising or lowering the door. modern openers use 310, 315 or 390 mhz. to prevent a thief using a replay attack, modern openers use a rolling code system. radio - controlled models β a popular hobby is playing with radio - controlled model boats, cars, airplanes, and helicopters ( quadcopters ) which are controlled by radio signals from a handheld console with a joystick. most recent transmitters use the 2. 4 ghz ism band with multiple control channels modulated with pwm, pcm or fsk. wireless doorbell β a residential doorbell that uses wireless technology to eliminate the need to run wires through the building walls. it consists of a doorbell button beside the door containing a small battery powered transmitter. when the doorbell is pressed it sends a signal to a receiver inside the house with a speaker that sounds chimes to indicate someone is at the door. they usually use the 2. 4 ghz ism band. the frequency channel used can usually be changed by the owner in case another nearby doorbell is using the same channel. = = = = scientific research = = = = radio astronomy is the scientific study of radio waves emitted by astronomical objects. radio astronomers use radio telescopes, large radio antennas and receivers, to receive and study the radio waves from astronomical radio sources. since astronomical radio sources are so far away, the radio waves from them are extremely weak, requiring extremely sensitive receivers, and radio telescopes are the most sensitive radio receivers in existence. they use
. this, he argued, would have been more persuasive and would have produced less controversy. the use of poetic imagery based on the concepts of the macrocosm and microcosm, " as above so below " to decide meaning such as edward w. james ' example of " mars above is red, so mars below means blood and war ", is a false cause fallacy. : 26 many astrologers claim that astrology is scientific. if one were to attempt to try to explain it scientifically, there are only four fundamental forces ( conventionally ), limiting the choice of possible natural mechanisms. : 65 some astrologers have proposed conventional causal agents such as electromagnetism and gravity. the strength of these forces drops off with distance. : 65 scientists reject these proposed mechanisms as implausible since, for example, the magnetic field, when measured from earth, of a large but distant planet such as jupiter is far smaller than that produced by ordinary household appliances. astronomer phil plait noted that in terms of magnitude, the sun is the only object with an electromagnetic field of note, but astrology isn ' t based just off the sun alone. : 65 while astrologers could try to suggest a fifth force, this is inconsistent with the trends in physics with the unification of electromagnetism and the weak force into the electroweak force. if the astrologer insisted on being inconsistent with the current understanding and evidential basis of physics, that would be an extraordinary claim. : 65 it would also be inconsistent with the other forces which drop off with distance. : 65 if distance is irrelevant, then, logically, all objects in space should be taken into account. : 66 carl jung sought to invoke synchronicity, the claim that two events have some sort of acausal connection, to explain the lack of statistically significant results on astrology from a single study he conducted. however, synchronicity itself is considered neither testable nor falsifiable. the study was subsequently heavily criticised for its non - random sample and its use of statistics and also its lack of consistency with astrology. = = psychology = = psychological studies have not found any robust relationship between astrological signs and life outcomes. for example, a study showed that zodiac signs are no more effective than random numbers in predicting subjective well - being and quality of life. it has also been shown that confirmation bias is a psychological factor that contributes to belief in astrology. : 344 : 180 β 181 :
Question: What is the most reasonable explanation for why a toy car rolls farther on a wood floor than on a thick carpet?
A) The car weighs more on the carpet.
B) The car weighs more on the floor.
C) The carpet has more resistance.
D) The floor has more traction.
|
C) The carpet has more resistance.
|
Context:
frits went. the first known auxin, indole - 3 - acetic acid ( iaa ), which promotes cell growth, was only isolated from plants about 50 years later. this compound mediates the tropic responses of shoots and roots towards light and gravity. the finding in 1939 that plant callus could be maintained in culture containing iaa, followed by the observation in 1947 that it could be induced to form roots and shoots by controlling the concentration of growth hormones were key steps in the development of plant biotechnology and genetic modification. cytokinins are a class of plant hormones named for their control of cell division ( especially cytokinesis ). the natural cytokinin zeatin was discovered in corn, zea mays, and is a derivative of the purine adenine. zeatin is produced in roots and transported to shoots in the xylem where it promotes cell division, bud development, and the greening of chloroplasts. the gibberelins, such as gibberelic acid are diterpenes synthesised from acetyl coa via the mevalonate pathway. they are involved in the promotion of germination and dormancy - breaking in seeds, in regulation of plant height by controlling stem elongation and the control of flowering. abscisic acid ( aba ) occurs in all land plants except liverworts, and is synthesised from carotenoids in the chloroplasts and other plastids. it inhibits cell division, promotes seed maturation, and dormancy, and promotes stomatal closure. it was so named because it was originally thought to control abscission. ethylene is a gaseous hormone that is produced in all higher plant tissues from methionine. it is now known to be the hormone that stimulates or regulates fruit ripening and abscission, and it, or the synthetic growth regulator ethephon which is rapidly metabolised to produce ethylene, are used on industrial scale to promote ripening of cotton, pineapples and other climacteric crops. another class of phytohormones is the jasmonates, first isolated from the oil of jasminum grandiflorum which regulates wound responses in plants by unblocking the expression of genes required in the systemic acquired resistance response to pathogen attack. in addition to being the primary energy source for plants, light functions as a signalling device, providing information to the plant, such as how
venus flytrap and bladderworts, and the pollinia of orchids. the hypothesis that plant growth and development is coordinated by plant hormones or plant growth regulators first emerged in the late 19th century. darwin experimented on the movements of plant shoots and roots towards light and gravity, and concluded " it is hardly an exaggeration to say that the tip of the radicle.. acts like the brain of one of the lower animals.. directing the several movements ". about the same time, the role of auxins ( from the greek auxein, to grow ) in control of plant growth was first outlined by the dutch scientist frits went. the first known auxin, indole - 3 - acetic acid ( iaa ), which promotes cell growth, was only isolated from plants about 50 years later. this compound mediates the tropic responses of shoots and roots towards light and gravity. the finding in 1939 that plant callus could be maintained in culture containing iaa, followed by the observation in 1947 that it could be induced to form roots and shoots by controlling the concentration of growth hormones were key steps in the development of plant biotechnology and genetic modification. cytokinins are a class of plant hormones named for their control of cell division ( especially cytokinesis ). the natural cytokinin zeatin was discovered in corn, zea mays, and is a derivative of the purine adenine. zeatin is produced in roots and transported to shoots in the xylem where it promotes cell division, bud development, and the greening of chloroplasts. the gibberelins, such as gibberelic acid are diterpenes synthesised from acetyl coa via the mevalonate pathway. they are involved in the promotion of germination and dormancy - breaking in seeds, in regulation of plant height by controlling stem elongation and the control of flowering. abscisic acid ( aba ) occurs in all land plants except liverworts, and is synthesised from carotenoids in the chloroplasts and other plastids. it inhibits cell division, promotes seed maturation, and dormancy, and promotes stomatal closure. it was so named because it was originally thought to control abscission. ethylene is a gaseous hormone that is produced in all higher plant tissues from methionine. it is now known to be the hormone that stimulates or regulates fruit ripening and abscission,
##aggeration to say that the tip of the radicle.. acts like the brain of one of the lower animals.. directing the several movements ". about the same time, the role of auxins ( from the greek auxein, to grow ) in control of plant growth was first outlined by the dutch scientist frits went. the first known auxin, indole - 3 - acetic acid ( iaa ), which promotes cell growth, was only isolated from plants about 50 years later. this compound mediates the tropic responses of shoots and roots towards light and gravity. the finding in 1939 that plant callus could be maintained in culture containing iaa, followed by the observation in 1947 that it could be induced to form roots and shoots by controlling the concentration of growth hormones were key steps in the development of plant biotechnology and genetic modification. cytokinins are a class of plant hormones named for their control of cell division ( especially cytokinesis ). the natural cytokinin zeatin was discovered in corn, zea mays, and is a derivative of the purine adenine. zeatin is produced in roots and transported to shoots in the xylem where it promotes cell division, bud development, and the greening of chloroplasts. the gibberelins, such as gibberelic acid are diterpenes synthesised from acetyl coa via the mevalonate pathway. they are involved in the promotion of germination and dormancy - breaking in seeds, in regulation of plant height by controlling stem elongation and the control of flowering. abscisic acid ( aba ) occurs in all land plants except liverworts, and is synthesised from carotenoids in the chloroplasts and other plastids. it inhibits cell division, promotes seed maturation, and dormancy, and promotes stomatal closure. it was so named because it was originally thought to control abscission. ethylene is a gaseous hormone that is produced in all higher plant tissues from methionine. it is now known to be the hormone that stimulates or regulates fruit ripening and abscission, and it, or the synthetic growth regulator ethephon which is rapidly metabolised to produce ethylene, are used on industrial scale to promote ripening of cotton, pineapples and other climacteric crops. another class of phytohormones is the jasmonates, first isolated
of embryophytes ( land plants ) is called phytology. bryology is the study of mosses ( and in the broader sense also liverworts and hornworts ). pteridology ( or filicology ) is the study of ferns and allied plants. a number of other taxa of ranks varying from family to subgenus have terms for their study, including agrostology ( or graminology ) for the study of grasses, synantherology for the study of composites, and batology for the study of brambles. study can also be divided by guild rather than clade or grade. for example, dendrology is the study of woody plants. many divisions of biology have botanical subfields. these are commonly denoted by prefixing the word plant ( e. g. plant taxonomy, plant ecology, plant anatomy, plant morphology, plant systematics ), or prefixing or substituting the prefix phyto - ( e. g. phytochemistry, phytogeography ). the study of fossil plants is called palaeobotany. other fields are denoted by adding or substituting the word botany ( e. g. systematic botany ). phytosociology is a subfield of plant ecology that classifies and studies communities of plants. the intersection of fields from the above pair of categories gives rise to fields such as bryogeography, the study of the distribution of mosses. different parts of plants also give rise to their own subfields, including xylology, carpology ( or fructology ), and palynology, these being the study of wood, fruit and pollen / spores respectively. botany also overlaps on the one hand with agriculture, horticulture and silviculture, and on the other hand with medicine and pharmacology, giving rise to fields such as agronomy, horticultural botany, phytopathology, and phytopharmacology. = = scope and importance = = the study of plants is vital because they underpin almost all animal life on earth by generating a large proportion of the oxygen and food that provide humans and other organisms with aerobic respiration with the chemical energy they need to exist. plants, algae and cyanobacteria are the major groups of organisms that carry out photosynthesis, a process that uses the energy of sunlight to convert water and carbon dioxide into sugars that can be used both as a source of chemical energy and of organic molecules that are used in
, dendrology is the study of woody plants. many divisions of biology have botanical subfields. these are commonly denoted by prefixing the word plant ( e. g. plant taxonomy, plant ecology, plant anatomy, plant morphology, plant systematics ), or prefixing or substituting the prefix phyto - ( e. g. phytochemistry, phytogeography ). the study of fossil plants is called palaeobotany. other fields are denoted by adding or substituting the word botany ( e. g. systematic botany ). phytosociology is a subfield of plant ecology that classifies and studies communities of plants. the intersection of fields from the above pair of categories gives rise to fields such as bryogeography, the study of the distribution of mosses. different parts of plants also give rise to their own subfields, including xylology, carpology ( or fructology ), and palynology, these being the study of wood, fruit and pollen / spores respectively. botany also overlaps on the one hand with agriculture, horticulture and silviculture, and on the other hand with medicine and pharmacology, giving rise to fields such as agronomy, horticultural botany, phytopathology, and phytopharmacology. = = scope and importance = = the study of plants is vital because they underpin almost all animal life on earth by generating a large proportion of the oxygen and food that provide humans and other organisms with aerobic respiration with the chemical energy they need to exist. plants, algae and cyanobacteria are the major groups of organisms that carry out photosynthesis, a process that uses the energy of sunlight to convert water and carbon dioxide into sugars that can be used both as a source of chemical energy and of organic molecules that are used in the structural components of cells. as a by - product of photosynthesis, plants release oxygen into the atmosphere, a gas that is required by nearly all living things to carry out cellular respiration. in addition, they are influential in the global carbon and water cycles and plant roots bind and stabilise soils, preventing soil erosion. plants are crucial to the future of human society as they provide food, oxygen, biochemicals, and products for people, as well as creating and preserving soil. historically, all living things were classified as either animals or plants and botany covered the study of all organisms not considered animals. botanists examine both
ranks varying from family to subgenus have terms for their study, including agrostology ( or graminology ) for the study of grasses, synantherology for the study of composites, and batology for the study of brambles. study can also be divided by guild rather than clade or grade. for example, dendrology is the study of woody plants. many divisions of biology have botanical subfields. these are commonly denoted by prefixing the word plant ( e. g. plant taxonomy, plant ecology, plant anatomy, plant morphology, plant systematics ), or prefixing or substituting the prefix phyto - ( e. g. phytochemistry, phytogeography ). the study of fossil plants is called palaeobotany. other fields are denoted by adding or substituting the word botany ( e. g. systematic botany ). phytosociology is a subfield of plant ecology that classifies and studies communities of plants. the intersection of fields from the above pair of categories gives rise to fields such as bryogeography, the study of the distribution of mosses. different parts of plants also give rise to their own subfields, including xylology, carpology ( or fructology ), and palynology, these being the study of wood, fruit and pollen / spores respectively. botany also overlaps on the one hand with agriculture, horticulture and silviculture, and on the other hand with medicine and pharmacology, giving rise to fields such as agronomy, horticultural botany, phytopathology, and phytopharmacology. = = scope and importance = = the study of plants is vital because they underpin almost all animal life on earth by generating a large proportion of the oxygen and food that provide humans and other organisms with aerobic respiration with the chemical energy they need to exist. plants, algae and cyanobacteria are the major groups of organisms that carry out photosynthesis, a process that uses the energy of sunlight to convert water and carbon dioxide into sugars that can be used both as a source of chemical energy and of organic molecules that are used in the structural components of cells. as a by - product of photosynthesis, plants release oxygen into the atmosphere, a gas that is required by nearly all living things to carry out cellular respiration. in addition, they are influential in the global carbon and water cycles and plant roots bind and stabilise soils, preventing
cellular and molecular biology of cereals, grasses and monocots generally. model plants such as arabidopsis thaliana are used for studying the molecular biology of plant cells and the chloroplast. ideally, these organisms have small genomes that are well known or completely sequenced, small stature and short generation times. corn has been used to study mechanisms of photosynthesis and phloem loading of sugar in c4 plants. the single celled green alga chlamydomonas reinhardtii, while not an embryophyte itself, contains a green - pigmented chloroplast related to that of land plants, making it useful for study. a red alga cyanidioschyzon merolae has also been used to study some basic chloroplast functions. spinach, peas, soybeans and a moss physcomitrella patens are commonly used to study plant cell biology. agrobacterium tumefaciens, a soil rhizosphere bacterium, can attach to plant cells and infect them with a callus - inducing ti plasmid by horizontal gene transfer, causing a callus infection called crown gall disease. schell and van montagu ( 1977 ) hypothesised that the ti plasmid could be a natural vector for introducing the nif gene responsible for nitrogen fixation in the root nodules of legumes and other plant species. today, genetic modification of the ti plasmid is one of the main techniques for introduction of transgenes to plants and the creation of genetically modified crops. = = = epigenetics = = = epigenetics is the study of heritable changes in gene function that cannot be explained by changes in the underlying dna sequence but cause the organism ' s genes to behave ( or " express themselves " ) differently. one example of epigenetic change is the marking of the genes by dna methylation which determines whether they will be expressed or not. gene expression can also be controlled by repressor proteins that attach to silencer regions of the dna and prevent that region of the dna code from being expressed. epigenetic marks may be added or removed from the dna during programmed stages of development of the plant, and are responsible, for example, for the differences between anthers, petals and normal leaves, despite the fact that they all have the same underlying genetic code. epigenetic changes may be temporary or may remain through successive cell divisions for the remainder of
from the oil of jasminum grandiflorum which regulates wound responses in plants by unblocking the expression of genes required in the systemic acquired resistance response to pathogen attack. in addition to being the primary energy source for plants, light functions as a signalling device, providing information to the plant, such as how much sunlight the plant receives each day. this can result in adaptive changes in a process known as photomorphogenesis. phytochromes are the photoreceptors in a plant that are sensitive to light. = = plant anatomy and morphology = = plant anatomy is the study of the structure of plant cells and tissues, whereas plant morphology is the study of their external form. all plants are multicellular eukaryotes, their dna stored in nuclei. the characteristic features of plant cells that distinguish them from those of animals and fungi include a primary cell wall composed of the polysaccharides cellulose, hemicellulose and pectin, larger vacuoles than in animal cells and the presence of plastids with unique photosynthetic and biosynthetic functions as in the chloroplasts. other plastids contain storage products such as starch ( amyloplasts ) or lipids ( elaioplasts ). uniquely, streptophyte cells and those of the green algal order trentepohliales divide by construction of a phragmoplast as a template for building a cell plate late in cell division. the bodies of vascular plants including clubmosses, ferns and seed plants ( gymnosperms and angiosperms ) generally have aerial and subterranean subsystems. the shoots consist of stems bearing green photosynthesising leaves and reproductive structures. the underground vascularised roots bear root hairs at their tips and generally lack chlorophyll. non - vascular plants, the liverworts, hornworts and mosses do not produce ground - penetrating vascular roots and most of the plant participates in photosynthesis. the sporophyte generation is nonphotosynthetic in liverworts but may be able to contribute part of its energy needs by photosynthesis in mosses and hornworts. the root system and the shoot system are interdependent β the usually nonphotosynthetic root system depends on the shoot system for food, and the usually photosynthetic shoot system depends on water and minerals from the root system. cells in each system are capable
groups of organisms. divisions related to the broader historical sense of botany include bacteriology, mycology ( or fungology ), and phycology β respectively, the study of bacteria, fungi, and algae β with lichenology as a subfield of mycology. the narrower sense of botany as the study of embryophytes ( land plants ) is called phytology. bryology is the study of mosses ( and in the broader sense also liverworts and hornworts ). pteridology ( or filicology ) is the study of ferns and allied plants. a number of other taxa of ranks varying from family to subgenus have terms for their study, including agrostology ( or graminology ) for the study of grasses, synantherology for the study of composites, and batology for the study of brambles. study can also be divided by guild rather than clade or grade. for example, dendrology is the study of woody plants. many divisions of biology have botanical subfields. these are commonly denoted by prefixing the word plant ( e. g. plant taxonomy, plant ecology, plant anatomy, plant morphology, plant systematics ), or prefixing or substituting the prefix phyto - ( e. g. phytochemistry, phytogeography ). the study of fossil plants is called palaeobotany. other fields are denoted by adding or substituting the word botany ( e. g. systematic botany ). phytosociology is a subfield of plant ecology that classifies and studies communities of plants. the intersection of fields from the above pair of categories gives rise to fields such as bryogeography, the study of the distribution of mosses. different parts of plants also give rise to their own subfields, including xylology, carpology ( or fructology ), and palynology, these being the study of wood, fruit and pollen / spores respectively. botany also overlaps on the one hand with agriculture, horticulture and silviculture, and on the other hand with medicine and pharmacology, giving rise to fields such as agronomy, horticultural botany, phytopathology, and phytopharmacology. = = scope and importance = = the study of plants is vital because they underpin almost all animal life on earth by generating a large proportion of the oxygen and food that provide humans and other organisms with aerobic respiration with the chemical
, the other can often regrow it. in fact it is possible to grow an entire plant from a single leaf, as is the case with plants in streptocarpus sect. saintpaulia, or even a single cell β which can dedifferentiate into a callus ( a mass of unspecialised cells ) that can grow into a new plant. in vascular plants, the xylem and phloem are the conductive tissues that transport resources between shoots and roots. roots are often adapted to store food such as sugars or starch, as in sugar beets and carrots. stems mainly provide support to the leaves and reproductive structures, but can store water in succulent plants such as cacti, food as in potato tubers, or reproduce vegetatively as in the stolons of strawberry plants or in the process of layering. leaves gather sunlight and carry out photosynthesis. large, flat, flexible, green leaves are called foliage leaves. gymnosperms, such as conifers, cycads, ginkgo, and gnetophytes are seed - producing plants with open seeds. angiosperms are seed - producing plants that produce flowers and have enclosed seeds. woody plants, such as azaleas and oaks, undergo a secondary growth phase resulting in two additional types of tissues : wood ( secondary xylem ) and bark ( secondary phloem and cork ). all gymnosperms and many angiosperms are woody plants. some plants reproduce sexually, some asexually, and some via both means. although reference to major morphological categories such as root, stem, leaf, and trichome are useful, one has to keep in mind that these categories are linked through intermediate forms so that a continuum between the categories results. furthermore, structures can be seen as processes, that is, process combinations. = = systematic botany = = systematic botany is part of systematic biology, which is concerned with the range and diversity of organisms and their relationships, particularly as determined by their evolutionary history. it involves, or is related to, biological classification, scientific taxonomy and phylogenetics. biological classification is the method by which botanists group organisms into categories such as genera or species. biological classification is a form of scientific taxonomy. modern taxonomy is rooted in the work of carl linnaeus, who grouped species according to shared physical characteristics. these groupings have since been revised to align better with the darwinian principle of common descent β grouping organisms
Question: What part of the scientific method involves measuring the growth of a plant over a one-month period?
A) analysis
B) observation
C) conclusion
D) hypothesis
|
B) observation
|
Context:
inflammatory reactions. because of this, they are of great interest for gene delivery and tissue engineering scaffolds. most hydroxyapatite ceramics are very porous and lack mechanical strength and are used to coat metal orthopedic devices to aid in forming a bond to bone or as bone fillers. they are also used as fillers for orthopedic plastic screws to aid in reducing the inflammation and increase absorption of these plastic materials. work is being done to make strong, fully dense nano crystalline hydroxyapatite ceramic materials for orthopedic weight bearing devices, replacing foreign metal and plastic orthopedic materials with a synthetic, but naturally occurring, bone mineral. ultimately these ceramic materials may be used as bone replacements or with the incorporation of protein collagens, synthetic bones. durable actinide - containing ceramic materials have many applications such as in nuclear fuels for burning excess pu and in chemically - inert sources of alpha irradiation for power supply of unmanned space vehicles or to produce electricity for microelectronic devices. both use and disposal of radioactive actinides require their immobilization in a durable host material. nuclear waste long - lived radionuclides such as actinides are immobilized using chemically - durable crystalline materials based on polycrystalline ceramics and large single crystals. alumina ceramics are widely utilized in the chemical industry due to their excellent chemical stability and high resistance to corrosion. it is used as acid - resistant pump impellers and pump bodies, ensuring long - lasting performance in transferring aggressive fluids. they are also used in acid - carrying pipe linings to prevent contamination and maintain fluid purity, which is crucial in industries like pharmaceuticals and food processing. valves made from alumina ceramics demonstrate exceptional durability and resistance to chemical attack, making them reliable for controlling the flow of corrosive liquids. = = glass - ceramics = = glass - ceramic materials share many properties with both glasses and ceramics. glass - ceramics have an amorphous phase and one or more crystalline phases and are produced by a so - called " controlled crystallization ", which is typically avoided in glass manufacturing. glass - ceramics often contain a crystalline phase which constitutes anywhere from 30 % [ m / m ] to 90 % [ m / m ] of its composition by volume, yielding an array of materials with interesting thermomechanical properties. in the processing of glass - ceramics, molten glass is cooled down gradually before reheating and annealing. in this heat
as dental implants and synthetic bones. hydroxyapatite, the natural mineral component of bone, has been made synthetically from a number of biological and chemical sources and can be formed into ceramic materials. orthopedic implants made from these materials bond readily to bone and other tissues in the body without rejection or inflammatory reactions. because of this, they are of great interest for gene delivery and tissue engineering scaffolds. most hydroxyapatite ceramics are very porous and lack mechanical strength and are used to coat metal orthopedic devices to aid in forming a bond to bone or as bone fillers. they are also used as fillers for orthopedic plastic screws to aid in reducing the inflammation and increase absorption of these plastic materials. work is being done to make strong, fully dense nano crystalline hydroxyapatite ceramic materials for orthopedic weight bearing devices, replacing foreign metal and plastic orthopedic materials with a synthetic, but naturally occurring, bone mineral. ultimately these ceramic materials may be used as bone replacements or with the incorporation of protein collagens, synthetic bones. durable actinide - containing ceramic materials have many applications such as in nuclear fuels for burning excess pu and in chemically - inert sources of alpha irradiation for power supply of unmanned space vehicles or to produce electricity for microelectronic devices. both use and disposal of radioactive actinides require their immobilization in a durable host material. nuclear waste long - lived radionuclides such as actinides are immobilized using chemically - durable crystalline materials based on polycrystalline ceramics and large single crystals. alumina ceramics are widely utilized in the chemical industry due to their excellent chemical stability and high resistance to corrosion. it is used as acid - resistant pump impellers and pump bodies, ensuring long - lasting performance in transferring aggressive fluids. they are also used in acid - carrying pipe linings to prevent contamination and maintain fluid purity, which is crucial in industries like pharmaceuticals and food processing. valves made from alumina ceramics demonstrate exceptional durability and resistance to chemical attack, making them reliable for controlling the flow of corrosive liquids. = = glass - ceramics = = glass - ceramic materials share many properties with both glasses and ceramics. glass - ceramics have an amorphous phase and one or more crystalline phases and are produced by a so - called " controlled crystallization ", which is typically avoided in glass manufacturing. glass - ceramics often contain a crystalline phase
. currently, even blades made of advanced metal alloys used in the engines ' hot section require cooling and careful limiting of operating temperatures. turbine engines made with ceramics could operate more efficiently, giving aircraft greater range and payload for a set amount of fuel. recently, there have been advances in ceramics which include bio - ceramics, such as dental implants and synthetic bones. hydroxyapatite, the natural mineral component of bone, has been made synthetically from a number of biological and chemical sources and can be formed into ceramic materials. orthopedic implants made from these materials bond readily to bone and other tissues in the body without rejection or inflammatory reactions. because of this, they are of great interest for gene delivery and tissue engineering scaffolds. most hydroxyapatite ceramics are very porous and lack mechanical strength and are used to coat metal orthopedic devices to aid in forming a bond to bone or as bone fillers. they are also used as fillers for orthopedic plastic screws to aid in reducing the inflammation and increase absorption of these plastic materials. work is being done to make strong, fully dense nano crystalline hydroxyapatite ceramic materials for orthopedic weight bearing devices, replacing foreign metal and plastic orthopedic materials with a synthetic, but naturally occurring, bone mineral. ultimately these ceramic materials may be used as bone replacements or with the incorporation of protein collagens, synthetic bones. durable actinide - containing ceramic materials have many applications such as in nuclear fuels for burning excess pu and in chemically - inert sources of alpha irradiation for power supply of unmanned space vehicles or to produce electricity for microelectronic devices. both use and disposal of radioactive actinides require their immobilization in a durable host material. nuclear waste long - lived radionuclides such as actinides are immobilized using chemically - durable crystalline materials based on polycrystalline ceramics and large single crystals. alumina ceramics are widely utilized in the chemical industry due to their excellent chemical stability and high resistance to corrosion. it is used as acid - resistant pump impellers and pump bodies, ensuring long - lasting performance in transferring aggressive fluids. they are also used in acid - carrying pipe linings to prevent contamination and maintain fluid purity, which is crucial in industries like pharmaceuticals and food processing. valves made from alumina ceramics demonstrate exceptional durability and resistance to chemical attack, making them reliable for controlling the flow of corrosive liquids. =
a graviton laser works, in principle, by the stimulated emission of coherent gravitons from a lasing medium. for significant amplification, we must have a very long path length and / or very high densities. black holes and the existence of weakly interacting sub - ev dark matter particles ( wisps ) solve both of these obstacles. orbiting trajectories for massless particles around black holes are well understood \ cite { mtw } and allow for arbitrarily long graviton path lengths. superradiance from kerr black holes of wisps can provide the sufficiently high density \ cite { abh }. this suggests that black holes can act as efficient graviton lasers. thus directed graviton laser beams have been emitted since the beginning of the universe and give rise to new sources of gravitational wave signals. to be in the path of particularly harmfully amplified graviton death rays will not be pleasant.
of ceramic parts in the requisite precision and durability is difficult. imperfection in the ceramic leads to cracks, which can lead to potentially dangerous equipment failure. such engines are possible in laboratory settings, but mass - production is not feasible with current technology. work is being done in developing ceramic parts for gas turbine engines. currently, even blades made of advanced metal alloys used in the engines ' hot section require cooling and careful limiting of operating temperatures. turbine engines made with ceramics could operate more efficiently, giving aircraft greater range and payload for a set amount of fuel. recently, there have been advances in ceramics which include bio - ceramics, such as dental implants and synthetic bones. hydroxyapatite, the natural mineral component of bone, has been made synthetically from a number of biological and chemical sources and can be formed into ceramic materials. orthopedic implants made from these materials bond readily to bone and other tissues in the body without rejection or inflammatory reactions. because of this, they are of great interest for gene delivery and tissue engineering scaffolds. most hydroxyapatite ceramics are very porous and lack mechanical strength and are used to coat metal orthopedic devices to aid in forming a bond to bone or as bone fillers. they are also used as fillers for orthopedic plastic screws to aid in reducing the inflammation and increase absorption of these plastic materials. work is being done to make strong, fully dense nano crystalline hydroxyapatite ceramic materials for orthopedic weight bearing devices, replacing foreign metal and plastic orthopedic materials with a synthetic, but naturally occurring, bone mineral. ultimately these ceramic materials may be used as bone replacements or with the incorporation of protein collagens, synthetic bones. durable actinide - containing ceramic materials have many applications such as in nuclear fuels for burning excess pu and in chemically - inert sources of alpha irradiation for power supply of unmanned space vehicles or to produce electricity for microelectronic devices. both use and disposal of radioactive actinides require their immobilization in a durable host material. nuclear waste long - lived radionuclides such as actinides are immobilized using chemically - durable crystalline materials based on polycrystalline ceramics and large single crystals. alumina ceramics are widely utilized in the chemical industry due to their excellent chemical stability and high resistance to corrosion. it is used as acid - resistant pump impellers and pump bodies, ensuring long - lasting performance in transferring aggressive
strongly. but the history of thought shows us that many people were totally committed to absurd beliefs. if the strengths of beliefs were a hallmark of knowledge, we should have to rank some tales about demons, angels, devils, and of heaven and hell as knowledge. scientists, on the other hand, are very sceptical even of their best theories. newton ' s is the most powerful theory science has yet produced, but newton himself never believed that bodies attract each other at a distance. so no degree of commitment to beliefs makes them knowledge. indeed, the hallmark of scientific behaviour is a certain scepticism even towards one ' s most cherished theories. blind commitment to a theory is not an intellectual virtue : it is an intellectual crime. thus a statement may be pseudoscientific even if it is eminently ' plausible ' and everybody believes in it, and it may be scientifically valuable even if it is unbelievable and nobody believes in it. a theory may even be of supreme scientific value even if no one understands it, let alone believes in it. the boundary between science and pseudoscience is disputed and difficult to determine analytically, even after more than a century of study by philosophers of science and scientists, and despite some basic agreements on the fundamentals of the scientific method. the concept of pseudoscience rests on an understanding that the scientific method has been misrepresented or misapplied with respect to a given theory, but many philosophers of science maintain that different kinds of methods are held as appropriate across different fields and different eras of human history. according to lakatos, the typical descriptive unit of great scientific achievements is not an isolated hypothesis but " a powerful problem - solving machinery, which, with the help of sophisticated mathematical techniques, digests anomalies and even turns them into positive evidence ". to popper, pseudoscience uses induction to generate theories, and only performs experiments to seek to verify them. to popper, falsifiability is what determines the scientific status of a theory. taking a historical approach, kuhn observed that scientists did not follow popper ' s rule, and might ignore falsifying data, unless overwhelming. to kuhn, puzzle - solving within a paradigm is science. lakatos attempted to resolve this debate, by suggesting history shows that science occurs in research programmes, competing according to how progressive they are. the leading idea of a programme could evolve, driven by its heuristic to make predictions that can be supported by evidence. feyerabend claimed that
the theory outright... lakatos sought to reconcile the rationalism of popperian falsificationism with what seemed to be its own refutation by history ". many philosophers have tried to solve the problem of demarcation in the following terms : a statement constitutes knowledge if sufficiently many people believe it sufficiently strongly. but the history of thought shows us that many people were totally committed to absurd beliefs. if the strengths of beliefs were a hallmark of knowledge, we should have to rank some tales about demons, angels, devils, and of heaven and hell as knowledge. scientists, on the other hand, are very sceptical even of their best theories. newton ' s is the most powerful theory science has yet produced, but newton himself never believed that bodies attract each other at a distance. so no degree of commitment to beliefs makes them knowledge. indeed, the hallmark of scientific behaviour is a certain scepticism even towards one ' s most cherished theories. blind commitment to a theory is not an intellectual virtue : it is an intellectual crime. thus a statement may be pseudoscientific even if it is eminently ' plausible ' and everybody believes in it, and it may be scientifically valuable even if it is unbelievable and nobody believes in it. a theory may even be of supreme scientific value even if no one understands it, let alone believes in it. the boundary between science and pseudoscience is disputed and difficult to determine analytically, even after more than a century of study by philosophers of science and scientists, and despite some basic agreements on the fundamentals of the scientific method. the concept of pseudoscience rests on an understanding that the scientific method has been misrepresented or misapplied with respect to a given theory, but many philosophers of science maintain that different kinds of methods are held as appropriate across different fields and different eras of human history. according to lakatos, the typical descriptive unit of great scientific achievements is not an isolated hypothesis but " a powerful problem - solving machinery, which, with the help of sophisticated mathematical techniques, digests anomalies and even turns them into positive evidence ". to popper, pseudoscience uses induction to generate theories, and only performs experiments to seek to verify them. to popper, falsifiability is what determines the scientific status of a theory. taking a historical approach, kuhn observed that scientists did not follow popper ' s rule, and might ignore falsifying data, unless overwhelming. to kuhn, puzzle - solving within
is also higher at high temperature, as shown by carnot ' s theorem. in a conventional metallic engine, much of the energy released from the fuel must be dissipated as waste heat in order to prevent a meltdown of the metallic parts. despite all of these desirable properties, such engines are not in production because the manufacturing of ceramic parts in the requisite precision and durability is difficult. imperfection in the ceramic leads to cracks, which can lead to potentially dangerous equipment failure. such engines are possible in laboratory settings, but mass - production is not feasible with current technology. work is being done in developing ceramic parts for gas turbine engines. currently, even blades made of advanced metal alloys used in the engines ' hot section require cooling and careful limiting of operating temperatures. turbine engines made with ceramics could operate more efficiently, giving aircraft greater range and payload for a set amount of fuel. recently, there have been advances in ceramics which include bio - ceramics, such as dental implants and synthetic bones. hydroxyapatite, the natural mineral component of bone, has been made synthetically from a number of biological and chemical sources and can be formed into ceramic materials. orthopedic implants made from these materials bond readily to bone and other tissues in the body without rejection or inflammatory reactions. because of this, they are of great interest for gene delivery and tissue engineering scaffolds. most hydroxyapatite ceramics are very porous and lack mechanical strength and are used to coat metal orthopedic devices to aid in forming a bond to bone or as bone fillers. they are also used as fillers for orthopedic plastic screws to aid in reducing the inflammation and increase absorption of these plastic materials. work is being done to make strong, fully dense nano crystalline hydroxyapatite ceramic materials for orthopedic weight bearing devices, replacing foreign metal and plastic orthopedic materials with a synthetic, but naturally occurring, bone mineral. ultimately these ceramic materials may be used as bone replacements or with the incorporation of protein collagens, synthetic bones. durable actinide - containing ceramic materials have many applications such as in nuclear fuels for burning excess pu and in chemically - inert sources of alpha irradiation for power supply of unmanned space vehicles or to produce electricity for microelectronic devices. both use and disposal of radioactive actinides require their immobilization in a durable host material. nuclear waste long - lived radionuclides such as actinides are immobilized using chemical
. this, he argued, would have been more persuasive and would have produced less controversy. the use of poetic imagery based on the concepts of the macrocosm and microcosm, " as above so below " to decide meaning such as edward w. james ' example of " mars above is red, so mars below means blood and war ", is a false cause fallacy. : 26 many astrologers claim that astrology is scientific. if one were to attempt to try to explain it scientifically, there are only four fundamental forces ( conventionally ), limiting the choice of possible natural mechanisms. : 65 some astrologers have proposed conventional causal agents such as electromagnetism and gravity. the strength of these forces drops off with distance. : 65 scientists reject these proposed mechanisms as implausible since, for example, the magnetic field, when measured from earth, of a large but distant planet such as jupiter is far smaller than that produced by ordinary household appliances. astronomer phil plait noted that in terms of magnitude, the sun is the only object with an electromagnetic field of note, but astrology isn ' t based just off the sun alone. : 65 while astrologers could try to suggest a fifth force, this is inconsistent with the trends in physics with the unification of electromagnetism and the weak force into the electroweak force. if the astrologer insisted on being inconsistent with the current understanding and evidential basis of physics, that would be an extraordinary claim. : 65 it would also be inconsistent with the other forces which drop off with distance. : 65 if distance is irrelevant, then, logically, all objects in space should be taken into account. : 66 carl jung sought to invoke synchronicity, the claim that two events have some sort of acausal connection, to explain the lack of statistically significant results on astrology from a single study he conducted. however, synchronicity itself is considered neither testable nor falsifiable. the study was subsequently heavily criticised for its non - random sample and its use of statistics and also its lack of consistency with astrology. = = psychology = = psychological studies have not found any robust relationship between astrological signs and life outcomes. for example, a study showed that zodiac signs are no more effective than random numbers in predicting subjective well - being and quality of life. it has also been shown that confirmation bias is a psychological factor that contributes to belief in astrology. : 344 : 180 β 181 :
what are the implications if the total ' information ' in the universe is conserved? black holes might be ' logic gates ' recomputing the ' lost information ' from incoming ' signals ' from outside their event horizons into outgoing ' signals ' representing evaporative or radiative decay ' products ' of the reconfiguration process of the black hole quantum logic ' gate '. apparent local imbalances in the information flow can be corrected by including the effects of the coupling of the vacuum ' reservoir ' of information as part of the total information involved in any evolutionary process. in this way perhaps the ' vacuum ' computes the future of the observable universe.
Question: Which statement on a toothpaste package best supports the claim of fighting cavities?
A) delivers great new flavor
B) patented triple cleansing action
C) guaranteed to make teeth sparkle
D) certified by the American Dental Association
|
D) certified by the American Dental Association
|
Context:
is said to have occurred. a chemical reaction is therefore a concept related to the " reaction " of a substance when it comes in close contact with another, whether as a mixture or a solution ; exposure to some form of energy, or both. it results in some energy exchange between the constituents of the reaction as well as with the system environment, which may be designed vessels β often laboratory glassware. chemical reactions can result in the formation or dissociation of molecules, that is, molecules breaking apart to form two or more molecules or rearrangement of atoms within or across molecules. chemical reactions usually involve the making or breaking of chemical bonds. oxidation, reduction, dissociation, acid β base neutralization and molecular rearrangement are some examples of common chemical reactions. a chemical reaction can be symbolically depicted through a chemical equation. while in a non - nuclear chemical reaction the number and kind of atoms on both sides of the equation are equal, for a nuclear reaction this holds true only for the nuclear particles viz. protons and neutrons. the sequence of steps in which the reorganization of chemical bonds may be taking place in the course of a chemical reaction is called its mechanism. a chemical reaction can be envisioned to take place in a number of steps, each of which may have a different speed. many reaction intermediates with variable stability can thus be envisaged during the course of a reaction. reaction mechanisms are proposed to explain the kinetics and the relative product mix of a reaction. many physical chemists specialize in exploring and proposing the mechanisms of various chemical reactions. several empirical rules, like the woodward β hoffmann rules often come in handy while proposing a mechanism for a chemical reaction. according to the iupac gold book, a chemical reaction is " a process that results in the interconversion of chemical species. " accordingly, a chemical reaction may be an elementary reaction or a stepwise reaction. an additional caveat is made, in that this definition includes cases where the interconversion of conformers is experimentally observable. such detectable chemical reactions normally involve sets of molecular entities as indicated by this definition, but it is often conceptually convenient to use the term also for changes involving single molecular entities ( i. e. ' microscopic chemical events ' ). = = = ions and salts = = = an ion is a charged species, an atom or a molecule, that has lost or gained one or more electrons. when an atom loses an electron and thus has more protons than electrons, the atom is a positively charged
with the system environment, which may be designed vessels β often laboratory glassware. chemical reactions can result in the formation or dissociation of molecules, that is, molecules breaking apart to form two or more molecules or rearrangement of atoms within or across molecules. chemical reactions usually involve the making or breaking of chemical bonds. oxidation, reduction, dissociation, acid β base neutralization and molecular rearrangement are some examples of common chemical reactions. a chemical reaction can be symbolically depicted through a chemical equation. while in a non - nuclear chemical reaction the number and kind of atoms on both sides of the equation are equal, for a nuclear reaction this holds true only for the nuclear particles viz. protons and neutrons. the sequence of steps in which the reorganization of chemical bonds may be taking place in the course of a chemical reaction is called its mechanism. a chemical reaction can be envisioned to take place in a number of steps, each of which may have a different speed. many reaction intermediates with variable stability can thus be envisaged during the course of a reaction. reaction mechanisms are proposed to explain the kinetics and the relative product mix of a reaction. many physical chemists specialize in exploring and proposing the mechanisms of various chemical reactions. several empirical rules, like the woodward β hoffmann rules often come in handy while proposing a mechanism for a chemical reaction. according to the iupac gold book, a chemical reaction is " a process that results in the interconversion of chemical species. " accordingly, a chemical reaction may be an elementary reaction or a stepwise reaction. an additional caveat is made, in that this definition includes cases where the interconversion of conformers is experimentally observable. such detectable chemical reactions normally involve sets of molecular entities as indicated by this definition, but it is often conceptually convenient to use the term also for changes involving single molecular entities ( i. e. ' microscopic chemical events ' ). = = = ions and salts = = = an ion is a charged species, an atom or a molecule, that has lost or gained one or more electrons. when an atom loses an electron and thus has more protons than electrons, the atom is a positively charged ion or cation. when an atom gains an electron and thus has more electrons than protons, the atom is a negatively charged ion or anion. cations and anions can form a crystalline lattice of neutral salts, such as the na + and clβ ions forming sodium chloride, or nacl. examples of
chemistry is the scientific study of the properties and behavior of matter. it is a physical science within the natural sciences that studies the chemical elements that make up matter and compounds made of atoms, molecules and ions : their composition, structure, properties, behavior and the changes they undergo during reactions with other substances. chemistry also addresses the nature of chemical bonds in chemical compounds. in the scope of its subject, chemistry occupies an intermediate position between physics and biology. it is sometimes called the central science because it provides a foundation for understanding both basic and applied scientific disciplines at a fundamental level. for example, chemistry explains aspects of plant growth ( botany ), the formation of igneous rocks ( geology ), how atmospheric ozone is formed and how environmental pollutants are degraded ( ecology ), the properties of the soil on the moon ( cosmochemistry ), how medications work ( pharmacology ), and how to collect dna evidence at a crime scene ( forensics ). chemistry has existed under various names since ancient times. it has evolved, and now chemistry encompasses various areas of specialisation, or subdisciplines, that continue to increase in number and interrelate to create further interdisciplinary fields of study. the applications of various fields of chemistry are used frequently for economic purposes in the chemical industry. = = etymology = = the word chemistry comes from a modification during the renaissance of the word alchemy, which referred to an earlier set of practices that encompassed elements of chemistry, metallurgy, philosophy, astrology, astronomy, mysticism, and medicine. alchemy is often associated with the quest to turn lead or other base metals into gold, though alchemists were also interested in many of the questions of modern chemistry. the modern word alchemy in turn is derived from the arabic word al - kimia ( Ψ§ΩΩΫΩ
ΫΨ§Ψ‘ ). this may have egyptian origins since al - kimia is derived from the ancient greek ΟΞ·ΞΌΞΉΞ±, which is in turn derived from the word kemet, which is the ancient name of egypt in the egyptian language. alternately, al - kimia may derive from ΟημΡια ' cast together '. = = modern principles = = the current model of atomic structure is the quantum mechanical model. traditional chemistry starts with the study of elementary particles, atoms, molecules, substances, metals, crystals and other aggregates of matter. matter can be studied in solid, liquid, gas and plasma states, in isolation or in combination. the interactions, reactions and transformations that
, but the other isolated chemical elements consist of either molecules or networks of atoms bonded to each other in some way. identifiable molecules compose familiar substances such as water, air, and many organic compounds like alcohol, sugar, gasoline, and the various pharmaceuticals. however, not all substances or chemical compounds consist of discrete molecules, and indeed most of the solid substances that make up the solid crust, mantle, and core of the earth are chemical compounds without molecules. these other types of substances, such as ionic compounds and network solids, are organized in such a way as to lack the existence of identifiable molecules per se. instead, these substances are discussed in terms of formula units or unit cells as the smallest repeating structure within the substance. examples of such substances are mineral salts ( such as table salt ), solids like carbon and diamond, metals, and familiar silica and silicate minerals such as quartz and granite. one of the main characteristics of a molecule is its geometry often called its structure. while the structure of diatomic, triatomic or tetra - atomic molecules may be trivial, ( linear, angular pyramidal etc. ) the structure of polyatomic molecules, that are constituted of more than six atoms ( of several elements ) can be crucial for its chemical nature. = = = = substance and mixture = = = = a chemical substance is a kind of matter with a definite composition and set of properties. a collection of substances is called a mixture. examples of mixtures are air and alloys. = = = = mole and amount of substance = = = = the mole is a unit of measurement that denotes an amount of substance ( also called chemical amount ). one mole is defined to contain exactly 6. 02214076Γ1023 particles ( atoms, molecules, ions, or electrons ), where the number of particles per mole is known as the avogadro constant. molar concentration is the amount of a particular substance per volume of solution, and is commonly reported in mol / dm3. = = = phase = = = in addition to the specific chemical properties that distinguish different chemical classifications, chemicals can exist in several phases. for the most part, the chemical classifications are independent of these bulk phase classifications ; however, some more exotic phases are incompatible with certain chemical properties. a phase is a set of states of a chemical system that have similar bulk structural properties, over a range of conditions, such as pressure or temperature. physical properties, such as density and refractive index tend to fall within values characteristic of the phase
modifying the nervous system. nuclear chemistry is the study of how subatomic particles come together and make nuclei. modern transmutation is a large component of nuclear chemistry, and the table of nuclides is an important result and tool for this field. in addition to medical applications, nuclear chemistry encompasses nuclear engineering which explores the topic of using nuclear power sources for generating energy. organic chemistry is the study of the structure, properties, composition, mechanisms, and reactions of organic compounds. an organic compound is defined as any compound based on a carbon skeleton. organic compounds can be classified, organized and understood in reactions by their functional groups, unit atoms or molecules that show characteristic chemical properties in a compound. physical chemistry is the study of the physical and fundamental basis of chemical systems and processes. in particular, the energetics and dynamics of such systems and processes are of interest to physical chemists. important areas of study include chemical thermodynamics, chemical kinetics, electrochemistry, statistical mechanics, spectroscopy, and more recently, astrochemistry. physical chemistry has large overlap with molecular physics. physical chemistry involves the use of infinitesimal calculus in deriving equations. it is usually associated with quantum chemistry and theoretical chemistry. physical chemistry is a distinct discipline from chemical physics, but again, there is very strong overlap. theoretical chemistry is the study of chemistry via fundamental theoretical reasoning ( usually within mathematics or physics ). in particular the application of quantum mechanics to chemistry is called quantum chemistry. since the end of the second world war, the development of computers has allowed a systematic development of computational chemistry, which is the art of developing and applying computer programs for solving chemical problems. theoretical chemistry has large overlap with ( theoretical and experimental ) condensed matter physics and molecular physics. other subdivisions include electrochemistry, femtochemistry, flavor chemistry, flow chemistry, immunohistochemistry, hydrogenation chemistry, mathematical chemistry, molecular mechanics, natural product chemistry, organometallic chemistry, petrochemistry, photochemistry, physical organic chemistry, polymer chemistry, radiochemistry, sonochemistry, supramolecular chemistry, synthetic chemistry, and many others. = = = interdisciplinary = = = interdisciplinary fields include agrochemistry, astrochemistry ( and cosmochemistry ), atmospheric chemistry, chemical engineering, chemical biology, chemo - informatics, environmental chemistry, geochemistry, green chemistry, immunochemistry, marine chemistry, materials science, mechanochemistry, medicinal chemistry, molecular biology,
= = organic compounds are molecules that contain carbon bonded to another element such as hydrogen. with the exception of water, nearly all the molecules that make up each organism contain carbon. carbon can form covalent bonds with up to four other atoms, enabling it to form diverse, large, and complex molecules. for example, a single carbon atom can form four single covalent bonds such as in methane, two double covalent bonds such as in carbon dioxide ( co2 ), or a triple covalent bond such as in carbon monoxide ( co ). moreover, carbon can form very long chains of interconnecting carbon β carbon bonds such as octane or ring - like structures such as glucose. the simplest form of an organic molecule is the hydrocarbon, which is a large family of organic compounds that are composed of hydrogen atoms bonded to a chain of carbon atoms. a hydrocarbon backbone can be substituted by other elements such as oxygen ( o ), hydrogen ( h ), phosphorus ( p ), and sulfur ( s ), which can change the chemical behavior of that compound. groups of atoms that contain these elements ( o -, h -, p -, and s - ) and are bonded to a central carbon atom or skeleton are called functional groups. there are six prominent functional groups that can be found in organisms : amino group, carboxyl group, carbonyl group, hydroxyl group, phosphate group, and sulfhydryl group. in 1953, the miller β urey experiment showed that organic compounds could be synthesized abiotically within a closed system mimicking the conditions of early earth, thus suggesting that complex organic molecules could have arisen spontaneously in early earth ( see abiogenesis ). = = = macromolecules = = = macromolecules are large molecules made up of smaller subunits or monomers. monomers include sugars, amino acids, and nucleotides. carbohydrates include monomers and polymers of sugars. lipids are the only class of macromolecules that are not made up of polymers. they include steroids, phospholipids, and fats, largely nonpolar and hydrophobic ( water - repelling ) substances. proteins are the most diverse of the macromolecules. they include enzymes, transport proteins, large signaling molecules, antibodies, and structural proteins. the basic unit ( or monomer ) of a protein is an amino acid. twenty amino acids are used in proteins. nucleic acids
the chemistry of condensed phases ( solids, liquids, polymers ) and interfaces between different phases. neurochemistry is the study of neurochemicals ; including transmitters, peptides, proteins, lipids, sugars, and nucleic acids ; their interactions, and the roles they play in forming, maintaining, and modifying the nervous system. nuclear chemistry is the study of how subatomic particles come together and make nuclei. modern transmutation is a large component of nuclear chemistry, and the table of nuclides is an important result and tool for this field. in addition to medical applications, nuclear chemistry encompasses nuclear engineering which explores the topic of using nuclear power sources for generating energy. organic chemistry is the study of the structure, properties, composition, mechanisms, and reactions of organic compounds. an organic compound is defined as any compound based on a carbon skeleton. organic compounds can be classified, organized and understood in reactions by their functional groups, unit atoms or molecules that show characteristic chemical properties in a compound. physical chemistry is the study of the physical and fundamental basis of chemical systems and processes. in particular, the energetics and dynamics of such systems and processes are of interest to physical chemists. important areas of study include chemical thermodynamics, chemical kinetics, electrochemistry, statistical mechanics, spectroscopy, and more recently, astrochemistry. physical chemistry has large overlap with molecular physics. physical chemistry involves the use of infinitesimal calculus in deriving equations. it is usually associated with quantum chemistry and theoretical chemistry. physical chemistry is a distinct discipline from chemical physics, but again, there is very strong overlap. theoretical chemistry is the study of chemistry via fundamental theoretical reasoning ( usually within mathematics or physics ). in particular the application of quantum mechanics to chemistry is called quantum chemistry. since the end of the second world war, the development of computers has allowed a systematic development of computational chemistry, which is the art of developing and applying computer programs for solving chemical problems. theoretical chemistry has large overlap with ( theoretical and experimental ) condensed matter physics and molecular physics. other subdivisions include electrochemistry, femtochemistry, flavor chemistry, flow chemistry, immunohistochemistry, hydrogenation chemistry, mathematical chemistry, molecular mechanics, natural product chemistry, organometallic chemistry, petrochemistry, photochemistry, physical organic chemistry, polymer chemistry, radiochemistry, sonochemistry, supramolecular chemistry, synthetic chemistry, and many others. = = = interdisciplinary = = = interdisciplinary fields include ag
are studied in chemistry are usually the result of interactions between atoms, leading to rearrangements of the chemical bonds which hold atoms together. such behaviors are studied in a chemistry laboratory. the chemistry laboratory stereotypically uses various forms of laboratory glassware. however glassware is not central to chemistry, and a great deal of experimental ( as well as applied / industrial ) chemistry is done without it. a chemical reaction is a transformation of some substances into one or more different substances. the basis of such a chemical transformation is the rearrangement of electrons in the chemical bonds between atoms. it can be symbolically depicted through a chemical equation, which usually involves atoms as subjects. the number of atoms on the left and the right in the equation for a chemical transformation is equal. ( when the number of atoms on either side is unequal, the transformation is referred to as a nuclear reaction or radioactive decay. ) the type of chemical reactions a substance may undergo and the energy changes that may accompany it are constrained by certain basic rules, known as chemical laws. energy and entropy considerations are invariably important in almost all chemical studies. chemical substances are classified in terms of their structure, phase, as well as their chemical compositions. they can be analyzed using the tools of chemical analysis, e. g. spectroscopy and chromatography. scientists engaged in chemical research are known as chemists. most chemists specialize in one or more sub - disciplines. several concepts are essential for the study of chemistry ; some of them are : = = = matter = = = in chemistry, matter is defined as anything that has rest mass and volume ( it takes up space ) and is made up of particles. the particles that make up matter have rest mass as well β not all particles have rest mass, such as the photon. matter can be a pure chemical substance or a mixture of substances. = = = = atom = = = = the atom is the basic unit of chemistry. it consists of a dense core called the atomic nucleus surrounded by a space occupied by an electron cloud. the nucleus is made up of positively charged protons and uncharged neutrons ( together called nucleons ), while the electron cloud consists of negatively charged electrons which orbit the nucleus. in a neutral atom, the negatively charged electrons balance out the positive charge of the protons. the nucleus is dense ; the mass of a nucleon is approximately 1, 836 times that of an electron, yet the radius of an atom is about 10, 000 times that of its nucleus. the atom
the topic of using nuclear power sources for generating energy. organic chemistry is the study of the structure, properties, composition, mechanisms, and reactions of organic compounds. an organic compound is defined as any compound based on a carbon skeleton. organic compounds can be classified, organized and understood in reactions by their functional groups, unit atoms or molecules that show characteristic chemical properties in a compound. physical chemistry is the study of the physical and fundamental basis of chemical systems and processes. in particular, the energetics and dynamics of such systems and processes are of interest to physical chemists. important areas of study include chemical thermodynamics, chemical kinetics, electrochemistry, statistical mechanics, spectroscopy, and more recently, astrochemistry. physical chemistry has large overlap with molecular physics. physical chemistry involves the use of infinitesimal calculus in deriving equations. it is usually associated with quantum chemistry and theoretical chemistry. physical chemistry is a distinct discipline from chemical physics, but again, there is very strong overlap. theoretical chemistry is the study of chemistry via fundamental theoretical reasoning ( usually within mathematics or physics ). in particular the application of quantum mechanics to chemistry is called quantum chemistry. since the end of the second world war, the development of computers has allowed a systematic development of computational chemistry, which is the art of developing and applying computer programs for solving chemical problems. theoretical chemistry has large overlap with ( theoretical and experimental ) condensed matter physics and molecular physics. other subdivisions include electrochemistry, femtochemistry, flavor chemistry, flow chemistry, immunohistochemistry, hydrogenation chemistry, mathematical chemistry, molecular mechanics, natural product chemistry, organometallic chemistry, petrochemistry, photochemistry, physical organic chemistry, polymer chemistry, radiochemistry, sonochemistry, supramolecular chemistry, synthetic chemistry, and many others. = = = interdisciplinary = = = interdisciplinary fields include agrochemistry, astrochemistry ( and cosmochemistry ), atmospheric chemistry, chemical engineering, chemical biology, chemo - informatics, environmental chemistry, geochemistry, green chemistry, immunochemistry, marine chemistry, materials science, mechanochemistry, medicinal chemistry, molecular biology, nanotechnology, oenology, pharmacology, phytochemistry, solid - state chemistry, surface science, thermochemistry, and many others. = = = industry = = = the chemical industry represents an important economic activity worldwide. the global top 50 chemical producers in 2013 had sales
analyzing their radiation spectra. the term chemical energy is often used to indicate the potential of a chemical substance to undergo a transformation through a chemical reaction or to transform other chemical substances. = = = reaction = = = when a chemical substance is transformed as a result of its interaction with another substance or with energy, a chemical reaction is said to have occurred. a chemical reaction is therefore a concept related to the " reaction " of a substance when it comes in close contact with another, whether as a mixture or a solution ; exposure to some form of energy, or both. it results in some energy exchange between the constituents of the reaction as well as with the system environment, which may be designed vessels β often laboratory glassware. chemical reactions can result in the formation or dissociation of molecules, that is, molecules breaking apart to form two or more molecules or rearrangement of atoms within or across molecules. chemical reactions usually involve the making or breaking of chemical bonds. oxidation, reduction, dissociation, acid β base neutralization and molecular rearrangement are some examples of common chemical reactions. a chemical reaction can be symbolically depicted through a chemical equation. while in a non - nuclear chemical reaction the number and kind of atoms on both sides of the equation are equal, for a nuclear reaction this holds true only for the nuclear particles viz. protons and neutrons. the sequence of steps in which the reorganization of chemical bonds may be taking place in the course of a chemical reaction is called its mechanism. a chemical reaction can be envisioned to take place in a number of steps, each of which may have a different speed. many reaction intermediates with variable stability can thus be envisaged during the course of a reaction. reaction mechanisms are proposed to explain the kinetics and the relative product mix of a reaction. many physical chemists specialize in exploring and proposing the mechanisms of various chemical reactions. several empirical rules, like the woodward β hoffmann rules often come in handy while proposing a mechanism for a chemical reaction. according to the iupac gold book, a chemical reaction is " a process that results in the interconversion of chemical species. " accordingly, a chemical reaction may be an elementary reaction or a stepwise reaction. an additional caveat is made, in that this definition includes cases where the interconversion of conformers is experimentally observable. such detectable chemical reactions normally involve sets of molecular entities as indicated by this definition, but it is often conceptually convenient to use the term also for changes involving single molecular entities (
Question: Which is a chemical compound?
A) aluminum
B) carbonated water
C) sodium chloride
D) carbon-14
|
C) sodium chloride
|
Context:
and their competitive or mutualistic interactions with other species. some ecologists even rely on empirical data from indigenous people that is gathered by ethnobotanists. this information can relay a great deal of information on how the land once was thousands of years ago and how it has changed over that time. the goals of plant ecology are to understand the causes of their distribution patterns, productivity, environmental impact, evolution, and responses to environmental change. plants depend on certain edaphic ( soil ) and climatic factors in their environment but can modify these factors too. for example, they can change their environment ' s albedo, increase runoff interception, stabilise mineral soils and develop their organic content, and affect local temperature. plants compete with other organisms in their ecosystem for resources. they interact with their neighbours at a variety of spatial scales in groups, populations and communities that collectively constitute vegetation. regions with characteristic vegetation types and dominant plants as well as similar abiotic and biotic factors, climate, and geography make up biomes like tundra or tropical rainforest. herbivores eat plants, but plants can defend themselves and some species are parasitic or even carnivorous. other organisms form mutually beneficial relationships with plants. for example, mycorrhizal fungi and rhizobia provide plants with nutrients in exchange for food, ants are recruited by ant plants to provide protection, honey bees, bats and other animals pollinate flowers and humans and other animals act as dispersal vectors to spread spores and seeds. = = = plants, climate and environmental change = = = plant responses to climate and other environmental changes can inform our understanding of how these changes affect ecosystem function and productivity. for example, plant phenology can be a useful proxy for temperature in historical climatology, and the biological impact of climate change and global warming. palynology, the analysis of fossil pollen deposits in sediments from thousands or millions of years ago allows the reconstruction of past climates. estimates of atmospheric co2 concentrations since the palaeozoic have been obtained from stomatal densities and the leaf shapes and sizes of ancient land plants. ozone depletion can expose plants to higher levels of ultraviolet radiation - b ( uv - b ), resulting in lower growth rates. moreover, information from studies of community ecology, plant systematics, and taxonomy is essential to understanding vegetation change, habitat destruction and species extinction. = = genetics = = inheritance in plants follows the same fundamental principles of genetics as in other multicellular organisms. gregor mendel discovered the genetic laws of inheritance by studying
soil erosion. plants are crucial to the future of human society as they provide food, oxygen, biochemicals, and products for people, as well as creating and preserving soil. historically, all living things were classified as either animals or plants and botany covered the study of all organisms not considered animals. botanists examine both the internal functions and processes within plant organelles, cells, tissues, whole plants, plant populations and plant communities. at each of these levels, a botanist may be concerned with the classification ( taxonomy ), phylogeny and evolution, structure ( anatomy and morphology ), or function ( physiology ) of plant life. the strictest definition of " plant " includes only the " land plants " or embryophytes, which include seed plants ( gymnosperms, including the pines, and flowering plants ) and the free - sporing cryptogams including ferns, clubmosses, liverworts, hornworts and mosses. embryophytes are multicellular eukaryotes descended from an ancestor that obtained its energy from sunlight by photosynthesis. they have life cycles with alternating haploid and diploid phases. the sexual haploid phase of embryophytes, known as the gametophyte, nurtures the developing diploid embryo sporophyte within its tissues for at least part of its life, even in the seed plants, where the gametophyte itself is nurtured by its parent sporophyte. other groups of organisms that were previously studied by botanists include bacteria ( now studied in bacteriology ), fungi ( mycology ) β including lichen - forming fungi ( lichenology ), non - chlorophyte algae ( phycology ), and viruses ( virology ). however, attention is still given to these groups by botanists, and fungi ( including lichens ) and photosynthetic protists are usually covered in introductory botany courses. palaeobotanists study ancient plants in the fossil record to provide information about the evolutionary history of plants. cyanobacteria, the first oxygen - releasing photosynthetic organisms on earth, are thought to have given rise to the ancestor of plants by entering into an endosymbiotic relationship with an early eukaryote, ultimately becoming the chloroplasts in plant cells. the new photosynthetic plants ( along with their algal relatives ) accelerated the rise in atmospheric oxygen started by the cyanobacteria, changing the
the structural components of cells. as a by - product of photosynthesis, plants release oxygen into the atmosphere, a gas that is required by nearly all living things to carry out cellular respiration. in addition, they are influential in the global carbon and water cycles and plant roots bind and stabilise soils, preventing soil erosion. plants are crucial to the future of human society as they provide food, oxygen, biochemicals, and products for people, as well as creating and preserving soil. historically, all living things were classified as either animals or plants and botany covered the study of all organisms not considered animals. botanists examine both the internal functions and processes within plant organelles, cells, tissues, whole plants, plant populations and plant communities. at each of these levels, a botanist may be concerned with the classification ( taxonomy ), phylogeny and evolution, structure ( anatomy and morphology ), or function ( physiology ) of plant life. the strictest definition of " plant " includes only the " land plants " or embryophytes, which include seed plants ( gymnosperms, including the pines, and flowering plants ) and the free - sporing cryptogams including ferns, clubmosses, liverworts, hornworts and mosses. embryophytes are multicellular eukaryotes descended from an ancestor that obtained its energy from sunlight by photosynthesis. they have life cycles with alternating haploid and diploid phases. the sexual haploid phase of embryophytes, known as the gametophyte, nurtures the developing diploid embryo sporophyte within its tissues for at least part of its life, even in the seed plants, where the gametophyte itself is nurtured by its parent sporophyte. other groups of organisms that were previously studied by botanists include bacteria ( now studied in bacteriology ), fungi ( mycology ) β including lichen - forming fungi ( lichenology ), non - chlorophyte algae ( phycology ), and viruses ( virology ). however, attention is still given to these groups by botanists, and fungi ( including lichens ) and photosynthetic protists are usually covered in introductory botany courses. palaeobotanists study ancient plants in the fossil record to provide information about the evolutionary history of plants. cyanobacteria, the first oxygen - releasing photosynthetic organisms on earth, are thought to have given rise to the
if wood has been with us since time immemorial, being part of our environment, housing and tools, now wood has gain momentum, as it is clear that wood improves our life style. because of the healthiness, resistance, ecology and comfort, wood is important for all of us, no matter what our life style is. woodtouch project aims to open a completely new market for furniture and interior design companies, enabling touch interaction between the user and wooden furniture surfaces. why not switch on or dim the lights touching a wooden table? why not turn on the heating system? why not use wood as a touch sensitive surface for domotic control? the furniture designed with this novel technology, offers a wooden outer image and has different touch sensitive areas over the ones the user is able to control all sorts of electric appliances touching over a wooden surface.
eat them. plants and other photosynthetic organisms are at the base of most food chains because they use the energy from the sun and nutrients from the soil and atmosphere, converting them into a form that can be used by animals. this is what ecologists call the first trophic level. the modern forms of the major staple foods, such as hemp, teff, maize, rice, wheat and other cereal grasses, pulses, bananas and plantains, as well as hemp, flax and cotton grown for their fibres, are the outcome of prehistoric selection over thousands of years from among wild ancestral plants with the most desirable characteristics. botanists study how plants produce food and how to increase yields, for example through plant breeding, making their work important to humanity ' s ability to feed the world and provide food security for future generations. botanists also study weeds, which are a considerable problem in agriculture, and the biology and control of plant pathogens in agriculture and natural ecosystems. ethnobotany is the study of the relationships between plants and people. when applied to the investigation of historical plant β people relationships ethnobotany may be referred to as archaeobotany or palaeoethnobotany. some of the earliest plant - people relationships arose between the indigenous people of canada in identifying edible plants from inedible plants. this relationship the indigenous people had with plants was recorded by ethnobotanists. = = plant biochemistry = = plant biochemistry is the study of the chemical processes used by plants. some of these processes are used in their primary metabolism like the photosynthetic calvin cycle and crassulacean acid metabolism. others make specialised materials like the cellulose and lignin used to build their bodies, and secondary products like resins and aroma compounds. plants and various other groups of photosynthetic eukaryotes collectively known as " algae " have unique organelles known as chloroplasts. chloroplasts are thought to be descended from cyanobacteria that formed endosymbiotic relationships with ancient plant and algal ancestors. chloroplasts and cyanobacteria contain the blue - green pigment chlorophyll a. chlorophyll a ( as well as its plant and green algal - specific cousin chlorophyll b ) absorbs light in the blue - violet and orange / red parts of the spectrum while reflecting and transmitting the green light that we see as the characteristic colour
equalizing the flow at these places, produces a lowering of the river above the rapids by facilitating the efflux, which may result in the appearance of fresh shoals at the low stage of the river. where, however, narrow rocky reefs or other hard shoals stretch across the bottom of a river and present obstacles to the erosion by the current of the soft materials forming the bed of the river above and below, their removal may result in permanent improvement by enabling the river to deepen its bed by natural scour. the capability of a river to provide a waterway for navigation during the summer or throughout the dry season depends on the depth that can be secured in the channel at the lowest stage. the problem in the dry season is the small discharge and deficiency in scour during this period. a typical solution is to restrict the width of the low - water channel, concentrate all of the flow in it, and also to fix its position so that it is scoured out every year by the floods which follow the deepest part of the bed along the line of the strongest current. this can be effected by closing subsidiary low - water channels with dikes across them, and narrowing the channel at the low stage by low - dipping cross dikes extending from the river banks down the slope and pointing slightly up - stream so as to direct the water flowing over them into a central channel. = = estuarine works = = the needs of navigation may also require that a stable, continuous, navigable channel is prolonged from the navigable river to deep water at the mouth of the estuary. the interaction of river flow and tide needs to be modeled by computer or using scale models, moulded to the configuration of the estuary under consideration and reproducing in miniature the tidal ebb and flow and fresh - water discharge over a bed of fine sand, in which various lines of training walls can be successively inserted. the models should be capable of furnishing valuable indications of the respective effects and comparative merits of the different schemes proposed for works. = = see also = = bridge scour flood control = = references = = = = external links = = u. s. army corps of engineers β civil works program river morphology and stream restoration references - wildland hydrology at the library of congress web archives ( archived 2002 - 08 - 13 )
= = = = = = environmental remediation = = = environmental remediation is the process through which contaminants or pollutants in soil, water and other media are removed to improve environmental quality. the main focus is the reduction of hazardous substances within the environment. some of the areas involved in environmental remediation include ; soil contamination, hazardous waste, groundwater contamination, oil, gas and chemical spills. there are three most common types of environmental remediation. these include soil, water, and sediment remediation. soil remediation consists of removing contaminants in soil, as these pose great risks to humans and the ecosystem. some examples of this are heavy metals, pesticides, and radioactive materials. depending on the contaminant the remedial processes can be physical, chemical, thermal, or biological. water remediation is one of the most important considering water is an essential natural resource. depending on the source of water there will be different contaminants. surface water contamination mainly consists of agricultural, animal, and industrial waste, as well as acid mine drainage. there has been a rise in the need for water remediation due to the increased discharge of industrial waste, leading to a demand for sustainable water solutions. the market for water remediation is expected to consistently increase to $ 19. 6 billion by 2030. sediment remediation consists of removing contaminated sediments. is it almost similar to soil remediation except it is often more sophisticated as it involves additional contaminants. to reduce the contaminants it is likely to use physical, chemical, and biological processes that help with source control, but if these processes are executed correctly, there ' s a risk of contamination resurfacing. = = = solid waste management = = = solid waste management is the purification, consumption, reuse, disposal, and treatment of solid waste that is undertaken by the government or the ruling bodies of a city / town. it refers to the collection, treatment, and disposal of non - soluble, solid waste material. solid waste is associated with both industrial, institutional, commercial and residential activities. hazardous solid waste, when improperly disposed can encourage the infestation of insects and rodents, contributing to the spread of diseases. some of the most common types of solid waste management include ; landfills, vermicomposting, composting, recycling, and incineration. however, a major barrier for solid waste management practices is the high costs associated with recycling
masculinity and warmth. the five phases β fire, earth, metal, wood, and water β described a cycle of transformations in nature. the water turned into wood, which turned into the fire when it burned. the ashes left by fire were earth. using these principles, chinese philosophers and doctors explored human anatomy, characterizing organs as predominantly yin or yang, and understood the relationship between the pulse, the heart, and the flow of blood in the body centuries before it became accepted in the west. little evidence survives of how ancient indian cultures around the indus river understood nature, but some of their perspectives may be reflected in the vedas, a set of sacred hindu texts. they reveal a conception of the universe as ever - expanding and constantly being recycled and reformed. surgeons in the ayurvedic tradition saw health and illness as a combination of three humors : wind, bile and phlegm. a healthy life resulted from a balance among these humors. in ayurvedic thought, the body consisted of five elements : earth, water, fire, wind, and space. ayurvedic surgeons performed complex surgeries and developed a detailed understanding of human anatomy. pre - socratic philosophers in ancient greek culture brought natural philosophy a step closer to direct inquiry about cause and effect in nature between 600 and 400 bc. however, an element of magic and mythology remained. natural phenomena such as earthquakes and eclipses were explained increasingly in the context of nature itself instead of being attributed to angry gods. thales of miletus, an early philosopher who lived from 625 to 546 bc, explained earthquakes by theorizing that the world floated on water and that water was the fundamental element in nature. in the 5th century bc, leucippus was an early exponent of atomism, the idea that the world is made up of fundamental indivisible particles. pythagoras applied greek innovations in mathematics to astronomy and suggested that the earth was spherical. = = = aristotelian natural philosophy ( 400 bc β 1100 ad ) = = = later socratic and platonic thought focused on ethics, morals, and art and did not attempt an investigation of the physical world ; plato criticized pre - socratic thinkers as materialists and anti - religionists. aristotle, however, a student of plato who lived from 384 to 322 bc, paid closer attention to the natural world in his philosophy. in his history of animals, he described the inner workings of 110 species, including the stingray, catfish and
horticultural botany, phytopathology, and phytopharmacology. = = scope and importance = = the study of plants is vital because they underpin almost all animal life on earth by generating a large proportion of the oxygen and food that provide humans and other organisms with aerobic respiration with the chemical energy they need to exist. plants, algae and cyanobacteria are the major groups of organisms that carry out photosynthesis, a process that uses the energy of sunlight to convert water and carbon dioxide into sugars that can be used both as a source of chemical energy and of organic molecules that are used in the structural components of cells. as a by - product of photosynthesis, plants release oxygen into the atmosphere, a gas that is required by nearly all living things to carry out cellular respiration. in addition, they are influential in the global carbon and water cycles and plant roots bind and stabilise soils, preventing soil erosion. plants are crucial to the future of human society as they provide food, oxygen, biochemicals, and products for people, as well as creating and preserving soil. historically, all living things were classified as either animals or plants and botany covered the study of all organisms not considered animals. botanists examine both the internal functions and processes within plant organelles, cells, tissues, whole plants, plant populations and plant communities. at each of these levels, a botanist may be concerned with the classification ( taxonomy ), phylogeny and evolution, structure ( anatomy and morphology ), or function ( physiology ) of plant life. the strictest definition of " plant " includes only the " land plants " or embryophytes, which include seed plants ( gymnosperms, including the pines, and flowering plants ) and the free - sporing cryptogams including ferns, clubmosses, liverworts, hornworts and mosses. embryophytes are multicellular eukaryotes descended from an ancestor that obtained its energy from sunlight by photosynthesis. they have life cycles with alternating haploid and diploid phases. the sexual haploid phase of embryophytes, known as the gametophyte, nurtures the developing diploid embryo sporophyte within its tissues for at least part of its life, even in the seed plants, where the gametophyte itself is nurtured by its parent sporophyte. other groups of organisms that were previously studied by botanists include bacteria ( now studied in bacteriology )
zero, called the positive numbers. another property required for a ring to be ordered is that, for each positive number, there exists a unique corresponding number less than 0 whose sum with the original positive number is 0. these numbers less than 0 are called the negative numbers. the numbers in each such pair are their respective additive inverses. this attribute of a number, being exclusively either zero ( 0 ), positive ( + ), or negative ( β ), is called its sign, and is often encoded to the real numbers 0, 1, and β1, respectively ( similar to the way the sign function is defined ). since rational and real numbers are also ordered rings ( in fact ordered fields ), the sign attribute also applies to these number systems. when a minus sign is used in between two numbers, it represents the binary operation of subtraction. when a minus sign is written before a single number, it represents the unary operation of yielding the additive inverse ( sometimes called negation ) of the operand. abstractly then, the difference of two number is the sum of the minuend with the additive inverse of the subtrahend. while 0 is its own additive inverse ( β0 = 0 ), the additive inverse of a positive number is negative, and the additive inverse of a negative number is positive. a double application of this operation is written as β ( β3 ) = 3. the plus sign is predominantly used in algebra to denote the binary operation of addition, and only rarely to emphasize the positivity of an expression. in common numeral notation ( used in arithmetic and elsewhere ), the sign of a number is often made explicit by placing a plus or a minus sign before the number. for example, + 3 denotes " positive three ", and β3 denotes " negative three " ( algebraically : the additive inverse of 3 ). without specific context ( or when no explicit sign is given ), a number is interpreted per default as positive. this notation establishes a strong association of the minus sign " β " with negative numbers, and the plus sign " + " with positive numbers. = = = sign of zero = = = within the convention of zero being neither positive nor negative, a specific sign - value 0 may be assigned to the number value 0. this is exploited in the sgn { \ displaystyle \ operatorname { sgn } } - function, as defined for real numbers. in arithmetic, + 0 and β0 both denote the same number 0. there is generally no danger of
Question: Which of these human activities in a forest has a positive effect on the ecosystem?
A) Clearing the land to build new homes
B) Planting new trees where old ones were cut down
C) Building a new highway through the area
D) Cutting down trees to make farmland
|
B) Planting new trees where old ones were cut down
|
Context:
becomes quite gentle. accordingly, in large basins, rivers in most cases begin as torrents with a variable flow, and end as gently flowing rivers with a comparatively regular discharge. the irregular flow of rivers throughout their course forms one of the main difficulties in devising works for mitigating inundations or for increasing the navigable capabilities of rivers. in tropical countries subject to periodical rains, the rivers are in flood during the rainy season and have hardly any flow during the rest of the year, while in temperate regions, where the rainfall is more evenly distributed throughout the year, evaporation causes the available rainfall to be much less in hot summer weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under
navigable capabilities of rivers. in tropical countries subject to periodical rains, the rivers are in flood during the rainy season and have hardly any flow during the rest of the year, while in temperate regions, where the rainfall is more evenly distributed throughout the year, evaporation causes the available rainfall to be much less in hot summer weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under ordinary conditions, most of the materials brought down from the high lands by torrential water courses are carried forward by the main river to the sea, or partially strewn over flat alluvial plains during floods ; the size of the materials forming the bed of the river or borne along by the stream is gradually reduced on proceeding sea
gravity induced condensation takes the form of momentum alignment in an ensemble of identical particles. use is made of a one - dimensional ising model to calculate the alignment per particle and the correlation length as a function of the temperature. these parameters indicate that momentum alignment is possible in the proximity of some astrophysical objects and in earth, or near earth laboratories. momenta oscillations behave as known spin oscillations and obey identical dispersion relations.
superheated droplets are proven to be excelent detectors for neutrons and could be used as a neutron dosimeter. to detect accurately the volume of the vapour formed upon nucleation and hence to observe the nucleation quantitatively an air displacement system has been developed.
weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under ordinary conditions, most of the materials brought down from the high lands by torrential water courses are carried forward by the main river to the sea, or partially strewn over flat alluvial plains during floods ; the size of the materials forming the bed of the river or borne along by the stream is gradually reduced on proceeding seawards, so that in the po river in italy, for instance, pebbles and gravel are found for about 140 miles below turin, sand along the next 100 miles, and silt and mud in the last 110 miles ( 176 km ). = = channelization = = the removal of obstructions, natural or artificial
approximately corresponds to the slope of the country it traverses ; as rivers rise close to the highest part of their basins, generally in hilly regions, their fall is rapid near their source and gradually diminishes, with occasional irregularities, until, in traversing plains along the latter part of their course, their fall usually becomes quite gentle. accordingly, in large basins, rivers in most cases begin as torrents with a variable flow, and end as gently flowing rivers with a comparatively regular discharge. the irregular flow of rivers throughout their course forms one of the main difficulties in devising works for mitigating inundations or for increasing the navigable capabilities of rivers. in tropical countries subject to periodical rains, the rivers are in flood during the rainy season and have hardly any flow during the rest of the year, while in temperate regions, where the rainfall is more evenly distributed throughout the year, evaporation causes the available rainfall to be much less in hot summer weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with
of substances dissolved in aqueous solution ( that is, in water ). less familiar phases include plasmas, bose β einstein condensates and fermionic condensates and the paramagnetic and ferromagnetic phases of magnetic materials. while most familiar phases deal with three - dimensional systems, it is also possible to define analogs in two - dimensional systems, which has received attention for its relevance to systems in biology. = = = bonding = = = atoms sticking together in molecules or crystals are said to be bonded with one another. a chemical bond may be visualized as the multipole balance between the positive charges in the nuclei and the negative charges oscillating about them. more than simple attraction and repulsion, the energies and distributions characterize the availability of an electron to bond to another atom. the chemical bond can be a covalent bond, an ionic bond, a hydrogen bond or just because of van der waals force. each of these kinds of bonds is ascribed to some potential. these potentials create the interactions which hold atoms together in molecules or crystals. in many simple compounds, valence bond theory, the valence shell electron pair repulsion model ( vsepr ), and the concept of oxidation number can be used to explain molecular structure and composition. an ionic bond is formed when a metal loses one or more of its electrons, becoming a positively charged cation, and the electrons are then gained by the non - metal atom, becoming a negatively charged anion. the two oppositely charged ions attract one another, and the ionic bond is the electrostatic force of attraction between them. for example, sodium ( na ), a metal, loses one electron to become an na + cation while chlorine ( cl ), a non - metal, gains this electron to become clβ. the ions are held together due to electrostatic attraction, and that compound sodium chloride ( nacl ), or common table salt, is formed. in a covalent bond, one or more pairs of valence electrons are shared by two atoms : the resulting electrically neutral group of bonded atoms is termed a molecule. atoms will share valence electrons in such a way as to create a noble gas electron configuration ( eight electrons in their outermost shell ) for each atom. atoms that tend to combine in such a way that they each have eight electrons in their valence shell are said to follow the octet rule. however, some elements like hydrogen and lithium need only two electrons in their outermost shell to
using only lidar or radar an accurate cloud boundary height estimate is often not possible. the combination of lidar and radar can give a reliable cloud boundary estimate in a much broader range of cases. however, also this combination with standard methods still can not measure the cloud boundaries in all cases. this will be illustrated with data from the clouds and radiation measurement campaigns, clara. rain is a problem : the radar has problems to measure the small cloud droplets in the presence of raindrops. similarly, few large particles below cloud base can obscure the cloud base in radar measurements. and the radar reflectivity can be very low at the cloud base of water clouds or in large regions of ice clouds, due to small particles. multiple cloud layers and clouds with specular reflections can pose problems for lidar. more advanced measurement techniques are suggested to solve these problems. an angle scanning lidar can, for example, detect specular reflections, while using information from the radars doppler velocity spectrum may help to detect clouds during rain.
the recent report on laser cooling of liquid may contradict the law of energy conservation.
hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under ordinary conditions, most of the materials brought down from the high lands by torrential water courses are carried forward by the main river to the sea, or partially strewn over flat alluvial plains during floods ; the size of the materials forming the bed of the river or borne along by the stream is gradually reduced on proceeding seawards, so that in the po river in italy, for instance, pebbles and gravel are found for about 140 miles below turin, sand along the next 100 miles, and silt and mud in the last 110 miles ( 176 km ). = = channelization = = the removal of obstructions, natural or artificial ( e. g., trunks of trees, boulders and accumulations of gravel ) from a river bed furnishes a simple and efficient means of increasing the discharging capacity of its channel. such removals will consequently lower the height of floods upstream. every impediment to the flow, in proportion to
Question: Which is an example of water condensing?
A) A puddle disappearing on a hot summer afternoon
B) Sweat forming on your forehead after you do a lot of exercise
C) Ice cubes melting when you put them out in the sun
D) Dew forming on plants during a cold night
|
D) Dew forming on plants during a cold night
|
Context:
inherited traits such as shape in pisum sativum ( peas ). what mendel learned from studying plants has had far - reaching benefits outside of botany. similarly, " jumping genes " were discovered by barbara mcclintock while she was studying maize. nevertheless, there are some distinctive genetic differences between plants and other organisms. species boundaries in plants may be weaker than in animals, and cross species hybrids are often possible. a familiar example is peppermint, mentha Γ piperita, a sterile hybrid between mentha aquatica and spearmint, mentha spicata. the many cultivated varieties of wheat are the result of multiple inter - and intra - specific crosses between wild species and their hybrids. angiosperms with monoecious flowers often have self - incompatibility mechanisms that operate between the pollen and stigma so that the pollen either fails to reach the stigma or fails to germinate and produce male gametes. this is one of several methods used by plants to promote outcrossing. in many land plants the male and female gametes are produced by separate individuals. these species are said to be dioecious when referring to vascular plant sporophytes and dioicous when referring to bryophyte gametophytes. charles darwin in his 1878 book the effects of cross and self - fertilization in the vegetable kingdom at the start of chapter xii noted " the first and most important of the conclusions which may be drawn from the observations given in this volume, is that generally cross - fertilisation is beneficial and self - fertilisation often injurious, at least with the plants on which i experimented. " an important adaptive benefit of outcrossing is that it allows the masking of deleterious mutations in the genome of progeny. this beneficial effect is also known as hybrid vigor or heterosis. once outcrossing is established, subsequent switching to inbreeding becomes disadvantageous since it allows expression of the previously masked deleterious recessive mutations, commonly referred to as inbreeding depression. unlike in higher animals, where parthenogenesis is rare, asexual reproduction may occur in plants by several different mechanisms. the formation of stem tubers in potato is one example. particularly in arctic or alpine habitats, where opportunities for fertilisation of flowers by animals are rare, plantlets or bulbs, may develop instead of flowers, replacing sexual reproduction with asexual reproduction and giving rise to clonal populations genetically identical to the parent. this is one
is the scientific study of inheritance. mendelian inheritance, specifically, is the process by which genes and traits are passed on from parents to offspring. it has several principles. the first is that genetic characteristics, alleles, are discrete and have alternate forms ( e. g., purple vs. white or tall vs. dwarf ), each inherited from one of two parents. based on the law of dominance and uniformity, which states that some alleles are dominant while others are recessive ; an organism with at least one dominant allele will display the phenotype of that dominant allele. during gamete formation, the alleles for each gene segregate, so that each gamete carries only one allele for each gene. heterozygotic individuals produce gametes with an equal frequency of two alleles. finally, the law of independent assortment, states that genes of different traits can segregate independently during the formation of gametes, i. e., genes are unlinked. an exception to this rule would include traits that are sex - linked. test crosses can be performed to experimentally determine the underlying genotype of an organism with a dominant phenotype. a punnett square can be used to predict the results of a test cross. the chromosome theory of inheritance, which states that genes are found on chromosomes, was supported by thomas morgans ' s experiments with fruit flies, which established the sex linkage between eye color and sex in these insects. = = = genes and dna = = = a gene is a unit of heredity that corresponds to a region of deoxyribonucleic acid ( dna ) that carries genetic information that controls form or function of an organism. dna is composed of two polynucleotide chains that coil around each other to form a double helix. it is found as linear chromosomes in eukaryotes, and circular chromosomes in prokaryotes. the set of chromosomes in a cell is collectively known as its genome. in eukaryotes, dna is mainly in the cell nucleus. in prokaryotes, the dna is held within the nucleoid. the genetic information is held within genes, and the complete assemblage in an organism is called its genotype. dna replication is a semiconservative process whereby each strand serves as a template for a new strand of dna. mutations are heritable changes in dna. they can arise spontaneously as a result of replication errors that were not corrected by proofreading or can
studies of the molecular genetics of model plants such as the thale cress, arabidopsis thaliana, a weedy species in the mustard family ( brassicaceae ). the genome or hereditary information contained in the genes of this species is encoded by about 135 million base pairs of dna, forming one of the smallest genomes among flowering plants. arabidopsis was the first plant to have its genome sequenced, in 2000. the sequencing of some other relatively small genomes, of rice ( oryza sativa ) and brachypodium distachyon, has made them important model species for understanding the genetics, cellular and molecular biology of cereals, grasses and monocots generally. model plants such as arabidopsis thaliana are used for studying the molecular biology of plant cells and the chloroplast. ideally, these organisms have small genomes that are well known or completely sequenced, small stature and short generation times. corn has been used to study mechanisms of photosynthesis and phloem loading of sugar in c4 plants. the single celled green alga chlamydomonas reinhardtii, while not an embryophyte itself, contains a green - pigmented chloroplast related to that of land plants, making it useful for study. a red alga cyanidioschyzon merolae has also been used to study some basic chloroplast functions. spinach, peas, soybeans and a moss physcomitrella patens are commonly used to study plant cell biology. agrobacterium tumefaciens, a soil rhizosphere bacterium, can attach to plant cells and infect them with a callus - inducing ti plasmid by horizontal gene transfer, causing a callus infection called crown gall disease. schell and van montagu ( 1977 ) hypothesised that the ti plasmid could be a natural vector for introducing the nif gene responsible for nitrogen fixation in the root nodules of legumes and other plant species. today, genetic modification of the ti plasmid is one of the main techniques for introduction of transgenes to plants and the creation of genetically modified crops. = = = epigenetics = = = epigenetics is the study of heritable changes in gene function that cannot be explained by changes in the underlying dna sequence but cause the organism ' s genes to behave ( or " express themselves " ) differently. one example
cellular and molecular biology of cereals, grasses and monocots generally. model plants such as arabidopsis thaliana are used for studying the molecular biology of plant cells and the chloroplast. ideally, these organisms have small genomes that are well known or completely sequenced, small stature and short generation times. corn has been used to study mechanisms of photosynthesis and phloem loading of sugar in c4 plants. the single celled green alga chlamydomonas reinhardtii, while not an embryophyte itself, contains a green - pigmented chloroplast related to that of land plants, making it useful for study. a red alga cyanidioschyzon merolae has also been used to study some basic chloroplast functions. spinach, peas, soybeans and a moss physcomitrella patens are commonly used to study plant cell biology. agrobacterium tumefaciens, a soil rhizosphere bacterium, can attach to plant cells and infect them with a callus - inducing ti plasmid by horizontal gene transfer, causing a callus infection called crown gall disease. schell and van montagu ( 1977 ) hypothesised that the ti plasmid could be a natural vector for introducing the nif gene responsible for nitrogen fixation in the root nodules of legumes and other plant species. today, genetic modification of the ti plasmid is one of the main techniques for introduction of transgenes to plants and the creation of genetically modified crops. = = = epigenetics = = = epigenetics is the study of heritable changes in gene function that cannot be explained by changes in the underlying dna sequence but cause the organism ' s genes to behave ( or " express themselves " ) differently. one example of epigenetic change is the marking of the genes by dna methylation which determines whether they will be expressed or not. gene expression can also be controlled by repressor proteins that attach to silencer regions of the dna and prevent that region of the dna code from being expressed. epigenetic marks may be added or removed from the dna during programmed stages of development of the plant, and are responsible, for example, for the differences between anthers, petals and normal leaves, despite the fact that they all have the same underlying genetic code. epigenetic changes may be temporary or may remain through successive cell divisions for the remainder of
. species boundaries in plants may be weaker than in animals, and cross species hybrids are often possible. a familiar example is peppermint, mentha Γ piperita, a sterile hybrid between mentha aquatica and spearmint, mentha spicata. the many cultivated varieties of wheat are the result of multiple inter - and intra - specific crosses between wild species and their hybrids. angiosperms with monoecious flowers often have self - incompatibility mechanisms that operate between the pollen and stigma so that the pollen either fails to reach the stigma or fails to germinate and produce male gametes. this is one of several methods used by plants to promote outcrossing. in many land plants the male and female gametes are produced by separate individuals. these species are said to be dioecious when referring to vascular plant sporophytes and dioicous when referring to bryophyte gametophytes. charles darwin in his 1878 book the effects of cross and self - fertilization in the vegetable kingdom at the start of chapter xii noted " the first and most important of the conclusions which may be drawn from the observations given in this volume, is that generally cross - fertilisation is beneficial and self - fertilisation often injurious, at least with the plants on which i experimented. " an important adaptive benefit of outcrossing is that it allows the masking of deleterious mutations in the genome of progeny. this beneficial effect is also known as hybrid vigor or heterosis. once outcrossing is established, subsequent switching to inbreeding becomes disadvantageous since it allows expression of the previously masked deleterious recessive mutations, commonly referred to as inbreeding depression. unlike in higher animals, where parthenogenesis is rare, asexual reproduction may occur in plants by several different mechanisms. the formation of stem tubers in potato is one example. particularly in arctic or alpine habitats, where opportunities for fertilisation of flowers by animals are rare, plantlets or bulbs, may develop instead of flowers, replacing sexual reproduction with asexual reproduction and giving rise to clonal populations genetically identical to the parent. this is one of several types of apomixis that occur in plants. apomixis can also happen in a seed, producing a seed that contains an embryo genetically identical to the parent. most sexually reproducing organisms are diploid, with paired chromosomes, but doubling of their chromosome number may occur due to errors in
generation times. corn has been used to study mechanisms of photosynthesis and phloem loading of sugar in c4 plants. the single celled green alga chlamydomonas reinhardtii, while not an embryophyte itself, contains a green - pigmented chloroplast related to that of land plants, making it useful for study. a red alga cyanidioschyzon merolae has also been used to study some basic chloroplast functions. spinach, peas, soybeans and a moss physcomitrella patens are commonly used to study plant cell biology. agrobacterium tumefaciens, a soil rhizosphere bacterium, can attach to plant cells and infect them with a callus - inducing ti plasmid by horizontal gene transfer, causing a callus infection called crown gall disease. schell and van montagu ( 1977 ) hypothesised that the ti plasmid could be a natural vector for introducing the nif gene responsible for nitrogen fixation in the root nodules of legumes and other plant species. today, genetic modification of the ti plasmid is one of the main techniques for introduction of transgenes to plants and the creation of genetically modified crops. = = = epigenetics = = = epigenetics is the study of heritable changes in gene function that cannot be explained by changes in the underlying dna sequence but cause the organism ' s genes to behave ( or " express themselves " ) differently. one example of epigenetic change is the marking of the genes by dna methylation which determines whether they will be expressed or not. gene expression can also be controlled by repressor proteins that attach to silencer regions of the dna and prevent that region of the dna code from being expressed. epigenetic marks may be added or removed from the dna during programmed stages of development of the plant, and are responsible, for example, for the differences between anthers, petals and normal leaves, despite the fact that they all have the same underlying genetic code. epigenetic changes may be temporary or may remain through successive cell divisions for the remainder of the cell ' s life. some epigenetic changes have been shown to be heritable, while others are reset in the germ cells. epigenetic changes in eukaryotic biology serve to regulate the process of cellular differentiation. during morphogenesis, totipotent stem cells become the various
phenotypic analysis. the new genetic material can be inserted randomly within the host genome or targeted to a specific location. the technique of gene targeting uses homologous recombination to make desired changes to a specific endogenous gene. this tends to occur at a relatively low frequency in plants and animals and generally requires the use of selectable markers. the frequency of gene targeting can be greatly enhanced through genome editing. genome editing uses artificially engineered nucleases that create specific double - stranded breaks at desired locations in the genome, and use the cell ' s endogenous mechanisms to repair the induced break by the natural processes of homologous recombination and nonhomologous end - joining. there are four families of engineered nucleases : meganucleases, zinc finger nucleases, transcription activator - like effector nucleases ( talens ), and the cas9 - guiderna system ( adapted from crispr ). talen and crispr are the two most commonly used and each has its own advantages. talens have greater target specificity, while crispr is easier to design and more efficient. in addition to enhancing gene targeting, engineered nucleases can be used to introduce mutations at endogenous genes that generate a gene knockout. = = applications = = genetic engineering has applications in medicine, research, industry and agriculture and can be used on a wide range of plants, animals and microorganisms. bacteria, the first organisms to be genetically modified, can have plasmid dna inserted containing new genes that code for medicines or enzymes that process food and other substrates. plants have been modified for insect protection, herbicide resistance, virus resistance, enhanced nutrition, tolerance to environmental pressures and the production of edible vaccines. most commercialised gmos are insect resistant or herbicide tolerant crop plants. genetically modified animals have been used for research, model animals and the production of agricultural or pharmaceutical products. the genetically modified animals include animals with genes knocked out, increased susceptibility to disease, hormones for extra growth and the ability to express proteins in their milk. = = = medicine = = = genetic engineering has many applications to medicine that include the manufacturing of drugs, creation of model animals that mimic human conditions and gene therapy. one of the earliest uses of genetic engineering was to mass - produce human insulin in bacteria. this application has now been applied to human growth hormones, follicle stimulating hormones ( for treating infertility ), human albumin,
smallest genomes among flowering plants. arabidopsis was the first plant to have its genome sequenced, in 2000. the sequencing of some other relatively small genomes, of rice ( oryza sativa ) and brachypodium distachyon, has made them important model species for understanding the genetics, cellular and molecular biology of cereals, grasses and monocots generally. model plants such as arabidopsis thaliana are used for studying the molecular biology of plant cells and the chloroplast. ideally, these organisms have small genomes that are well known or completely sequenced, small stature and short generation times. corn has been used to study mechanisms of photosynthesis and phloem loading of sugar in c4 plants. the single celled green alga chlamydomonas reinhardtii, while not an embryophyte itself, contains a green - pigmented chloroplast related to that of land plants, making it useful for study. a red alga cyanidioschyzon merolae has also been used to study some basic chloroplast functions. spinach, peas, soybeans and a moss physcomitrella patens are commonly used to study plant cell biology. agrobacterium tumefaciens, a soil rhizosphere bacterium, can attach to plant cells and infect them with a callus - inducing ti plasmid by horizontal gene transfer, causing a callus infection called crown gall disease. schell and van montagu ( 1977 ) hypothesised that the ti plasmid could be a natural vector for introducing the nif gene responsible for nitrogen fixation in the root nodules of legumes and other plant species. today, genetic modification of the ti plasmid is one of the main techniques for introduction of transgenes to plants and the creation of genetically modified crops. = = = epigenetics = = = epigenetics is the study of heritable changes in gene function that cannot be explained by changes in the underlying dna sequence but cause the organism ' s genes to behave ( or " express themselves " ) differently. one example of epigenetic change is the marking of the genes by dna methylation which determines whether they will be expressed or not. gene expression can also be controlled by repressor proteins that attach to silencer regions of the dna and prevent that region of the dna code from being expressed. epigenetic marks may be added
of several methods used by plants to promote outcrossing. in many land plants the male and female gametes are produced by separate individuals. these species are said to be dioecious when referring to vascular plant sporophytes and dioicous when referring to bryophyte gametophytes. charles darwin in his 1878 book the effects of cross and self - fertilization in the vegetable kingdom at the start of chapter xii noted " the first and most important of the conclusions which may be drawn from the observations given in this volume, is that generally cross - fertilisation is beneficial and self - fertilisation often injurious, at least with the plants on which i experimented. " an important adaptive benefit of outcrossing is that it allows the masking of deleterious mutations in the genome of progeny. this beneficial effect is also known as hybrid vigor or heterosis. once outcrossing is established, subsequent switching to inbreeding becomes disadvantageous since it allows expression of the previously masked deleterious recessive mutations, commonly referred to as inbreeding depression. unlike in higher animals, where parthenogenesis is rare, asexual reproduction may occur in plants by several different mechanisms. the formation of stem tubers in potato is one example. particularly in arctic or alpine habitats, where opportunities for fertilisation of flowers by animals are rare, plantlets or bulbs, may develop instead of flowers, replacing sexual reproduction with asexual reproduction and giving rise to clonal populations genetically identical to the parent. this is one of several types of apomixis that occur in plants. apomixis can also happen in a seed, producing a seed that contains an embryo genetically identical to the parent. most sexually reproducing organisms are diploid, with paired chromosomes, but doubling of their chromosome number may occur due to errors in cytokinesis. this can occur early in development to produce an autopolyploid or partly autopolyploid organism, or during normal processes of cellular differentiation to produce some cell types that are polyploid ( endopolyploidy ), or during gamete formation. an allopolyploid plant may result from a hybridisation event between two different species. both autopolyploid and allopolyploid plants can often reproduce normally, but may be unable to cross - breed successfully with the parent population because there is a mismatch in chromosome numbers. these plants that are reproductively isolated from the parent
mitochondria and chloroplasts in plants is non - mendelian. chloroplasts are inherited through the male parent in gymnosperms but often through the female parent in flowering plants. = = = molecular genetics = = = a considerable amount of new knowledge about plant function comes from studies of the molecular genetics of model plants such as the thale cress, arabidopsis thaliana, a weedy species in the mustard family ( brassicaceae ). the genome or hereditary information contained in the genes of this species is encoded by about 135 million base pairs of dna, forming one of the smallest genomes among flowering plants. arabidopsis was the first plant to have its genome sequenced, in 2000. the sequencing of some other relatively small genomes, of rice ( oryza sativa ) and brachypodium distachyon, has made them important model species for understanding the genetics, cellular and molecular biology of cereals, grasses and monocots generally. model plants such as arabidopsis thaliana are used for studying the molecular biology of plant cells and the chloroplast. ideally, these organisms have small genomes that are well known or completely sequenced, small stature and short generation times. corn has been used to study mechanisms of photosynthesis and phloem loading of sugar in c4 plants. the single celled green alga chlamydomonas reinhardtii, while not an embryophyte itself, contains a green - pigmented chloroplast related to that of land plants, making it useful for study. a red alga cyanidioschyzon merolae has also been used to study some basic chloroplast functions. spinach, peas, soybeans and a moss physcomitrella patens are commonly used to study plant cell biology. agrobacterium tumefaciens, a soil rhizosphere bacterium, can attach to plant cells and infect them with a callus - inducing ti plasmid by horizontal gene transfer, causing a callus infection called crown gall disease. schell and van montagu ( 1977 ) hypothesised that the ti plasmid could be a natural vector for introducing the nif gene responsible for nitrogen fixation in the root nodules of legumes and other plant species. today, genetic modification of the ti plasmid is one of the main techniques for introduction of transgenes to plants
Question: Gregor Mendel studied ratios related to traits passed from parent pea plants to offspring pea plants. Mendel used large numbers of plants to avoid "sampling error." This changed the way scientists studied genetics because it included ideas of
A) nuclei and DNA.
B) atoms and molecules.
C) ecology and environment.
D) mathematics and probability.
|
D) mathematics and probability.
|
Context:
in the year 1598 philipp uffenbach published a printed diptych sundial, which is a forerunner of franz ritters horizantal sundial. uffenbach ' s sundial contains apart from the usual information on a sundial ascending signs of the zodiac, several brigthest stars, an almucantar and most important the oldest gnomonic world map known so far. the sundial is constructed for the polar height of 50 1 / 6 degrees, the height of frankfurt / main the town of his citizenship.
variation in total solar irradiance is thought to have little effect on the earth ' s surface temperature because of the thermal time constant - - the characteristic response time of the earth ' s global surface temperature to changes in forcing. this time constant is large enough to smooth annual variations but not necessarily variations having a longer period such as those due to solar inertial motion ; the magnitude of these surface temperature variations is estimated.
weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under ordinary conditions, most of the materials brought down from the high lands by torrential water courses are carried forward by the main river to the sea, or partially strewn over flat alluvial plains during floods ; the size of the materials forming the bed of the river or borne along by the stream is gradually reduced on proceeding seawards, so that in the po river in italy, for instance, pebbles and gravel are found for about 140 miles below turin, sand along the next 100 miles, and silt and mud in the last 110 miles ( 176 km ). = = channelization = = the removal of obstructions, natural or artificial
oscillations of the sun have been used to understand its interior structure. the extension of similar studies to more distant stars has raised many difficulties despite the strong efforts of the international community over the past decades. the corot ( convection rotation and planetary transits ) satellite, launched in december 2006, has now measured oscillations and the stellar granulation signature in three main sequence stars that are noticeably hotter than the sun. the oscillation amplitudes are about 1. 5 times as large as those in the sun ; the stellar granulation is up to three times as high. the stellar amplitudes are about 25 % below the theoretic values, providing a measurement of the nonadiabaticity of the process ruling the oscillations in the outer layers of the stars.
navigable capabilities of rivers. in tropical countries subject to periodical rains, the rivers are in flood during the rainy season and have hardly any flow during the rest of the year, while in temperate regions, where the rainfall is more evenly distributed throughout the year, evaporation causes the available rainfall to be much less in hot summer weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under ordinary conditions, most of the materials brought down from the high lands by torrential water courses are carried forward by the main river to the sea, or partially strewn over flat alluvial plains during floods ; the size of the materials forming the bed of the river or borne along by the stream is gradually reduced on proceeding sea
cools and solidifies. through subduction, oceanic crust and lithosphere vehemently returns to the convecting mantle. volcanoes result primarily from the melting of subducted crust material. crust material that is forced into the asthenosphere melts, and some portion of the melted material becomes light enough to rise to the surface β giving birth to volcanoes. = = atmospheric science = = atmospheric science initially developed in the late - 19th century as a means to forecast the weather through meteorology, the study of weather. atmospheric chemistry was developed in the 20th century to measure air pollution and expanded in the 1970s in response to acid rain. climatology studies the climate and climate change. the troposphere, stratosphere, mesosphere, thermosphere, and exosphere are the five layers which make up earth ' s atmosphere. 75 % of the mass in the atmosphere is located within the troposphere, the lowest layer. in all, the atmosphere is made up of about 78. 0 % nitrogen, 20. 9 % oxygen, and 0. 92 % argon, and small amounts of other gases including co2 and water vapor. water vapor and co2 cause the earth ' s atmosphere to catch and hold the sun ' s energy through the greenhouse effect. this makes earth ' s surface warm enough for liquid water and life. in addition to trapping heat, the atmosphere also protects living organisms by shielding the earth ' s surface from cosmic rays. the magnetic field β created by the internal motions of the core β produces the magnetosphere which protects earth ' s atmosphere from the solar wind. as the earth is 4. 5 billion years old, it would have lost its atmosphere by now if there were no protective magnetosphere. = = earth ' s magnetic field = = = = hydrology = = hydrology is the study of the hydrosphere and the movement of water on earth. it emphasizes the study of how humans use and interact with freshwater supplies. study of water ' s movement is closely related to geomorphology and other branches of earth science. applied hydrology involves engineering to maintain aquatic environments and distribute water supplies. subdisciplines of hydrology include oceanography, hydrogeology, ecohydrology, and glaciology. oceanography is the study of oceans. hydrogeology is the study of groundwater. it includes the mapping of groundwater supplies and the analysis of groundwater contaminants. applied hydrogeology seeks to prevent contamination of groundwater and mineral springs and make
becomes quite gentle. accordingly, in large basins, rivers in most cases begin as torrents with a variable flow, and end as gently flowing rivers with a comparatively regular discharge. the irregular flow of rivers throughout their course forms one of the main difficulties in devising works for mitigating inundations or for increasing the navigable capabilities of rivers. in tropical countries subject to periodical rains, the rivers are in flood during the rainy season and have hardly any flow during the rest of the year, while in temperate regions, where the rainfall is more evenly distributed throughout the year, evaporation causes the available rainfall to be much less in hot summer weather than in the winter months, so that the rivers fall to their low stage in the summer and are liable to be in flood in the winter. in fact, with a temperate climate, the year may be divided into a warm and a cold season, extending from may to october and from november to april in the northern hemisphere respectively ; the rivers are low and moderate floods are of rare occurrence during the warm period, and the rivers are high and subject to occasional heavy floods after a considerable rainfall during the cold period in most years. the only exceptions are rivers which have their sources amongst mountains clad with perpetual snow and are fed by glaciers ; their floods occur in the summer from the melting of snow and ice, as exemplified by the rhone above the lake of geneva, and the arve which joins it below. but even these rivers are liable to have their flow modified by the influx of tributaries subject to different conditions, so that the rhone below lyon has a more uniform discharge than most rivers, as the summer floods of the arve are counteracted to a great extent by the low stage of the saone flowing into the rhone at lyon, which has its floods in the winter when the arve, on the contrary, is low. another serious obstacle encountered in river engineering consists in the large quantity of detritus they bring down in flood - time, derived mainly from the disintegration of the surface layers of the hills and slopes in the upper parts of the valleys by glaciers, frost and rain. the power of a current to transport materials varies with its velocity, so that torrents with a rapid fall near the sources of rivers can carry down rocks, boulders and large stones, which are by degrees ground by attrition in their onward course into slate, gravel, sand and silt, simultaneously with the gradual reduction in fall, and, consequently, in the transporting force of the current. accordingly, under
the first observations of saturn ' s visible - wavelength aurora were made by the cassini camera. the aurora was observed between 2006 and 2013 in the northern and southern hemispheres. the color of the aurora changes from pink at a few hundred km above the horizon to purple at 1000 - 1500 km above the horizon. the spectrum observed in 9 filters spanning wavelengths from 250 nm to 1000 nm has a prominent h - alpha line and roughly agrees with laboratory simulated auroras. auroras in both hemispheres vary dramatically with longitude. auroras form bright arcs between 70 and 80 degree latitude north and between 65 and 80 degree latitude south, which sometimes spiral around the pole, and sometimes form double arcs. a large 10, 000 - km - scale longitudinal brightness structure persists for more than 100 hours. this structure rotates approximately together with saturn. on top of the large steady structure, the auroras brighten suddenly on the timescales of a few minutes. these brightenings repeat with a period of about 1 hour. smaller, 1000 - km - scale structures may move faster or lag behind saturn ' s rotation on timescales of tens of minutes. the persistence of nearly - corotating large bright longitudinal structure in the auroral oval seen in two movies spanning 8 and 11 rotations gives an estimate on the period of 10. 65 $ \ pm $ 0. 15 h for 2009 in the northern oval and 10. 8 $ \ pm $ 0. 1 h for 2012 in the southern oval. the 2009 north aurora period is close to the north branch of saturn kilometric radiation ( skr ) detected at that time.
observed solar neutrino fluxes are employed to constrain the interior composition of the sun. including the effects of neutrino flavor mixing, the results from homestake, sudbury, and gallium experiments constrain the mg, si, and fe abundances in the solar interior to be within a factor 0. 89 to 1. 34 of the surface values with 68 % confidence. if the o and / or ne abundances are increased in the interior to resolve helioseismic discrepancies with recent standard solar models, then the nominal interior mg, si, and fe abundances are constrained to a range of 0. 83 to 1. 24 relative to the surface. additional research is needed to determine whether the sun ' s interior is metal poor relative to its surface.
a letter to the editor shortly summing up ten or so years of research into the h - index.
Question: What season is the Northern Hemisphere experiencing when it is tilted directly toward the Sun?
A) fall
B) winter
C) spring
D) summer
|
D) summer
|
Context:
have evolved from the earliest emergence of life to present day. earth formed about 4. 5 billion years ago and all life on earth, both living and extinct, descended from a last universal common ancestor that lived about 3. 5 billion years ago. geologists have developed a geologic time scale that divides the history of the earth into major divisions, starting with four eons ( hadean, archean, proterozoic, and phanerozoic ), the first three of which are collectively known as the precambrian, which lasted approximately 4 billion years. each eon can be divided into eras, with the phanerozoic eon that began 539 million years ago being subdivided into paleozoic, mesozoic, and cenozoic eras. these three eras together comprise eleven periods ( cambrian, ordovician, silurian, devonian, carboniferous, permian, triassic, jurassic, cretaceous, tertiary, and quaternary ). the similarities among all known present - day species indicate that they have diverged through the process of evolution from their common ancestor. biologists regard the ubiquity of the genetic code as evidence of universal common descent for all bacteria, archaea, and eukaryotes. microbial mats of coexisting bacteria and archaea were the dominant form of life in the early archean eon and many of the major steps in early evolution are thought to have taken place in this environment. the earliest evidence of eukaryotes dates from 1. 85 billion years ago, and while they may have been present earlier, their diversification accelerated when they started using oxygen in their metabolism. later, around 1. 7 billion years ago, multicellular organisms began to appear, with differentiated cells performing specialised functions. algae - like multicellular land plants are dated back to about 1 billion years ago, although evidence suggests that microorganisms formed the earliest terrestrial ecosystems, at least 2. 7 billion years ago. microorganisms are thought to have paved the way for the inception of land plants in the ordovician period. land plants were so successful that they are thought to have contributed to the late devonian extinction event. ediacara biota appear during the ediacaran period, while vertebrates, along with most other modern phyla originated about 525 million years ago during the cambrian explosion. during the permian period, synapsids, including the ancestors of mammals, dominated the land, but most of this group became
they may have been present earlier, their diversification accelerated when they started using oxygen in their metabolism. later, around 1. 7 billion years ago, multicellular organisms began to appear, with differentiated cells performing specialised functions. algae - like multicellular land plants are dated back to about 1 billion years ago, although evidence suggests that microorganisms formed the earliest terrestrial ecosystems, at least 2. 7 billion years ago. microorganisms are thought to have paved the way for the inception of land plants in the ordovician period. land plants were so successful that they are thought to have contributed to the late devonian extinction event. ediacara biota appear during the ediacaran period, while vertebrates, along with most other modern phyla originated about 525 million years ago during the cambrian explosion. during the permian period, synapsids, including the ancestors of mammals, dominated the land, but most of this group became extinct in the permian β triassic extinction event 252 million years ago. during the recovery from this catastrophe, archosaurs became the most abundant land vertebrates ; one archosaur group, the dinosaurs, dominated the jurassic and cretaceous periods. after the cretaceous β paleogene extinction event 66 million years ago killed off the non - avian dinosaurs, mammals increased rapidly in size and diversity. such mass extinctions may have accelerated evolution by providing opportunities for new groups of organisms to diversify. = = diversity = = = = = bacteria and archaea = = = bacteria are a type of cell that constitute a large domain of prokaryotic microorganisms. typically a few micrometers in length, bacteria have a number of shapes, ranging from spheres to rods and spirals. bacteria were among the first life forms to appear on earth, and are present in most of its habitats. bacteria inhabit soil, water, acidic hot springs, radioactive waste, and the deep biosphere of the earth ' s crust. bacteria also live in symbiotic and parasitic relationships with plants and animals. most bacteria have not been characterised, and only about 27 percent of the bacterial phyla have species that can be grown in the laboratory. archaea constitute the other domain of prokaryotic cells and were initially classified as bacteria, receiving the name archaebacteria ( in the archaebacteria kingdom ), a term that has fallen out of use. archaeal cells have unique properties separating them from the other two domains, bacteria and eukaryota. archaea
. microbial mats of coexisting bacteria and archaea were the dominant form of life in the early archean eon and many of the major steps in early evolution are thought to have taken place in this environment. the earliest evidence of eukaryotes dates from 1. 85 billion years ago, and while they may have been present earlier, their diversification accelerated when they started using oxygen in their metabolism. later, around 1. 7 billion years ago, multicellular organisms began to appear, with differentiated cells performing specialised functions. algae - like multicellular land plants are dated back to about 1 billion years ago, although evidence suggests that microorganisms formed the earliest terrestrial ecosystems, at least 2. 7 billion years ago. microorganisms are thought to have paved the way for the inception of land plants in the ordovician period. land plants were so successful that they are thought to have contributed to the late devonian extinction event. ediacara biota appear during the ediacaran period, while vertebrates, along with most other modern phyla originated about 525 million years ago during the cambrian explosion. during the permian period, synapsids, including the ancestors of mammals, dominated the land, but most of this group became extinct in the permian β triassic extinction event 252 million years ago. during the recovery from this catastrophe, archosaurs became the most abundant land vertebrates ; one archosaur group, the dinosaurs, dominated the jurassic and cretaceous periods. after the cretaceous β paleogene extinction event 66 million years ago killed off the non - avian dinosaurs, mammals increased rapidly in size and diversity. such mass extinctions may have accelerated evolution by providing opportunities for new groups of organisms to diversify. = = diversity = = = = = bacteria and archaea = = = bacteria are a type of cell that constitute a large domain of prokaryotic microorganisms. typically a few micrometers in length, bacteria have a number of shapes, ranging from spheres to rods and spirals. bacteria were among the first life forms to appear on earth, and are present in most of its habitats. bacteria inhabit soil, water, acidic hot springs, radioactive waste, and the deep biosphere of the earth ' s crust. bacteria also live in symbiotic and parasitic relationships with plants and animals. most bacteria have not been characterised, and only about 27 percent of the bacterial phyla have species that can be grown in the laboratory. archaea constitute the other domain of
, tertiary, and quaternary ). the similarities among all known present - day species indicate that they have diverged through the process of evolution from their common ancestor. biologists regard the ubiquity of the genetic code as evidence of universal common descent for all bacteria, archaea, and eukaryotes. microbial mats of coexisting bacteria and archaea were the dominant form of life in the early archean eon and many of the major steps in early evolution are thought to have taken place in this environment. the earliest evidence of eukaryotes dates from 1. 85 billion years ago, and while they may have been present earlier, their diversification accelerated when they started using oxygen in their metabolism. later, around 1. 7 billion years ago, multicellular organisms began to appear, with differentiated cells performing specialised functions. algae - like multicellular land plants are dated back to about 1 billion years ago, although evidence suggests that microorganisms formed the earliest terrestrial ecosystems, at least 2. 7 billion years ago. microorganisms are thought to have paved the way for the inception of land plants in the ordovician period. land plants were so successful that they are thought to have contributed to the late devonian extinction event. ediacara biota appear during the ediacaran period, while vertebrates, along with most other modern phyla originated about 525 million years ago during the cambrian explosion. during the permian period, synapsids, including the ancestors of mammals, dominated the land, but most of this group became extinct in the permian β triassic extinction event 252 million years ago. during the recovery from this catastrophe, archosaurs became the most abundant land vertebrates ; one archosaur group, the dinosaurs, dominated the jurassic and cretaceous periods. after the cretaceous β paleogene extinction event 66 million years ago killed off the non - avian dinosaurs, mammals increased rapidly in size and diversity. such mass extinctions may have accelerated evolution by providing opportunities for new groups of organisms to diversify. = = diversity = = = = = bacteria and archaea = = = bacteria are a type of cell that constitute a large domain of prokaryotic microorganisms. typically a few micrometers in length, bacteria have a number of shapes, ranging from spheres to rods and spirals. bacteria were among the first life forms to appear on earth, and are present in most of its habitats. bacteria inhabit soil, water, acidic hot springs, radioactive
into major divisions, starting with four eons ( hadean, archean, proterozoic, and phanerozoic ), the first three of which are collectively known as the precambrian, which lasted approximately 4 billion years. each eon can be divided into eras, with the phanerozoic eon that began 539 million years ago being subdivided into paleozoic, mesozoic, and cenozoic eras. these three eras together comprise eleven periods ( cambrian, ordovician, silurian, devonian, carboniferous, permian, triassic, jurassic, cretaceous, tertiary, and quaternary ). the similarities among all known present - day species indicate that they have diverged through the process of evolution from their common ancestor. biologists regard the ubiquity of the genetic code as evidence of universal common descent for all bacteria, archaea, and eukaryotes. microbial mats of coexisting bacteria and archaea were the dominant form of life in the early archean eon and many of the major steps in early evolution are thought to have taken place in this environment. the earliest evidence of eukaryotes dates from 1. 85 billion years ago, and while they may have been present earlier, their diversification accelerated when they started using oxygen in their metabolism. later, around 1. 7 billion years ago, multicellular organisms began to appear, with differentiated cells performing specialised functions. algae - like multicellular land plants are dated back to about 1 billion years ago, although evidence suggests that microorganisms formed the earliest terrestrial ecosystems, at least 2. 7 billion years ago. microorganisms are thought to have paved the way for the inception of land plants in the ordovician period. land plants were so successful that they are thought to have contributed to the late devonian extinction event. ediacara biota appear during the ediacaran period, while vertebrates, along with most other modern phyla originated about 525 million years ago during the cambrian explosion. during the permian period, synapsids, including the ancestors of mammals, dominated the land, but most of this group became extinct in the permian β triassic extinction event 252 million years ago. during the recovery from this catastrophe, archosaurs became the most abundant land vertebrates ; one archosaur group, the dinosaurs, dominated the jurassic and cretaceous periods. after the cretaceous β paleogene extinction event 66 million years ago killed off
variation in total solar irradiance is thought to have little effect on the earth ' s surface temperature because of the thermal time constant - - the characteristic response time of the earth ' s global surface temperature to changes in forcing. this time constant is large enough to smooth annual variations but not necessarily variations having a longer period such as those due to solar inertial motion ; the magnitude of these surface temperature variations is estimated.
##ning an animal for its hide, and even forming other tools out of softer materials such as bone and wood. the earliest stone tools were irrelevant, being little more than a fractured rock. in the acheulian era, beginning approximately 1. 65 million years ago, methods of working these stones into specific shapes, such as hand axes emerged. this early stone age is described as the lower paleolithic. the middle paleolithic, approximately 300, 000 years ago, saw the introduction of the prepared - core technique, where multiple blades could be rapidly formed from a single core stone. the upper paleolithic, beginning approximately 40, 000 years ago, saw the introduction of pressure flaking, where a wood, bone, or antler punch could be used to shape a stone very finely. the end of the last ice age about 10, 000 years ago is taken as the end point of the upper paleolithic and the beginning of the epipaleolithic / mesolithic. the mesolithic technology included the use of microliths as composite stone tools, along with wood, bone, and antler tools. the later stone age, during which the rudiments of agricultural technology were developed, is called the neolithic period. during this period, polished stone tools were made from a variety of hard rocks such as flint, jade, jadeite, and greenstone, largely by working exposures as quarries, but later the valuable rocks were pursued by tunneling underground, the first steps in mining technology. the polished axes were used for forest clearance and the establishment of crop farming and were so effective as to remain in use when bronze and iron appeared. these stone axes were used alongside a continued use of stone tools such as a range of projectiles, knives, and scrapers, as well as tools, made from organic materials such as wood, bone, and antler. stone age cultures developed music and engaged in organized warfare. stone age humans developed ocean - worthy outrigger canoe technology, leading to migration across the malay archipelago, across the indian ocean to madagascar and also across the pacific ocean, which required knowledge of the ocean currents, weather patterns, sailing, and celestial navigation. although paleolithic cultures left no written records, the shift from nomadic life to settlement and agriculture can be inferred from a range of archaeological evidence. such evidence includes ancient tools, cave paintings, and other prehistoric art, such as the venus of willendorf. human remains also provide direct evidence, both through the examination of bones, and
smallest genomes among flowering plants. arabidopsis was the first plant to have its genome sequenced, in 2000. the sequencing of some other relatively small genomes, of rice ( oryza sativa ) and brachypodium distachyon, has made them important model species for understanding the genetics, cellular and molecular biology of cereals, grasses and monocots generally. model plants such as arabidopsis thaliana are used for studying the molecular biology of plant cells and the chloroplast. ideally, these organisms have small genomes that are well known or completely sequenced, small stature and short generation times. corn has been used to study mechanisms of photosynthesis and phloem loading of sugar in c4 plants. the single celled green alga chlamydomonas reinhardtii, while not an embryophyte itself, contains a green - pigmented chloroplast related to that of land plants, making it useful for study. a red alga cyanidioschyzon merolae has also been used to study some basic chloroplast functions. spinach, peas, soybeans and a moss physcomitrella patens are commonly used to study plant cell biology. agrobacterium tumefaciens, a soil rhizosphere bacterium, can attach to plant cells and infect them with a callus - inducing ti plasmid by horizontal gene transfer, causing a callus infection called crown gall disease. schell and van montagu ( 1977 ) hypothesised that the ti plasmid could be a natural vector for introducing the nif gene responsible for nitrogen fixation in the root nodules of legumes and other plant species. today, genetic modification of the ti plasmid is one of the main techniques for introduction of transgenes to plants and the creation of genetically modified crops. = = = epigenetics = = = epigenetics is the study of heritable changes in gene function that cannot be explained by changes in the underlying dna sequence but cause the organism ' s genes to behave ( or " express themselves " ) differently. one example of epigenetic change is the marking of the genes by dna methylation which determines whether they will be expressed or not. gene expression can also be controlled by repressor proteins that attach to silencer regions of the dna and prevent that region of the dna code from being expressed. epigenetic marks may be added
##wi, turkana, dating from 3. 3 million years ago. stone tools diversified through the pleistocene period, which ended ~ 12, 000 years ago. the earliest evidence of warfare between two groups is recorded at the site of nataruk in turkana, kenya, where human skeletons with major traumatic injuries to the head, neck, ribs, knees and hands, including an embedded obsidian bladelet on a skull, are evidence of inter - group conflict between groups of nomadic hunter - gatherers 10, 000 years ago. humans entered the bronze age as they learned to smelt copper into an alloy with tin to make weapons. in asia where copper - tin ores are rare, this development was delayed until trading in bronze began in the third millennium bce. in the middle east and southern european regions, the bronze age follows the neolithic period, but in other parts of the world, the copper age is a transition from neolithic to the bronze age. although the iron age generally follows the bronze age, in some areas the iron age intrudes directly on the neolithic from outside the region, with the exception of sub - saharan africa where it was developed independently. the first large - scale use of iron weapons began in asia minor around the 14th century bce and in central europe around the 11th century bce followed by the middle east ( about 1000 bce ) and india and china. the assyrians are credited with the introduction of horse cavalry in warfare and the extensive use of iron weapons by 1100 bce. assyrians were also the first to use iron - tipped arrows. = = = post - classical technology = = = the wujing zongyao ( essentials of the military arts ), written by zeng gongliang, ding du, and others at the order of emperor renzong around 1043 during the song dynasty illustrate the eras focus on advancing intellectual issues and military technology due to the significance of warfare between the song and the liao, jin, and yuan to their north. the book covers topics of military strategy, training, and the production and employment of advanced weaponry. advances in military technology aided the song dynasty in its defense against hostile neighbors to the north. the flamethrower found its origins in byzantine - era greece, employing greek fire ( a chemically complex, highly flammable petrol fluid ) in a device with a siphon hose by the 7th century. : 77 the earliest reference to greek fire in china was made in 917, written by wu renchen in his spring and autumn annals of the ten kingdoms. : 80 in 91
are the stone - paved streets of the city - state of ur, dating to c. 4, 000 bce, and timber roads leading through the swamps of glastonbury, england, dating to around the same period. the first long - distance road, which came into use around 3, 500 bce, spanned 2, 400 km from the persian gulf to the mediterranean sea, but was not paved and was only partially maintained. in around 2, 000 bce, the minoans on the greek island of crete built a 50 km road leading from the palace of gortyn on the south side of the island, through the mountains, to the palace of knossos on the north side of the island. unlike the earlier road, the minoan road was completely paved. ancient minoan private homes had running water. a bathtub virtually identical to modern ones was unearthed at the palace of knossos. several minoan private homes also had toilets, which could be flushed by pouring water down the drain. the ancient romans had many public flush toilets, which emptied into an extensive sewage system. the primary sewer in rome was the cloaca maxima ; construction began on it in the sixth century bce and it is still in use today. the ancient romans also had a complex system of aqueducts, which were used to transport water across long distances. the first roman aqueduct was built in 312 bce. the eleventh and final ancient roman aqueduct was built in 226 ce. put together, the roman aqueducts extended over 450 km, but less than 70 km of this was above ground and supported by arches. = = = pre - modern = = = innovations continued through the middle ages with the introduction of silk production ( in asia and later europe ), the horse collar, and horseshoes. simple machines ( such as the lever, the screw, and the pulley ) were combined into more complicated tools, such as the wheelbarrow, windmills, and clocks. a system of universities developed and spread scientific ideas and practices, including oxford and cambridge. the renaissance era produced many innovations, including the introduction of the movable type printing press to europe, which facilitated the communication of knowledge. technology became increasingly influenced by science, beginning a cycle of mutual advancement. = = = modern = = = starting in the united kingdom in the 18th century, the discovery of steam power set off the industrial revolution, which saw wide - ranging technological discoveries, particularly in the areas of agriculture, manufacturing, mining, metallurgy, and transport, and the
Question: Which species have been on the Earth for the shortest amount of time?
A) Humans
B) Insects
C) Fish
D) Reptiles
|
A) Humans
|
Context:
classes according to pore size : the form and shape of the membrane pores are highly dependent on the manufacturing process and are often difficult to specify. therefore, for characterization, test filtrations are carried out and the pore diameter refers to the diameter of the smallest particles which could not pass through the membrane. the rejection can be determined in various ways and provides an indirect measurement of the pore size. one possibility is the filtration of macromolecules ( often dextran, polyethylene glycol or albumin ), another is measurement of the cut - off by gel permeation chromatography. these methods are used mainly to measure membranes for ultrafiltration applications. another testing method is the filtration of particles with defined size and their measurement with a particle sizer or by laser induced breakdown spectroscopy ( libs ). a vivid characterization is to measure the rejection of dextran blue or other colored molecules. the retention of bacteriophage and bacteria, the so - called " bacteria challenge test ", can also provide information about the pore size. to determine the pore diameter, physical methods such as porosimeter ( mercury, liquid - liquid porosimeter and bubble point test ) are also used, but a certain form of the pores ( such as cylindrical or concatenated spherical holes ) is assumed. such methods are used for membranes whose pore geometry does not match the ideal, and we get " nominal " pore diameter, which characterizes the membrane, but does not necessarily reflect its actual filtration behavior and selectivity. the selectivity is highly dependent on the separation process, the composition of the membrane and its electrochemical properties in addition to the pore size. with high selectivity, isotopes can be enriched ( uranium enrichment ) in nuclear engineering or industrial gases like nitrogen can be recovered ( gas separation ). ideally, even racemics can be enriched with a suitable membrane. when choosing membranes selectivity has priority over a high permeability, as low flows can easily be offset by increasing the filter surface with a modular structure. in gas phase filtration different deposition mechanisms are operative, so that particles having sizes below the pore size of the membrane can be retained as well. = = membrane classification = = bio - membrane is classified in two categories, synthetic membrane and natural membrane. synthetic membranes further classified in organic and inorganic membranes. organic membrane sub classified polymeric membranes and inorganic membrane sub classified ceramic polymers. = = synthesis of biomass membrane
frits went. the first known auxin, indole - 3 - acetic acid ( iaa ), which promotes cell growth, was only isolated from plants about 50 years later. this compound mediates the tropic responses of shoots and roots towards light and gravity. the finding in 1939 that plant callus could be maintained in culture containing iaa, followed by the observation in 1947 that it could be induced to form roots and shoots by controlling the concentration of growth hormones were key steps in the development of plant biotechnology and genetic modification. cytokinins are a class of plant hormones named for their control of cell division ( especially cytokinesis ). the natural cytokinin zeatin was discovered in corn, zea mays, and is a derivative of the purine adenine. zeatin is produced in roots and transported to shoots in the xylem where it promotes cell division, bud development, and the greening of chloroplasts. the gibberelins, such as gibberelic acid are diterpenes synthesised from acetyl coa via the mevalonate pathway. they are involved in the promotion of germination and dormancy - breaking in seeds, in regulation of plant height by controlling stem elongation and the control of flowering. abscisic acid ( aba ) occurs in all land plants except liverworts, and is synthesised from carotenoids in the chloroplasts and other plastids. it inhibits cell division, promotes seed maturation, and dormancy, and promotes stomatal closure. it was so named because it was originally thought to control abscission. ethylene is a gaseous hormone that is produced in all higher plant tissues from methionine. it is now known to be the hormone that stimulates or regulates fruit ripening and abscission, and it, or the synthetic growth regulator ethephon which is rapidly metabolised to produce ethylene, are used on industrial scale to promote ripening of cotton, pineapples and other climacteric crops. another class of phytohormones is the jasmonates, first isolated from the oil of jasminum grandiflorum which regulates wound responses in plants by unblocking the expression of genes required in the systemic acquired resistance response to pathogen attack. in addition to being the primary energy source for plants, light functions as a signalling device, providing information to the plant, such as how
three of what is called the six simple machines, from which all machines are based. these machines are the inclined plane, the wedge, and the lever, which allowed the ancient egyptians to move millions of limestone blocks which weighed approximately 3. 5 tons ( 7, 000 lbs. ) each into place to create structures like the great pyramid of giza, which is 481 feet ( 147 meters ) high. they also made writing medium similar to paper from papyrus, which joshua mark states is the foundation for modern paper. papyrus is a plant ( cyperus papyrus ) which grew in plentiful amounts in the egyptian delta and throughout the nile river valley during ancient times. the papyrus was harvested by field workers and brought to processing centers where it was cut into thin strips. the strips were then laid - out side by side and covered in plant resin. the second layer of strips was laid on perpendicularly, then both pressed together until the sheet was dry. the sheets were then joined to form a roll and later used for writing. egyptian society made several significant advances during dynastic periods in many areas of technology. according to hossam elanzeery, they were the first civilization to use timekeeping devices such as sundials, shadow clocks, and obelisks and successfully leveraged their knowledge of astronomy to create a calendar model that society still uses today. they developed shipbuilding technology that saw them progress from papyrus reed vessels to cedar wood ships while also pioneering the use of rope trusses and stem - mounted rudders. the egyptians also used their knowledge of anatomy to lay the foundation for many modern medical techniques and practiced the earliest known version of neuroscience. elanzeery also states that they used and furthered mathematical science, as evidenced in the building of the pyramids. ancient egyptians also invented and pioneered many food technologies that have become the basis of modern food technology processes. based on paintings and reliefs found in tombs, as well as archaeological artifacts, scholars like paul t nicholson believe that the ancient egyptians established systematic farming practices, engaged in cereal processing, brewed beer and baked bread, processed meat, practiced viticulture and created the basis for modern wine production, and created condiments to complement, preserve and mask the flavors of their food. = = = = indus valley = = = = the indus valley civilization, situated in a resource - rich area ( in modern pakistan and northwestern india ), is notable for its early application of city planning, sanitation technologies, and plumbing. indus valley construction and architecture, called ' vaastu
the broad definition of " utilizing a biotechnological system to make products ". indeed, the cultivation of plants may be viewed as the earliest biotechnological enterprise. agriculture has been theorized to have become the dominant way of producing food since the neolithic revolution. through early biotechnology, the earliest farmers selected and bred the best - suited crops ( e. g., those with the highest yields ) to produce enough food to support a growing population. as crops and fields became increasingly large and difficult to maintain, it was discovered that specific organisms and their by - products could effectively fertilize, restore nitrogen, and control pests. throughout the history of agriculture, farmers have inadvertently altered the genetics of their crops through introducing them to new environments and breeding them with other plants β one of the first forms of biotechnology. these processes also were included in early fermentation of beer. these processes were introduced in early mesopotamia, egypt, china and india, and still use the same basic biological methods. in brewing, malted grains ( containing enzymes ) convert starch from grains into sugar and then adding specific yeasts to produce beer. in this process, carbohydrates in the grains broke down into alcohols, such as ethanol. later, other cultures produced the process of lactic acid fermentation, which produced other preserved foods, such as soy sauce. fermentation was also used in this time period to produce leavened bread. although the process of fermentation was not fully understood until louis pasteur ' s work in 1857, it is still the first use of biotechnology to convert a food source into another form. before the time of charles darwin ' s work and life, animal and plant scientists had already used selective breeding. darwin added to that body of work with his scientific observations about the ability of science to change species. these accounts contributed to darwin ' s theory of natural selection. for thousands of years, humans have used selective breeding to improve the production of crops and livestock to use them for food. in selective breeding, organisms with desirable characteristics are mated to produce offspring with the same characteristics. for example, this technique was used with corn to produce the largest and sweetest crops. in the early twentieth century scientists gained a greater understanding of microbiology and explored ways of manufacturing specific products. in 1917, chaim weizmann first used a pure microbiological culture in an industrial process, that of manufacturing corn starch using clostridium acetobutylicum, to produce acetone, which the united
significantly greater strength and fracture toughness. another major change in the body during the firing or sintering process will be the establishment of the polycrystalline nature of the solid. significant grain growth tends to occur during sintering, with this growth depending on temperature and duration of the sintering process. the growth of grains will result in some form of grain size distribution, which will have a significant impact on the ultimate physical properties of the material. in particular, abnormal grain growth in which certain grains grow very large in a matrix of finer grains will significantly alter the physical and mechanical properties of the obtained ceramic. in the sintered body, grain sizes are a product of the thermal processing parameters as well as the initial particle size, or possibly the sizes of aggregates or particle clusters which arise during the initial stages of processing. the ultimate microstructure ( and thus the physical properties ) of the final product will be limited by and subject to the form of the structural template or precursor which is created in the initial stages of chemical synthesis and physical forming. hence the importance of chemical powder and polymer processing as it pertains to the synthesis of industrial ceramics, glasses and glass - ceramics. there are numerous possible refinements of the sintering process. some of the most common involve pressing the green body to give the densification a head start and reduce the sintering time needed. sometimes organic binders such as polyvinyl alcohol are added to hold the green body together ; these burn out during the firing ( at 200 β 350 Β°c ). sometimes organic lubricants are added during pressing to increase densification. it is common to combine these, and add binders and lubricants to a powder, then press. ( the formulation of these organic chemical additives is an art in itself. this is particularly important in the manufacture of high performance ceramics such as those used by the billions for electronics, in capacitors, inductors, sensors, etc. ) a slurry can be used in place of a powder, and then cast into a desired shape, dried and then sintered. indeed, traditional pottery is done with this type of method, using a plastic mixture worked with the hands. if a mixture of different materials is used together in a ceramic, the sintering temperature is sometimes above the melting point of one minor component β a liquid phase sintering. this results in shorter sintering times compared to solid state sintering. such liquid phase sintering involves in faster diffusion processes and may result in abnormal grain
of tool usage was found in ethiopia within the great rift valley, dating back to 2. 5 million years ago. the earliest methods of stone tool making, known as the oldowan " industry ", date back to at least 2. 3 million years ago. this era of stone tool use is called the paleolithic, or " old stone age ", and spans all of human history up to the development of agriculture approximately 12, 000 years ago. to make a stone tool, a " core " of hard stone with specific flaking properties ( such as flint ) was struck with a hammerstone. this flaking produced sharp edges which could be used as tools, primarily in the form of choppers or scrapers. these tools greatly aided the early humans in their hunter - gatherer lifestyle to perform a variety of tasks including butchering carcasses ( and breaking bones to get at the marrow ) ; chopping wood ; cracking open nuts ; skinning an animal for its hide, and even forming other tools out of softer materials such as bone and wood. the earliest stone tools were irrelevant, being little more than a fractured rock. in the acheulian era, beginning approximately 1. 65 million years ago, methods of working these stones into specific shapes, such as hand axes emerged. this early stone age is described as the lower paleolithic. the middle paleolithic, approximately 300, 000 years ago, saw the introduction of the prepared - core technique, where multiple blades could be rapidly formed from a single core stone. the upper paleolithic, beginning approximately 40, 000 years ago, saw the introduction of pressure flaking, where a wood, bone, or antler punch could be used to shape a stone very finely. the end of the last ice age about 10, 000 years ago is taken as the end point of the upper paleolithic and the beginning of the epipaleolithic / mesolithic. the mesolithic technology included the use of microliths as composite stone tools, along with wood, bone, and antler tools. the later stone age, during which the rudiments of agricultural technology were developed, is called the neolithic period. during this period, polished stone tools were made from a variety of hard rocks such as flint, jade, jadeite, and greenstone, largely by working exposures as quarries, but later the valuable rocks were pursued by tunneling underground, the first steps in mining technology. the polished axes were used for forest clearance and the establishment of crop
there are a variety of bioreactors designed for 3d cell cultures. there are small plastic cylindrical chambers, as well as glass chambers, with regulated internal humidity and moisture specifically engineered for the purpose of growing cells in three dimensions. the bioreactor uses bioactive synthetic materials such as polyethylene terephthalate membranes to surround the spheroid cells in an environment that maintains high levels of nutrients. they are easy to open and close, so that cell spheroids can be removed for testing, yet the chamber is able to maintain 100 % humidity throughout. this humidity is important to achieve maximum cell growth and function. the bioreactor chamber is part of a larger device that rotates to ensure equal cell growth in each direction across three dimensions. quinxell technologies now under quintech life sciences from singapore has developed a bioreactor known as the tisxell biaxial bioreactor which is specially designed for the purpose of tissue engineering. it is the first bioreactor in the world to have a spherical glass chamber with biaxial rotation ; specifically to mimic the rotation of the fetus in the womb ; which provides a conducive environment for the growth of tissues. multiple forms of mechanical stimulation have also been combined into a single bioreactor. using gene expression analysis, one academic study found that applying a combination of cyclic strain and ultrasound stimulation to pre - osteoblast cells in a bioreactor accelerated matrix maturation and differentiation. the technology of this combined stimulation bioreactor could be used to grow bone cells more quickly and effectively in future clinical stem cell therapies. mc2 biotek has also developed a bioreactor known as prototissue that uses gas exchange to maintain high oxygen levels within the cell chamber ; improving upon previous bioreactors, since the higher oxygen levels help the cell grow and undergo normal cell respiration. active areas of research on bioreactors includes increasing production scale and refining the physiological environment, both of which could improve the efficiency and efficacy of bioreactors in research or clinical use. bioreactors are currently used to study, among other things, cell and tissue level therapies, cell and tissue response to specific physiological environment changes, and development of disease and injury. = = = long fiber generation = = = in 2013, a group from the university of tokyo developed cell laden fibers up to a meter in length and on the order of 100 ΞΌm in size. these fibers were created using a microfluidic device that forms a
##tion, and pasteurization in order to become products that can be sold. there are three levels of food processing : primary, secondary, and tertiary. primary food processing involves turning agricultural products into other products that can be turned into food, secondary food processing is the making of food from readily available ingredients, and tertiary food processing is commercial production of ready - to eat or heat - and - serve foods. drying, pickling, salting, and fermenting foods were some of the oldest food processing techniques used to preserve food by preventing yeasts, molds, and bacteria to cause spoiling. methods for preserving food have evolved to meet current standards of food safety but still use the same processes as the past. biochemical engineers also work to improve the nutritional value of food products, such as in golden rice, which was developed to prevent vitamin a deficiency in certain areas where this was an issue. efforts to advance preserving technologies can also ensure lasting retention of nutrients as foods are stored. packaging plays a key role in preserving as well as ensuring the safety of the food by protecting the product from contamination, physical damage, and tampering. packaging can also make it easier to transport and serve food. a common job for biochemical engineers working in the food industry is to design ways to perform all these processes on a large scale in order to meet the demands of the population. responsibilities for this career path include designing and performing experiments, optimizing processes, consulting with groups to develop new technologies, and preparing project plans for equipment and facilities. = = = pharmaceuticals = = = in the pharmaceutical industry, bioprocess engineering plays a crucial role in the large - scale production of biopharmaceuticals, such as monoclonal antibodies, vaccines, and therapeutic proteins. the development and optimization of bioreactors and fermentation systems are essential for the mass production of these products, ensuring consistent quality and high yields. for example, recombinant proteins like insulin and erythropoietin are produced through cell culture systems using genetically modified cells. the bioprocess engineer β s role is to optimize variables like temperature, ph, nutrient availability, and oxygen levels to maximize the efficiency of these systems. the growing field of gene therapy also relies on bioprocessing techniques to produce viral vectors, which are used to deliver therapeutic genes to patients. this involves scaling up processes from laboratory to industrial scale while maintaining safety and regulatory compliance. as the demand for biopharmaceutical products increases, advancements
process of lactic acid fermentation, which produced other preserved foods, such as soy sauce. fermentation was also used in this time period to produce leavened bread. although the process of fermentation was not fully understood until louis pasteur ' s work in 1857, it is still the first use of biotechnology to convert a food source into another form. before the time of charles darwin ' s work and life, animal and plant scientists had already used selective breeding. darwin added to that body of work with his scientific observations about the ability of science to change species. these accounts contributed to darwin ' s theory of natural selection. for thousands of years, humans have used selective breeding to improve the production of crops and livestock to use them for food. in selective breeding, organisms with desirable characteristics are mated to produce offspring with the same characteristics. for example, this technique was used with corn to produce the largest and sweetest crops. in the early twentieth century scientists gained a greater understanding of microbiology and explored ways of manufacturing specific products. in 1917, chaim weizmann first used a pure microbiological culture in an industrial process, that of manufacturing corn starch using clostridium acetobutylicum, to produce acetone, which the united kingdom desperately needed to manufacture explosives during world war i. biotechnology has also led to the development of antibiotics. in 1928, alexander fleming discovered the mold penicillium. his work led to the purification of the antibiotic formed by the mold by howard florey, ernst boris chain and norman heatley β to form what we today know as penicillin. in 1940, penicillin became available for medicinal use to treat bacterial infections in humans. the field of modern biotechnology is generally thought of as having been born in 1971 when paul berg ' s ( stanford ) experiments in gene splicing had early success. herbert w. boyer ( univ. calif. at san francisco ) and stanley n. cohen ( stanford ) significantly advanced the new technology in 1972 by transferring genetic material into a bacterium, such that the imported material would be reproduced. the commercial viability of a biotechnology industry was significantly expanded on june 16, 1980, when the united states supreme court ruled that a genetically modified microorganism could be patented in the case of diamond v. chakrabarty. indian - born ananda chakrabarty, working for general electric, had modified a bacterium ( of the genus pseudomonas ) capable of breaking down crude oil, which he proposed to
molecular diffusion processes give rise to significant changes in the primary microstructural features. this includes the gradual elimination of porosity, which is typically accompanied by a net shrinkage and overall densification of the component. thus, the pores in the object may close up, resulting in a denser product of significantly greater strength and fracture toughness. another major change in the body during the firing or sintering process will be the establishment of the polycrystalline nature of the solid. significant grain growth tends to occur during sintering, with this growth depending on temperature and duration of the sintering process. the growth of grains will result in some form of grain size distribution, which will have a significant impact on the ultimate physical properties of the material. in particular, abnormal grain growth in which certain grains grow very large in a matrix of finer grains will significantly alter the physical and mechanical properties of the obtained ceramic. in the sintered body, grain sizes are a product of the thermal processing parameters as well as the initial particle size, or possibly the sizes of aggregates or particle clusters which arise during the initial stages of processing. the ultimate microstructure ( and thus the physical properties ) of the final product will be limited by and subject to the form of the structural template or precursor which is created in the initial stages of chemical synthesis and physical forming. hence the importance of chemical powder and polymer processing as it pertains to the synthesis of industrial ceramics, glasses and glass - ceramics. there are numerous possible refinements of the sintering process. some of the most common involve pressing the green body to give the densification a head start and reduce the sintering time needed. sometimes organic binders such as polyvinyl alcohol are added to hold the green body together ; these burn out during the firing ( at 200 β 350 Β°c ). sometimes organic lubricants are added during pressing to increase densification. it is common to combine these, and add binders and lubricants to a powder, then press. ( the formulation of these organic chemical additives is an art in itself. this is particularly important in the manufacture of high performance ceramics such as those used by the billions for electronics, in capacitors, inductors, sensors, etc. ) a slurry can be used in place of a powder, and then cast into a desired shape, dried and then sintered. indeed, traditional pottery is done with this type of method, using a plastic mixture worked with the hands.
Question: Which tool should be used to compare the growth of an algal colony growing in a petri dish?
A) a scale
B) a ruler
C) a hand lens
D) a balance scale
|
C) a hand lens
|
Context:
. historically, metallurgy has predominately focused on the production of metals. metal production begins with the processing of ores to extract the metal, and includes the mixture of metals to make alloys. metal alloys are often a blend of at least two different metallic elements. however, non - metallic elements are often added to alloys in order to achieve properties suitable for an application. the study of metal production is subdivided into ferrous metallurgy ( also known as black metallurgy ) and non - ferrous metallurgy, also known as colored metallurgy. ferrous metallurgy involves processes and alloys based on iron, while non - ferrous metallurgy involves processes and alloys based on other metals. the production of ferrous metals accounts for 95 % of world metal production. modern metallurgists work in both emerging and traditional areas as part of an interdisciplinary team alongside material scientists and other engineers. some traditional areas include mineral processing, metal production, heat treatment, failure analysis, and the joining of metals ( including welding, brazing, and soldering ). emerging areas for metallurgists include nanotechnology, superconductors, composites, biomedical materials, electronic materials ( semiconductors ) and surface engineering. = = etymology and pronunciation = = metallurgy derives from the ancient greek ΞΌΞ΅ΟαλλοΟ
ΟΞ³ΞΏΟ, metallourgos, " worker in metal ", from ΞΌΞ΅Οαλλον, metallon, " mine, metal " + Ξ΅ΟΞ³ΞΏΞ½, ergon, " work " the word was originally an alchemist ' s term for the extraction of metals from minerals, the ending - urgy signifying a process, especially manufacturing : it was discussed in this sense in the 1797 encyclopΓ¦dia britannica. in the late 19th century, metallurgy ' s definition was extended to the more general scientific study of metals, alloys, and related processes. in english, the pronunciation is the more common one in the united kingdom. the pronunciation is the more common one in the us and is the first - listed variant in various american dictionaries, including merriam - webster collegiate and american heritage. = = history = = the earliest metal employed by humans appears to be gold, which can be found " native ". small amounts of natural gold, dating to the late paleolithic period, 40, 000 bc, have been found in spanish caves. silver, copper, tin and meteoric iron
applications, where neither weight nor corrosion are a major concern. cast irons, including ductile iron, are also part of the iron - carbon system. iron - manganese - chromium alloys ( hadfield - type steels ) are also used in non - magnetic applications such as directional drilling. other engineering metals include aluminium, chromium, copper, magnesium, nickel, titanium, zinc, and silicon. these metals are most often used as alloys with the noted exception of silicon, which is not a metal. other forms include : stainless steel, particularly austenitic stainless steels, galvanized steel, nickel alloys, titanium alloys, or occasionally copper alloys are used, where resistance to corrosion is important. aluminium alloys and magnesium alloys are commonly used, when a lightweight strong part is required such as in automotive and aerospace applications. copper - nickel alloys ( such as monel ) are used in highly corrosive environments and for non - magnetic applications. nickel - based superalloys like inconel are used in high - temperature applications such as gas turbines, turbochargers, pressure vessels, and heat exchangers. for extremely high temperatures, single crystal alloys are used to minimize creep. in modern electronics, high purity single crystal silicon is essential for metal - oxide - silicon transistors ( mos ) and integrated circuits. = = production = = in production engineering, metallurgy is concerned with the production of metallic components for use in consumer or engineering products. this involves production of alloys, shaping, heat treatment and surface treatment of product. the task of the metallurgist is to achieve balance between material properties, such as cost, weight, strength, toughness, hardness, corrosion, fatigue resistance and performance in temperature extremes. to achieve this goal, the operating environment must be carefully considered. determining the hardness of the metal using the rockwell, vickers, and brinell hardness scales is a commonly used practice that helps better understand the metal ' s elasticity and plasticity for different applications and production processes. in a saltwater environment, most ferrous metals and some non - ferrous alloys corrode quickly. metals exposed to cold or cryogenic conditions may undergo a ductile to brittle transition and lose their toughness, becoming more brittle and prone to cracking. metals under continual cyclic loading can suffer from metal fatigue. metals under constant stress at elevated temperatures can creep. = = = metalworking processes = = = casting β molten metal is poured into a shaped mold. variants of casting include sand casting, investment
the valuable metals into individual constituents. = = metal and its alloys = = much effort has been placed on understanding iron β carbon alloy system, which includes steels and cast irons. plain carbon steels ( those that contain essentially only carbon as an alloying element ) are used in low - cost, high - strength applications, where neither weight nor corrosion are a major concern. cast irons, including ductile iron, are also part of the iron - carbon system. iron - manganese - chromium alloys ( hadfield - type steels ) are also used in non - magnetic applications such as directional drilling. other engineering metals include aluminium, chromium, copper, magnesium, nickel, titanium, zinc, and silicon. these metals are most often used as alloys with the noted exception of silicon, which is not a metal. other forms include : stainless steel, particularly austenitic stainless steels, galvanized steel, nickel alloys, titanium alloys, or occasionally copper alloys are used, where resistance to corrosion is important. aluminium alloys and magnesium alloys are commonly used, when a lightweight strong part is required such as in automotive and aerospace applications. copper - nickel alloys ( such as monel ) are used in highly corrosive environments and for non - magnetic applications. nickel - based superalloys like inconel are used in high - temperature applications such as gas turbines, turbochargers, pressure vessels, and heat exchangers. for extremely high temperatures, single crystal alloys are used to minimize creep. in modern electronics, high purity single crystal silicon is essential for metal - oxide - silicon transistors ( mos ) and integrated circuits. = = production = = in production engineering, metallurgy is concerned with the production of metallic components for use in consumer or engineering products. this involves production of alloys, shaping, heat treatment and surface treatment of product. the task of the metallurgist is to achieve balance between material properties, such as cost, weight, strength, toughness, hardness, corrosion, fatigue resistance and performance in temperature extremes. to achieve this goal, the operating environment must be carefully considered. determining the hardness of the metal using the rockwell, vickers, and brinell hardness scales is a commonly used practice that helps better understand the metal ' s elasticity and plasticity for different applications and production processes. in a saltwater environment, most ferrous metals and some non - ferrous alloys corrode quickly. metals exposed to cold or cryogenic conditions may undergo a ductile to brittle
is further subdivided into two broad categories : chemical metallurgy and physical metallurgy. chemical metallurgy is chiefly concerned with the reduction and oxidation of metals, and the chemical performance of metals. subjects of study in chemical metallurgy include mineral processing, the extraction of metals, thermodynamics, electrochemistry, and chemical degradation ( corrosion ). in contrast, physical metallurgy focuses on the mechanical properties of metals, the physical properties of metals, and the physical performance of metals. topics studied in physical metallurgy include crystallography, material characterization, mechanical metallurgy, phase transformations, and failure mechanisms. historically, metallurgy has predominately focused on the production of metals. metal production begins with the processing of ores to extract the metal, and includes the mixture of metals to make alloys. metal alloys are often a blend of at least two different metallic elements. however, non - metallic elements are often added to alloys in order to achieve properties suitable for an application. the study of metal production is subdivided into ferrous metallurgy ( also known as black metallurgy ) and non - ferrous metallurgy, also known as colored metallurgy. ferrous metallurgy involves processes and alloys based on iron, while non - ferrous metallurgy involves processes and alloys based on other metals. the production of ferrous metals accounts for 95 % of world metal production. modern metallurgists work in both emerging and traditional areas as part of an interdisciplinary team alongside material scientists and other engineers. some traditional areas include mineral processing, metal production, heat treatment, failure analysis, and the joining of metals ( including welding, brazing, and soldering ). emerging areas for metallurgists include nanotechnology, superconductors, composites, biomedical materials, electronic materials ( semiconductors ) and surface engineering. = = etymology and pronunciation = = metallurgy derives from the ancient greek ΞΌΞ΅ΟαλλοΟ
ΟΞ³ΞΏΟ, metallourgos, " worker in metal ", from ΞΌΞ΅Οαλλον, metallon, " mine, metal " + Ξ΅ΟΞ³ΞΏΞ½, ergon, " work " the word was originally an alchemist ' s term for the extraction of metals from minerals, the ending - urgy signifying a process, especially manufacturing : it was discussed in this sense in the 1797 encyclopΓ¦dia britannica. in the late 19th century, metallurgy '
##chemistry, and chemical degradation ( corrosion ). in contrast, physical metallurgy focuses on the mechanical properties of metals, the physical properties of metals, and the physical performance of metals. topics studied in physical metallurgy include crystallography, material characterization, mechanical metallurgy, phase transformations, and failure mechanisms. historically, metallurgy has predominately focused on the production of metals. metal production begins with the processing of ores to extract the metal, and includes the mixture of metals to make alloys. metal alloys are often a blend of at least two different metallic elements. however, non - metallic elements are often added to alloys in order to achieve properties suitable for an application. the study of metal production is subdivided into ferrous metallurgy ( also known as black metallurgy ) and non - ferrous metallurgy, also known as colored metallurgy. ferrous metallurgy involves processes and alloys based on iron, while non - ferrous metallurgy involves processes and alloys based on other metals. the production of ferrous metals accounts for 95 % of world metal production. modern metallurgists work in both emerging and traditional areas as part of an interdisciplinary team alongside material scientists and other engineers. some traditional areas include mineral processing, metal production, heat treatment, failure analysis, and the joining of metals ( including welding, brazing, and soldering ). emerging areas for metallurgists include nanotechnology, superconductors, composites, biomedical materials, electronic materials ( semiconductors ) and surface engineering. = = etymology and pronunciation = = metallurgy derives from the ancient greek ΞΌΞ΅ΟαλλοΟ
ΟΞ³ΞΏΟ, metallourgos, " worker in metal ", from ΞΌΞ΅Οαλλον, metallon, " mine, metal " + Ξ΅ΟΞ³ΞΏΞ½, ergon, " work " the word was originally an alchemist ' s term for the extraction of metals from minerals, the ending - urgy signifying a process, especially manufacturing : it was discussed in this sense in the 1797 encyclopΓ¦dia britannica. in the late 19th century, metallurgy ' s definition was extended to the more general scientific study of metals, alloys, and related processes. in english, the pronunciation is the more common one in the united kingdom. the pronunciation is the more common one in the us and is the first - listed variant in various american dictionaries, including merriam - webster collegiate
is collected and processed to extract valuable metals. ore bodies often contain more than one valuable metal. tailings of a previous process may be used as a feed in another process to extract a secondary product from the original ore. additionally, a concentrate may contain more than one valuable metal. that concentrate would then be processed to separate the valuable metals into individual constituents. = = metal and its alloys = = much effort has been placed on understanding iron β carbon alloy system, which includes steels and cast irons. plain carbon steels ( those that contain essentially only carbon as an alloying element ) are used in low - cost, high - strength applications, where neither weight nor corrosion are a major concern. cast irons, including ductile iron, are also part of the iron - carbon system. iron - manganese - chromium alloys ( hadfield - type steels ) are also used in non - magnetic applications such as directional drilling. other engineering metals include aluminium, chromium, copper, magnesium, nickel, titanium, zinc, and silicon. these metals are most often used as alloys with the noted exception of silicon, which is not a metal. other forms include : stainless steel, particularly austenitic stainless steels, galvanized steel, nickel alloys, titanium alloys, or occasionally copper alloys are used, where resistance to corrosion is important. aluminium alloys and magnesium alloys are commonly used, when a lightweight strong part is required such as in automotive and aerospace applications. copper - nickel alloys ( such as monel ) are used in highly corrosive environments and for non - magnetic applications. nickel - based superalloys like inconel are used in high - temperature applications such as gas turbines, turbochargers, pressure vessels, and heat exchangers. for extremely high temperatures, single crystal alloys are used to minimize creep. in modern electronics, high purity single crystal silicon is essential for metal - oxide - silicon transistors ( mos ) and integrated circuits. = = production = = in production engineering, metallurgy is concerned with the production of metallic components for use in consumer or engineering products. this involves production of alloys, shaping, heat treatment and surface treatment of product. the task of the metallurgist is to achieve balance between material properties, such as cost, weight, strength, toughness, hardness, corrosion, fatigue resistance and performance in temperature extremes. to achieve this goal, the operating environment must be carefully considered. determining the hardness of the metal using the rockwell, vickers, and brinell hardness scales
metallurgy is a domain of materials science and engineering that studies the physical and chemical behavior of metallic elements, their inter - metallic compounds, and their mixtures, which are known as alloys. metallurgy encompasses both the science and the technology of metals, including the production of metals and the engineering of metal components used in products for both consumers and manufacturers. metallurgy is distinct from the craft of metalworking. metalworking relies on metallurgy in a similar manner to how medicine relies on medical science for technical advancement. a specialist practitioner of metallurgy is known as a metallurgist. the science of metallurgy is further subdivided into two broad categories : chemical metallurgy and physical metallurgy. chemical metallurgy is chiefly concerned with the reduction and oxidation of metals, and the chemical performance of metals. subjects of study in chemical metallurgy include mineral processing, the extraction of metals, thermodynamics, electrochemistry, and chemical degradation ( corrosion ). in contrast, physical metallurgy focuses on the mechanical properties of metals, the physical properties of metals, and the physical performance of metals. topics studied in physical metallurgy include crystallography, material characterization, mechanical metallurgy, phase transformations, and failure mechanisms. historically, metallurgy has predominately focused on the production of metals. metal production begins with the processing of ores to extract the metal, and includes the mixture of metals to make alloys. metal alloys are often a blend of at least two different metallic elements. however, non - metallic elements are often added to alloys in order to achieve properties suitable for an application. the study of metal production is subdivided into ferrous metallurgy ( also known as black metallurgy ) and non - ferrous metallurgy, also known as colored metallurgy. ferrous metallurgy involves processes and alloys based on iron, while non - ferrous metallurgy involves processes and alloys based on other metals. the production of ferrous metals accounts for 95 % of world metal production. modern metallurgists work in both emerging and traditional areas as part of an interdisciplinary team alongside material scientists and other engineers. some traditional areas include mineral processing, metal production, heat treatment, failure analysis, and the joining of metals ( including welding, brazing, and soldering ). emerging areas for metallurgists include nanotechnology, superconductors, composites, biomedical materials, electronic materials ( semiconductors ) and surface engineering
joints. = = = metal alloys = = = the alloys of iron ( steel, stainless steel, cast iron, tool steel, alloy steels ) make up the largest proportion of metals today both by quantity and commercial value. iron alloyed with various proportions of carbon gives low, mid and high carbon steels. an iron - carbon alloy is only considered steel if the carbon level is between 0. 01 % and 2. 00 % by weight. for steels, the hardness and tensile strength of the steel is related to the amount of carbon present, with increasing carbon levels also leading to lower ductility and toughness. heat treatment processes such as quenching and tempering can significantly change these properties, however. in contrast, certain metal alloys exhibit unique properties where their size and density remain unchanged across a range of temperatures. cast iron is defined as an iron β carbon alloy with more than 2. 00 %, but less than 6. 67 % carbon. stainless steel is defined as a regular steel alloy with greater than 10 % by weight alloying content of chromium. nickel and molybdenum are typically also added in stainless steels. other significant metallic alloys are those of aluminium, titanium, copper and magnesium. copper alloys have been known for a long time ( since the bronze age ), while the alloys of the other three metals have been relatively recently developed. due to the chemical reactivity of these metals, the electrolytic extraction processes required were only developed relatively recently. the alloys of aluminium, titanium and magnesium are also known and valued for their high strength to weight ratios and, in the case of magnesium, their ability to provide electromagnetic shielding. these materials are ideal for situations where high strength to weight ratios are more important than bulk cost, such as in the aerospace industry and certain automotive engineering applications. = = = semiconductors = = = a semiconductor is a material that has a resistivity between a conductor and insulator. modern day electronics run on semiconductors, and the industry had an estimated us $ 530 billion market in 2021. its electronic properties can be greatly altered through intentionally introducing impurities in a process referred to as doping. semiconductor materials are used to build diodes, transistors, light - emitting diodes ( leds ), and analog and digital electric circuits, among their many uses. semiconductor devices have replaced thermionic devices like vacuum tubes in most applications. semiconductor devices are manufactured both as single discrete devices and as integrated circuits ( ics ), which consist of a number β from a
an alternative explanation of 1 / f - noise in manganites is suggested and discussed
near east, about 3, 500 bc, it was discovered that by combining copper and tin, a superior metal could be made, an alloy called bronze. this represented a major technological shift known as the bronze age. the extraction of iron from its ore into a workable metal is much more difficult than for copper or tin. the process appears to have been invented by the hittites in about 1200 bc, beginning the iron age. the secret of extracting and working iron was a key factor in the success of the philistines. historical developments in ferrous metallurgy can be found in a wide variety of past cultures and civilizations. this includes the ancient and medieval kingdoms and empires of the middle east and near east, ancient iran, ancient egypt, ancient nubia, and anatolia in present - day turkey, ancient nok, carthage, the celts, greeks and romans of ancient europe, medieval europe, ancient and medieval china, ancient and medieval india, ancient and medieval japan, amongst others. a 16th century book by georg agricola, de re metallica, describes the highly developed and complex processes of mining metal ores, metal extraction, and metallurgy of the time. agricola has been described as the " father of metallurgy ". = = extraction = = extractive metallurgy is the practice of removing valuable metals from an ore and refining the extracted raw metals into a purer form. in order to convert a metal oxide or sulphide to a purer metal, the ore must be reduced physically, chemically, or electrolytically. extractive metallurgists are interested in three primary streams : feed, concentrate ( metal oxide / sulphide ) and tailings ( waste ). after mining, large pieces of the ore feed are broken through crushing or grinding in order to obtain particles small enough, where each particle is either mostly valuable or mostly waste. concentrating the particles of value in a form supporting separation enables the desired metal to be removed from waste products. mining may not be necessary, if the ore body and physical environment are conducive to leaching. leaching dissolves minerals in an ore body and results in an enriched solution. the solution is collected and processed to extract valuable metals. ore bodies often contain more than one valuable metal. tailings of a previous process may be used as a feed in another process to extract a secondary product from the original ore. additionally, a concentrate may contain more than one valuable metal. that concentrate would then be processed to separate
Question: Use The Periodic Table of the Elements and the information below to answer the question. Magnesium metal (Mg) is grayish-white in color and reacts actively with water. Fluorine (F2) is a greenish-yellow gas at room temperature and is a member of the halogen family. These two elements react to produce magnesium fluoride (MgF2), a chemical commonly used in making windows and lenses. How do the elements magnesium (Mg) and fluorine (F2) produce the compound magnesium fluoride (MgF2)?
A) by nuclear connection
B) by physically combining
C) by magnetic connection
D) by chemically combining
|
D) by chemically combining
|
Context:
the boron buckyball avoids the high symmetry icosahedral cage structure. the previously reported ih symmetric structure is not an energy minimum in the potential energy surface and exhibits a spontaneous symmetry breaking to yield a puckered cage with a rare th symmetry. the homo - lumo gap is twice as large as the reported value and amounts to 1. 94 ev at b3lyp / 6 - 31g ( d ) level. the valence orbital structure of boron buckyball is identical to the one in the carbon analogue.
angles. stealth aircraft such as the f - 117 use a different arrangement, tilting the tail surfaces to reduce corner reflections formed between them. a more radical method is to omit the tail, as in the b - 2 spirit. the b - 2 ' s clean, low - drag flying wing configuration gives it exceptional range and reduces its radar profile. the flying wing design most closely resembles a so - called infinite flat plate ( as vertical control surfaces dramatically increase rcs ), the perfect stealth shape, as it would have no angles to reflect back radar waves. in addition to altering the tail, stealth design must bury the engines within the wing or fuselage, or in some cases where stealth is applied to an extant aircraft, install baffles in the air intakes, so that the compressor blades are not visible to radar. a stealthy shape must be devoid of complex bumps or protrusions of any kind, meaning that weapons, fuel tanks, and other stores must not be carried externally. any stealthy vehicle becomes un - stealthy when a door or hatch opens. parallel alignment of edges or even surfaces is also often used in stealth designs. the technique involves using a small number of edge orientations in the shape of the structure. for example, on the f - 22a raptor, the leading edges of the wing and the tail planes are set at the same angle. other smaller structures, such as the air intake bypass doors and the air refueling aperture, also use the same angles. the effect of this is to return a narrow radar signal in a very specific direction away from the radar emitter rather than returning a diffuse signal detectable at many angles. the effect is sometimes called " glitter " after the very brief signal seen when the reflected beam passes across a detector. it can be difficult for the radar operator to distinguish between a glitter event and a digital glitch in the processing system. stealth airframes sometimes display distinctive serrations on some exposed edges, such as the engine ports. the yf - 23 has such serrations on the exhaust ports. this is another example in the parallel alignment of features, this time on the external airframe. the shaping requirements detracted greatly from the f - 117 ' s aerodynamic properties. it is inherently unstable, and cannot be flown without a fly - by - wire control system. similarly, coating the cockpit canopy with a thin film transparent conductor ( vapor - deposited gold or indium tin oxide ) helps to reduce the aircraft ' s radar profile, because radar waves would normally enter the cockpit
and reduces its radar profile. the flying wing design most closely resembles a so - called infinite flat plate ( as vertical control surfaces dramatically increase rcs ), the perfect stealth shape, as it would have no angles to reflect back radar waves. in addition to altering the tail, stealth design must bury the engines within the wing or fuselage, or in some cases where stealth is applied to an extant aircraft, install baffles in the air intakes, so that the compressor blades are not visible to radar. a stealthy shape must be devoid of complex bumps or protrusions of any kind, meaning that weapons, fuel tanks, and other stores must not be carried externally. any stealthy vehicle becomes un - stealthy when a door or hatch opens. parallel alignment of edges or even surfaces is also often used in stealth designs. the technique involves using a small number of edge orientations in the shape of the structure. for example, on the f - 22a raptor, the leading edges of the wing and the tail planes are set at the same angle. other smaller structures, such as the air intake bypass doors and the air refueling aperture, also use the same angles. the effect of this is to return a narrow radar signal in a very specific direction away from the radar emitter rather than returning a diffuse signal detectable at many angles. the effect is sometimes called " glitter " after the very brief signal seen when the reflected beam passes across a detector. it can be difficult for the radar operator to distinguish between a glitter event and a digital glitch in the processing system. stealth airframes sometimes display distinctive serrations on some exposed edges, such as the engine ports. the yf - 23 has such serrations on the exhaust ports. this is another example in the parallel alignment of features, this time on the external airframe. the shaping requirements detracted greatly from the f - 117 ' s aerodynamic properties. it is inherently unstable, and cannot be flown without a fly - by - wire control system. similarly, coating the cockpit canopy with a thin film transparent conductor ( vapor - deposited gold or indium tin oxide ) helps to reduce the aircraft ' s radar profile, because radar waves would normally enter the cockpit, reflect off objects ( the inside of a cockpit has a complex shape, with a pilot helmet alone forming a sizeable return ), and possibly return to the radar, but the conductive coating creates a controlled shape that deflects the incoming radar waves away from the radar. the coating is thin enough that it has
or fuselage, or in some cases where stealth is applied to an extant aircraft, install baffles in the air intakes, so that the compressor blades are not visible to radar. a stealthy shape must be devoid of complex bumps or protrusions of any kind, meaning that weapons, fuel tanks, and other stores must not be carried externally. any stealthy vehicle becomes un - stealthy when a door or hatch opens. parallel alignment of edges or even surfaces is also often used in stealth designs. the technique involves using a small number of edge orientations in the shape of the structure. for example, on the f - 22a raptor, the leading edges of the wing and the tail planes are set at the same angle. other smaller structures, such as the air intake bypass doors and the air refueling aperture, also use the same angles. the effect of this is to return a narrow radar signal in a very specific direction away from the radar emitter rather than returning a diffuse signal detectable at many angles. the effect is sometimes called " glitter " after the very brief signal seen when the reflected beam passes across a detector. it can be difficult for the radar operator to distinguish between a glitter event and a digital glitch in the processing system. stealth airframes sometimes display distinctive serrations on some exposed edges, such as the engine ports. the yf - 23 has such serrations on the exhaust ports. this is another example in the parallel alignment of features, this time on the external airframe. the shaping requirements detracted greatly from the f - 117 ' s aerodynamic properties. it is inherently unstable, and cannot be flown without a fly - by - wire control system. similarly, coating the cockpit canopy with a thin film transparent conductor ( vapor - deposited gold or indium tin oxide ) helps to reduce the aircraft ' s radar profile, because radar waves would normally enter the cockpit, reflect off objects ( the inside of a cockpit has a complex shape, with a pilot helmet alone forming a sizeable return ), and possibly return to the radar, but the conductive coating creates a controlled shape that deflects the incoming radar waves away from the radar. the coating is thin enough that it has no adverse effect on pilot vision. = = = = ships = = = = ships have also adopted similar methods. though the earlier american arleigh burke - class destroyers incorporated some signature - reduction features. the norwegian skjold - class corvettes was the first coastal defence and the french la fayette - class frigates the
reflect radar waves back to the emitting radar is with orthogonal metal plates, forming a corner reflector consisting of either a dihedral ( two plates ) or a trihedral ( three orthogonal plates ). this configuration occurs in the tail of a conventional aircraft, where the vertical and horizontal components of the tail are set at right angles. stealth aircraft such as the f - 117 use a different arrangement, tilting the tail surfaces to reduce corner reflections formed between them. a more radical method is to omit the tail, as in the b - 2 spirit. the b - 2 ' s clean, low - drag flying wing configuration gives it exceptional range and reduces its radar profile. the flying wing design most closely resembles a so - called infinite flat plate ( as vertical control surfaces dramatically increase rcs ), the perfect stealth shape, as it would have no angles to reflect back radar waves. in addition to altering the tail, stealth design must bury the engines within the wing or fuselage, or in some cases where stealth is applied to an extant aircraft, install baffles in the air intakes, so that the compressor blades are not visible to radar. a stealthy shape must be devoid of complex bumps or protrusions of any kind, meaning that weapons, fuel tanks, and other stores must not be carried externally. any stealthy vehicle becomes un - stealthy when a door or hatch opens. parallel alignment of edges or even surfaces is also often used in stealth designs. the technique involves using a small number of edge orientations in the shape of the structure. for example, on the f - 22a raptor, the leading edges of the wing and the tail planes are set at the same angle. other smaller structures, such as the air intake bypass doors and the air refueling aperture, also use the same angles. the effect of this is to return a narrow radar signal in a very specific direction away from the radar emitter rather than returning a diffuse signal detectable at many angles. the effect is sometimes called " glitter " after the very brief signal seen when the reflected beam passes across a detector. it can be difficult for the radar operator to distinguish between a glitter event and a digital glitch in the processing system. stealth airframes sometimes display distinctive serrations on some exposed edges, such as the engine ports. the yf - 23 has such serrations on the exhaust ports. this is another example in the parallel alignment of features, this time on the external airframe. the shaping requirements detracted greatly from the f - 117 '
insights from stripe incommensurabilities and antiferromagnetic stability indicate that the magnetic moments of both host cu ^ 2 + ions and cu atoms from electron doping support the thermal hall effect in cuprates, whereas those of o atoms from hole doping oppose it.
elevons, flaps, and flaperons into wings to perform the aerodynamic purpose with the advantages of lower rcs for stealth, via simpler geometries and lower complexity ( mechanically simpler, fewer or no moving parts or surfaces, less maintenance ), and lower mass, cost ( up to 50 % less ), drag ( up to 15 % less during use ), and inertia ( for faster, stronger control response to change vehicle orientation to reduce detection ). two promising approaches are flexible wings, and fluidics. in flexible wings, much or all of a wing surface can change shape in flight to deflect air flow. adaptive compliant wings are a military and commercial effort. the x - 53 active aeroelastic wing was a us air force, boeing, and nasa effort. in fluidics, fluid injection into airflows is being researched for use in aircraft to control direction, in two ways : circulation control and thrust vectoring. in both, larger more complex mechanical parts are replaced by smaller, simpler, lower mass fluidic systems, in which larger forces in fluids are diverted by smaller jets or flows of fluid intermittently, to change the direction of vehicles. mechanical control surfaces that must move cause an important part of aircraft radar cross - section. omitting mechanical control surfaces can reduce radar returns. as of 2023, at least two countries are known to be researching fluidic control. in britain, bae systems has tested two fluidically controlled unmanned aircraft, one starting in 2010 named demon, and another starting in 2017 named magma, with the university of manchester. in the united states, the defense advanced research projects agency ( darpa ) program named control of revolutionary aircraft with novel effectors ( crane ) seeks "... to design, build, and flight test a novel x - plane that incorporates active flow control ( afc ) as a primary design consideration.... in 2023, the aircraft received its official designation as x - 65. " in january 2024, construction began, at boeing subsidiary aurora flight sciences. according to darpa, the aurora x - 65 could be completed and unveiled as soon as early 2025, with the first flight occurring in summer 2025. in circulation control, near the trailing edges of wings, aircraft flight control systems are replaced by slots which emit fluid flows. = = list of stealth aircraft = = f - 117 nighthawk b - 2 spirit f - 22 raptor f - 35 lightning ii j - 20 su - 57 b - 21 raider fc - 31 su -
to that of a flat crack through the plain matrix. the magnitude of the toughening is determined by the mismatch strain caused by thermal contraction incompatibility and the microfracture resistance of the particle / matrix interface. the toughening becomes noticeable with a narrow size distribution of appropriately sized particles, and researchers typically accept that deflection effects in materials with roughly equiaxial grains may increase the fracture toughness by about twice the grain boundary value. the model reveals that the increase in toughness is dependent on particle shape and the volume fraction of the second phase, with the most effective morphology being the rod of high aspect ratio, which can account for a fourfold increase in fracture toughness. the toughening arises primarily from the twist of the crack front between particles, as indicated by deflection profiles. disc - shaped particles and spheres are less effective in toughening. fracture toughness, regardless of morphology, is determined by the twist of the crack front at its most severe configuration, rather than the initial tilt of the crack front. only for disc - shaped particles does the initial tilting of the crack front provide significant toughening ; however, the twist component still overrides the tilt - derived toughening. additional important features of the deflection analysis include the appearance of asymptotic toughening for the three morphologies at volume fractions in excess of 0. 2. it is also noted that a significant influence on the toughening by spherical particles is exerted by the interparticle spacing distribution ; greater toughening is afforded when spheres are nearly contacting such that twist angles approach Ο / 2. these predictions provide the basis for the design of high - toughness two - phase ceramic materials. the ideal second phase, in addition to maintaining chemical compatibility, should be present in amounts of 10 to 20 volume percent. greater amounts may diminish the toughness increase due to overlapping particles. particles with high aspect ratios, especially those with rod - shaped morphologies, are most suitable for maximum toughening. this model is often used to determine the factors that contribute to the increase in fracture toughness in ceramics which is ultimately useful in the development of advanced ceramic materials with improved performance. = = theory of chemical processing = = = = = microstructural uniformity = = = in the processing of fine ceramics, the irregular particle sizes and shapes in a typical powder often lead to non - uniform packing morphologies that result in packing density variations in the powder compact. uncontrolled aggl
accept that deflection effects in materials with roughly equiaxial grains may increase the fracture toughness by about twice the grain boundary value. the model reveals that the increase in toughness is dependent on particle shape and the volume fraction of the second phase, with the most effective morphology being the rod of high aspect ratio, which can account for a fourfold increase in fracture toughness. the toughening arises primarily from the twist of the crack front between particles, as indicated by deflection profiles. disc - shaped particles and spheres are less effective in toughening. fracture toughness, regardless of morphology, is determined by the twist of the crack front at its most severe configuration, rather than the initial tilt of the crack front. only for disc - shaped particles does the initial tilting of the crack front provide significant toughening ; however, the twist component still overrides the tilt - derived toughening. additional important features of the deflection analysis include the appearance of asymptotic toughening for the three morphologies at volume fractions in excess of 0. 2. it is also noted that a significant influence on the toughening by spherical particles is exerted by the interparticle spacing distribution ; greater toughening is afforded when spheres are nearly contacting such that twist angles approach Ο / 2. these predictions provide the basis for the design of high - toughness two - phase ceramic materials. the ideal second phase, in addition to maintaining chemical compatibility, should be present in amounts of 10 to 20 volume percent. greater amounts may diminish the toughness increase due to overlapping particles. particles with high aspect ratios, especially those with rod - shaped morphologies, are most suitable for maximum toughening. this model is often used to determine the factors that contribute to the increase in fracture toughness in ceramics which is ultimately useful in the development of advanced ceramic materials with improved performance. = = theory of chemical processing = = = = = microstructural uniformity = = = in the processing of fine ceramics, the irregular particle sizes and shapes in a typical powder often lead to non - uniform packing morphologies that result in packing density variations in the powder compact. uncontrolled agglomeration of powders due to attractive van der waals forces can also give rise to in microstructural inhomogeneities. differential stresses that develop as a result of non - uniform drying shrinkage are directly related to the rate at which the solvent can be removed, and thus highly dependent upon the
; and the wafer surfaces are sufficiently clean. the most stringent criteria for wafer bonding is usually the direct fusion wafer bonding since even one or more small particulates can render the bonding unsuccessful. in comparison, wafer bonding methods that use intermediary layers are often far more forgiving. both bulk and surface silicon micromachining are used in the industrial production of sensors, ink - jet nozzles, and other devices. but in many cases the distinction between these two has diminished. a new etching technology, deep reactive - ion etching, has made it possible to combine good performance typical of bulk micromachining with comb structures and in - plane operation typical of surface micromachining. while it is common in surface micromachining to have structural layer thickness in the range of 2 ΞΌm, in har silicon micromachining the thickness can be from 10 to 100 ΞΌm. the materials commonly used in har silicon micromachining are thick polycrystalline silicon, known as epi - poly, and bonded silicon - on - insulator ( soi ) wafers although processes for bulk silicon wafer also have been created ( scream ). bonding a second wafer by glass frit bonding, anodic bonding or alloy bonding is used to protect the mems structures. integrated circuits are typically not combined with har silicon micromachining. = = applications = = some common commercial applications of mems include : inkjet printers, which use piezoelectrics or thermal bubble ejection to deposit ink on paper. accelerometers in modern cars for a large number of purposes including airbag deployment and electronic stability control. inertial measurement units ( imus ) : mems accelerometers. mems gyroscopes in remote controlled, or autonomous, helicopters, planes and multirotors ( also known as drones ), used for automatically sensing and balancing flying characteristics of roll, pitch and yaw. mems magnetic field sensor ( magnetometer ) may also be incorporated in such devices to provide directional heading. mems inertial navigation systems ( inss ) of modern cars, airplanes, submarines and other vehicles to detect yaw, pitch, and roll ; for example, the autopilot of an airplane. accelerometers in consumer electronics devices such as game controllers ( nintendo wii ), personal media players / cell phones ( virtually all smartphones, various htc pda models ), augmented
Question: Male barn swallow birds with symmetrically-shaped wings find mates more quickly than males with asymmetrical wing shapes. Which statement best explains the significance of symmetry in the barn swallow population?
A) Symmetrical wings are attractive to females.
B) Symmetrical wings indicate genetic mutations.
C) Symmetrical wings indicate that males are good predators.
D) Symmetrical wings are signs that a bird is very aggressive.
|
A) Symmetrical wings are attractive to females.
|
Context:
the thickness and the density of the material to be measured. the method is used for containers of liquids or of grainy substances thickness gauges : if the material is of constant density, the signal measured by the radiation detector depends on the thickness of the material. this is useful for continuous production, like of paper, rubber, etc. electrostatic control - to avoid the build - up of static electricity in production of paper, plastics, synthetic textiles, etc., a ribbon - shaped source of the alpha emitter 241am can be placed close to the material at the end of the production line. the source ionizes the air to remove electric charges on the material. radioactive tracers - since radioactive isotopes behave, chemically, mostly like the inactive element, the behavior of a certain chemical substance can be followed by tracing the radioactivity. examples : adding a gamma tracer to a gas or liquid in a closed system makes it possible to find a hole in a tube. adding a tracer to the surface of the component of a motor makes it possible to measure wear by measuring the activity of the lubricating oil. oil and gas exploration - nuclear well logging is used to help predict the commercial viability of new or existing wells. the technology involves the use of a neutron or gamma - ray source and a radiation detector which are lowered into boreholes to determine the properties of the surrounding rock such as porosity and lithography. [ 1 ] road construction - nuclear moisture / density gauges are used to determine the density of soils, asphalt, and concrete. typically a cesium - 137 source is used. = = = commercial applications = = = radioluminescence tritium illumination : tritium is used with phosphor in rifle sights to increase nighttime firing accuracy. some runway markers and building exit signs use the same technology, to remain illuminated during blackouts. betavoltaics. smoke detector : an ionization smoke detector includes a tiny mass of radioactive americium - 241, which is a source of alpha radiation. two ionisation chambers are placed next to each other. both contain a small source of 241am that gives rise to a small constant current. one is closed and serves for comparison, the other is open to ambient air ; it has a gridded electrode. when smoke enters the open chamber, the current is disrupted as the smoke particles attach to the charged ions and restore them to a neutral electrical state. this reduces the current in the open chamber. when the current drops below a certain threshold, the
##sphere ( or lithosphere ). earth science can be considered to be a branch of planetary science but with a much older history. = = geology = = geology is broadly the study of earth ' s structure, substance, and processes. geology is largely the study of the lithosphere, or earth ' s surface, including the crust and rocks. it includes the physical characteristics and processes that occur in the lithosphere as well as how they are affected by geothermal energy. it incorporates aspects of chemistry, physics, and biology as elements of geology interact. historical geology is the application of geology to interpret earth history and how it has changed over time. geochemistry studies the chemical components and processes of the earth. geophysics studies the physical properties of the earth. paleontology studies fossilized biological material in the lithosphere. planetary geology studies geoscience as it pertains to extraterrestrial bodies. geomorphology studies the origin of landscapes. structural geology studies the deformation of rocks to produce mountains and lowlands. resource geology studies how energy resources can be obtained from minerals. environmental geology studies how pollution and contaminants affect soil and rock. mineralogy is the study of minerals and includes the study of mineral formation, crystal structure, hazards associated with minerals, and the physical and chemical properties of minerals. petrology is the study of rocks, including the formation and composition of rocks. petrography is a branch of petrology that studies the typology and classification of rocks. = = earth ' s interior = = plate tectonics, mountain ranges, volcanoes, and earthquakes are geological phenomena that can be explained in terms of physical and chemical processes in the earth ' s crust. beneath the earth ' s crust lies the mantle which is heated by the radioactive decay of heavy elements. the mantle is not quite solid and consists of magma which is in a state of semi - perpetual convection. this convection process causes the lithospheric plates to move, albeit slowly. the resulting process is known as plate tectonics. areas of the crust where new crust is created are called divergent boundaries, those where it is brought back into the earth are convergent boundaries and those where plates slide past each other, but no new lithospheric material is created or destroyed, are referred to as transform ( or conservative ) boundaries. earthquakes result from the movement of the lithospheric plates, and they often occur near convergent boundaries where parts of the crust are forced into the earth as
##morphology studies the origin of landscapes. structural geology studies the deformation of rocks to produce mountains and lowlands. resource geology studies how energy resources can be obtained from minerals. environmental geology studies how pollution and contaminants affect soil and rock. mineralogy is the study of minerals and includes the study of mineral formation, crystal structure, hazards associated with minerals, and the physical and chemical properties of minerals. petrology is the study of rocks, including the formation and composition of rocks. petrography is a branch of petrology that studies the typology and classification of rocks. = = earth ' s interior = = plate tectonics, mountain ranges, volcanoes, and earthquakes are geological phenomena that can be explained in terms of physical and chemical processes in the earth ' s crust. beneath the earth ' s crust lies the mantle which is heated by the radioactive decay of heavy elements. the mantle is not quite solid and consists of magma which is in a state of semi - perpetual convection. this convection process causes the lithospheric plates to move, albeit slowly. the resulting process is known as plate tectonics. areas of the crust where new crust is created are called divergent boundaries, those where it is brought back into the earth are convergent boundaries and those where plates slide past each other, but no new lithospheric material is created or destroyed, are referred to as transform ( or conservative ) boundaries. earthquakes result from the movement of the lithospheric plates, and they often occur near convergent boundaries where parts of the crust are forced into the earth as part of subduction. plate tectonics might be thought of as the process by which the earth is resurfaced. as the result of seafloor spreading, new crust and lithosphere is created by the flow of magma from the mantle to the near surface, through fissures, where it cools and solidifies. through subduction, oceanic crust and lithosphere vehemently returns to the convecting mantle. volcanoes result primarily from the melting of subducted crust material. crust material that is forced into the asthenosphere melts, and some portion of the melted material becomes light enough to rise to the surface β giving birth to volcanoes. = = atmospheric science = = atmospheric science initially developed in the late - 19th century as a means to forecast the weather through meteorology, the study of weather. atmospheric chemistry was developed in the 20th century to measure air pollution and expanded in the 1970s in response to
, crystal structure, hazards associated with minerals, and the physical and chemical properties of minerals. petrology is the study of rocks, including the formation and composition of rocks. petrography is a branch of petrology that studies the typology and classification of rocks. = = earth ' s interior = = plate tectonics, mountain ranges, volcanoes, and earthquakes are geological phenomena that can be explained in terms of physical and chemical processes in the earth ' s crust. beneath the earth ' s crust lies the mantle which is heated by the radioactive decay of heavy elements. the mantle is not quite solid and consists of magma which is in a state of semi - perpetual convection. this convection process causes the lithospheric plates to move, albeit slowly. the resulting process is known as plate tectonics. areas of the crust where new crust is created are called divergent boundaries, those where it is brought back into the earth are convergent boundaries and those where plates slide past each other, but no new lithospheric material is created or destroyed, are referred to as transform ( or conservative ) boundaries. earthquakes result from the movement of the lithospheric plates, and they often occur near convergent boundaries where parts of the crust are forced into the earth as part of subduction. plate tectonics might be thought of as the process by which the earth is resurfaced. as the result of seafloor spreading, new crust and lithosphere is created by the flow of magma from the mantle to the near surface, through fissures, where it cools and solidifies. through subduction, oceanic crust and lithosphere vehemently returns to the convecting mantle. volcanoes result primarily from the melting of subducted crust material. crust material that is forced into the asthenosphere melts, and some portion of the melted material becomes light enough to rise to the surface β giving birth to volcanoes. = = atmospheric science = = atmospheric science initially developed in the late - 19th century as a means to forecast the weather through meteorology, the study of weather. atmospheric chemistry was developed in the 20th century to measure air pollution and expanded in the 1970s in response to acid rain. climatology studies the climate and climate change. the troposphere, stratosphere, mesosphere, thermosphere, and exosphere are the five layers which make up earth ' s atmosphere. 75 % of the mass in the atmosphere is located within the troposphere, the lowest
of measuring methods. x - rays and gamma rays are used in industrial radiography to make images of the inside of solid products, as a means of nondestructive testing and inspection. the piece to be radiographed is placed between the source and a photographic film in a cassette. after a certain exposure time, the film is developed and it shows any internal defects of the material. gauges - gauges use the exponential absorption law of gamma rays level indicators : source and detector are placed at opposite sides of a container, indicating the presence or absence of material in the horizontal radiation path. beta or gamma sources are used, depending on the thickness and the density of the material to be measured. the method is used for containers of liquids or of grainy substances thickness gauges : if the material is of constant density, the signal measured by the radiation detector depends on the thickness of the material. this is useful for continuous production, like of paper, rubber, etc. electrostatic control - to avoid the build - up of static electricity in production of paper, plastics, synthetic textiles, etc., a ribbon - shaped source of the alpha emitter 241am can be placed close to the material at the end of the production line. the source ionizes the air to remove electric charges on the material. radioactive tracers - since radioactive isotopes behave, chemically, mostly like the inactive element, the behavior of a certain chemical substance can be followed by tracing the radioactivity. examples : adding a gamma tracer to a gas or liquid in a closed system makes it possible to find a hole in a tube. adding a tracer to the surface of the component of a motor makes it possible to measure wear by measuring the activity of the lubricating oil. oil and gas exploration - nuclear well logging is used to help predict the commercial viability of new or existing wells. the technology involves the use of a neutron or gamma - ray source and a radiation detector which are lowered into boreholes to determine the properties of the surrounding rock such as porosity and lithography. [ 1 ] road construction - nuclear moisture / density gauges are used to determine the density of soils, asphalt, and concrete. typically a cesium - 137 source is used. = = = commercial applications = = = radioluminescence tritium illumination : tritium is used with phosphor in rifle sights to increase nighttime firing accuracy. some runway markers and building exit signs use the same technology, to remain illuminated during blackouts. betavoltaics
classes according to pore size : the form and shape of the membrane pores are highly dependent on the manufacturing process and are often difficult to specify. therefore, for characterization, test filtrations are carried out and the pore diameter refers to the diameter of the smallest particles which could not pass through the membrane. the rejection can be determined in various ways and provides an indirect measurement of the pore size. one possibility is the filtration of macromolecules ( often dextran, polyethylene glycol or albumin ), another is measurement of the cut - off by gel permeation chromatography. these methods are used mainly to measure membranes for ultrafiltration applications. another testing method is the filtration of particles with defined size and their measurement with a particle sizer or by laser induced breakdown spectroscopy ( libs ). a vivid characterization is to measure the rejection of dextran blue or other colored molecules. the retention of bacteriophage and bacteria, the so - called " bacteria challenge test ", can also provide information about the pore size. to determine the pore diameter, physical methods such as porosimeter ( mercury, liquid - liquid porosimeter and bubble point test ) are also used, but a certain form of the pores ( such as cylindrical or concatenated spherical holes ) is assumed. such methods are used for membranes whose pore geometry does not match the ideal, and we get " nominal " pore diameter, which characterizes the membrane, but does not necessarily reflect its actual filtration behavior and selectivity. the selectivity is highly dependent on the separation process, the composition of the membrane and its electrochemical properties in addition to the pore size. with high selectivity, isotopes can be enriched ( uranium enrichment ) in nuclear engineering or industrial gases like nitrogen can be recovered ( gas separation ). ideally, even racemics can be enriched with a suitable membrane. when choosing membranes selectivity has priority over a high permeability, as low flows can easily be offset by increasing the filter surface with a modular structure. in gas phase filtration different deposition mechanisms are operative, so that particles having sizes below the pore size of the membrane can be retained as well. = = membrane classification = = bio - membrane is classified in two categories, synthetic membrane and natural membrane. synthetic membranes further classified in organic and inorganic membranes. organic membrane sub classified polymeric membranes and inorganic membrane sub classified ceramic polymers. = = synthesis of biomass membrane
consisting of several distinct layers, often referred to as spheres : the lithosphere, the hydrosphere, the atmosphere, and the biosphere, this concept of spheres is a useful tool for understanding the earth ' s surface and its various processes these correspond to rocks, water, air and life. also included by some are the cryosphere ( corresponding to ice ) as a distinct portion of the hydrosphere and the pedosphere ( corresponding to soil ) as an active and intermixed sphere. the following fields of science are generally categorized within the earth sciences : geology describes the rocky parts of the earth ' s crust ( or lithosphere ) and its historic development. major subdisciplines are mineralogy and petrology, geomorphology, paleontology, stratigraphy, structural geology, engineering geology, and sedimentology. physical geography focuses on geography as an earth science. physical geography is the study of earth ' s seasons, climate, atmosphere, soil, streams, landforms, and oceans. physical geography can be divided into several branches or related fields, as follows : geomorphology, biogeography, environmental geography, palaeogeography, climatology, meteorology, coastal geography, hydrology, ecology, glaciology. geophysics and geodesy investigate the shape of the earth, its reaction to forces and its magnetic and gravity fields. geophysicists explore the earth ' s core and mantle as well as the tectonic and seismic activity of the lithosphere. geophysics is commonly used to supplement the work of geologists in developing a comprehensive understanding of crustal geology, particularly in mineral and petroleum exploration. seismologists use geophysics to understand plate tectonic movement, as well as predict seismic activity. geochemistry studies the processes that control the abundance, composition, and distribution of chemical compounds and isotopes in geologic environments. geochemists use the tools and principles of chemistry to study the earth ' s composition, structure, processes, and other physical aspects. major subdisciplines are aqueous geochemistry, cosmochemistry, isotope geochemistry and biogeochemistry. soil science covers the outermost layer of the earth ' s crust that is subject to soil formation processes ( or pedosphere ). major subdivisions in this field of study include edaphology and pedology. ecology covers the interactions between organisms and their environment. this field of study differentiates the study of earth
##thic, or " old stone age ", and spans all of human history up to the development of agriculture approximately 12, 000 years ago. to make a stone tool, a " core " of hard stone with specific flaking properties ( such as flint ) was struck with a hammerstone. this flaking produced sharp edges which could be used as tools, primarily in the form of choppers or scrapers. these tools greatly aided the early humans in their hunter - gatherer lifestyle to perform a variety of tasks including butchering carcasses ( and breaking bones to get at the marrow ) ; chopping wood ; cracking open nuts ; skinning an animal for its hide, and even forming other tools out of softer materials such as bone and wood. the earliest stone tools were irrelevant, being little more than a fractured rock. in the acheulian era, beginning approximately 1. 65 million years ago, methods of working these stones into specific shapes, such as hand axes emerged. this early stone age is described as the lower paleolithic. the middle paleolithic, approximately 300, 000 years ago, saw the introduction of the prepared - core technique, where multiple blades could be rapidly formed from a single core stone. the upper paleolithic, beginning approximately 40, 000 years ago, saw the introduction of pressure flaking, where a wood, bone, or antler punch could be used to shape a stone very finely. the end of the last ice age about 10, 000 years ago is taken as the end point of the upper paleolithic and the beginning of the epipaleolithic / mesolithic. the mesolithic technology included the use of microliths as composite stone tools, along with wood, bone, and antler tools. the later stone age, during which the rudiments of agricultural technology were developed, is called the neolithic period. during this period, polished stone tools were made from a variety of hard rocks such as flint, jade, jadeite, and greenstone, largely by working exposures as quarries, but later the valuable rocks were pursued by tunneling underground, the first steps in mining technology. the polished axes were used for forest clearance and the establishment of crop farming and were so effective as to remain in use when bronze and iron appeared. these stone axes were used alongside a continued use of stone tools such as a range of projectiles, knives, and scrapers, as well as tools, made from organic materials such as wood, bone, and antler. stone age cultures
has rest mass and volume ( it takes up space ) and is made up of particles. the particles that make up matter have rest mass as well β not all particles have rest mass, such as the photon. matter can be a pure chemical substance or a mixture of substances. = = = = atom = = = = the atom is the basic unit of chemistry. it consists of a dense core called the atomic nucleus surrounded by a space occupied by an electron cloud. the nucleus is made up of positively charged protons and uncharged neutrons ( together called nucleons ), while the electron cloud consists of negatively charged electrons which orbit the nucleus. in a neutral atom, the negatively charged electrons balance out the positive charge of the protons. the nucleus is dense ; the mass of a nucleon is approximately 1, 836 times that of an electron, yet the radius of an atom is about 10, 000 times that of its nucleus. the atom is also the smallest entity that can be envisaged to retain the chemical properties of the element, such as electronegativity, ionization potential, preferred oxidation state ( s ), coordination number, and preferred types of bonds to form ( e. g., metallic, ionic, covalent ). = = = = element = = = = a chemical element is a pure substance which is composed of a single type of atom, characterized by its particular number of protons in the nuclei of its atoms, known as the atomic number and represented by the symbol z. the mass number is the sum of the number of protons and neutrons in a nucleus. although all the nuclei of all atoms belonging to one element will have the same atomic number, they may not necessarily have the same mass number ; atoms of an element which have different mass numbers are known as isotopes. for example, all atoms with 6 protons in their nuclei are atoms of the chemical element carbon, but atoms of carbon may have mass numbers of 12 or 13. the standard presentation of the chemical elements is in the periodic table, which orders elements by atomic number. the periodic table is arranged in groups, or columns, and periods, or rows. the periodic table is useful in identifying periodic trends. = = = = compound = = = = a compound is a pure chemical substance composed of more than one element. the properties of a compound bear little similarity to those of its elements. the standard nomenclature of compounds is set by the international union of pure and applied chemistry ( iupac ). organic compounds are named
of tool usage was found in ethiopia within the great rift valley, dating back to 2. 5 million years ago. the earliest methods of stone tool making, known as the oldowan " industry ", date back to at least 2. 3 million years ago. this era of stone tool use is called the paleolithic, or " old stone age ", and spans all of human history up to the development of agriculture approximately 12, 000 years ago. to make a stone tool, a " core " of hard stone with specific flaking properties ( such as flint ) was struck with a hammerstone. this flaking produced sharp edges which could be used as tools, primarily in the form of choppers or scrapers. these tools greatly aided the early humans in their hunter - gatherer lifestyle to perform a variety of tasks including butchering carcasses ( and breaking bones to get at the marrow ) ; chopping wood ; cracking open nuts ; skinning an animal for its hide, and even forming other tools out of softer materials such as bone and wood. the earliest stone tools were irrelevant, being little more than a fractured rock. in the acheulian era, beginning approximately 1. 65 million years ago, methods of working these stones into specific shapes, such as hand axes emerged. this early stone age is described as the lower paleolithic. the middle paleolithic, approximately 300, 000 years ago, saw the introduction of the prepared - core technique, where multiple blades could be rapidly formed from a single core stone. the upper paleolithic, beginning approximately 40, 000 years ago, saw the introduction of pressure flaking, where a wood, bone, or antler punch could be used to shape a stone very finely. the end of the last ice age about 10, 000 years ago is taken as the end point of the upper paleolithic and the beginning of the epipaleolithic / mesolithic. the mesolithic technology included the use of microliths as composite stone tools, along with wood, bone, and antler tools. the later stone age, during which the rudiments of agricultural technology were developed, is called the neolithic period. during this period, polished stone tools were made from a variety of hard rocks such as flint, jade, jadeite, and greenstone, largely by working exposures as quarries, but later the valuable rocks were pursued by tunneling underground, the first steps in mining technology. the polished axes were used for forest clearance and the establishment of crop
Question: Which of the following would be most useful for calculating the density of a rock sample?
A) microscope and balance
B) graduated cylinder and balance
C) microscope and graduated cylinder
D) beaker and graduated cylinder
|
B) graduated cylinder and balance
|
Context:
outer satellites of the planets have distant, eccentric orbits that can be highly inclined or even retrograde relative to the equatorial planes of their planets. these irregular orbits cannot have formed by circumplanetary accretion and are likely products of early capture from heliocentric orbit. the irregular satellites may be the only small bodies remaining which are still relatively near their formation locations within the giant planet region. the study of the irregular satellites provides a unique window on processes operating in the young solar system and allows us to probe possible planet formation mechanisms and the composition of the solar nebula between the rocky objects in the main asteroid belt and the very volatile rich objects in the kuiper belt. the gas and ice giant planets all appear to have very similar irregular satellite systems irrespective of their mass or formation timescales and mechanisms. water ice has been detected on some of the outer satellites of saturn and neptune whereas none has been observed on jupiter ' s outer satellites.
in mathematics, a reflection ( also spelled reflexion ) is a mapping from a euclidean space to itself that is an isometry with a hyperplane as the set of fixed points ; this set is called the axis ( in dimension 2 ) or plane ( in dimension 3 ) of reflection. the image of a figure by a reflection is its mirror image in the axis or plane of reflection. for example the mirror image of the small latin letter p for a reflection with respect to a vertical axis ( a vertical reflection ) would look like q. its image by reflection in a horizontal axis ( a horizontal reflection ) would look like b. a reflection is an involution : when applied twice in succession, every point returns to its original location, and every geometrical object is restored to its original state. the term reflection is sometimes used for a larger class of mappings from a euclidean space to itself, namely the non - identity isometries that are involutions. the set of fixed points ( the " mirror " ) of such an isometry is an affine subspace, but is possibly smaller than a hyperplane. for instance a reflection through a point is an involutive isometry with just one fixed point ; the image of the letter p under it would look like a d. this operation is also known as a central inversion ( coxeter 1969, Β§ 7. 2 ), and exhibits euclidean space as a symmetric space. in a euclidean vector space, the reflection in the point situated at the origin is the same as vector negation. other examples include reflections in a line in three - dimensional space. typically, however, unqualified use of the term " reflection " means reflection in a hyperplane. some mathematicians use " flip " as a synonym for " reflection ". = = construction = = in a plane ( or, respectively, 3 - dimensional ) geometry, to find the reflection of a point drop a perpendicular from the point to the line ( plane ) used for reflection, and extend it the same distance on the other side. to find the reflection of a figure, reflect each point in the figure. to reflect point p through the line ab using compass and straightedge, proceed as follows ( see figure ) : step 1 ( red ) : construct a circle with center at p and some fixed radius r to create points a β² and b β² on the line ab, which will be equidistant from p. step 2 ( green ) : construct circles centered at a β² and b β² having radius r
the gas giant planets in the solar system have a retinue of icy moons, and we expect giant exoplanets to have similar satellite systems. if a jupiter - like planet were to migrate toward its parent star the icy moons orbiting it would evaporate, creating atmospheres and possible habitable surface oceans. here, we examine how long the surface ice and possible oceans would last before being hydrodynamically lost to space. the hydrodynamic loss rate from the moons is determined, in large part, by the stellar flux available for absorption, which increases as the giant planet and icy moons migrate closer to the star. at some planet - star distance the stellar flux incident on the icy moons becomes so great that they enter a runaway greenhouse state. this runaway greenhouse state rapidly transfers all available surface water to the atmosphere as vapor, where it is easily lost from the small moons. however, for icy moons of ganymede ' s size around a sun - like star we found that surface water ( either ice or liquid ) can persist indefinitely outside the runaway greenhouse orbital distance. in contrast, the surface water on smaller moons of europa ' s size will only persist on timescales greater than 1 gyr at distances ranging 1. 49 to 0. 74 au around a sun - like star for bond albedos of 0. 2 and 0. 8, where the lower albedo becomes relevant if ice melts. consequently, small moons can lose their icy shells, which would create a torus of h atoms around their host planet that might be detectable in future observations.
also launched missions to mercury in 2004, with the messenger probe demonstrating as the first use of a solar sail. nasa also launched probes to the outer solar system starting in the 1960s. pioneer 10 was the first probe to the outer planets, flying by jupiter, while pioneer 11 provided the first close up view of the planet. both probes became the first objects to leave the solar system. the voyager program launched in 1977, conducting flybys of jupiter and saturn, neptune, and uranus on a trajectory to leave the solar system. the galileo spacecraft, deployed from the space shuttle flight sts - 34, was the first spacecraft to orbit jupiter, discovering evidence of subsurface oceans on the europa and observed that the moon may hold ice or liquid water. a joint nasa - european space agency - italian space agency mission, cassini β huygens, was sent to saturn ' s moon titan, which, along with mars and europa, are the only celestial bodies in the solar system suspected of being capable of harboring life. cassini discovered three new moons of saturn and the huygens probe entered titan ' s atmosphere. the mission discovered evidence of liquid hydrocarbon lakes on titan and subsurface water oceans on the moon of enceladus, which could harbor life. finally launched in 2006, the new horizons mission was the first spacecraft to visit pluto and the kuiper belt. beyond interplanetary probes, nasa has launched many space telescopes. launched in the 1960s, the orbiting astronomical observatory were nasa ' s first orbital telescopes, providing ultraviolet, gamma - ray, x - ray, and infrared observations. nasa launched the orbiting geophysical observatory in the 1960s and 1970s to look down at earth and observe its interactions with the sun. the uhuru satellite was the first dedicated x - ray telescope, mapping 85 % of the sky and discovering a large number of black holes. launched in the 1990s and early 2000s, the great observatories program are among nasa ' s most powerful telescopes. the hubble space telescope was launched in 1990 on sts - 31 from the discovery and could view galaxies 15 billion light years away. a major defect in the telescope ' s mirror could have crippled the program, had nasa not used computer enhancement to compensate for the imperfection and launched five space shuttle servicing flights to replace the damaged components. the compton gamma ray observatory was launched from the atlantis on sts - 37 in 1991, discovering a possible source of antimatter at the center of the milky way and observing that the majority of gamma - ray bursts
three major planets, venus, earth, and mercury formed out of the solar nebula. a fourth planetesimal, theia, also formed near earth where it collided in a giant impact, rebounding as the planet mars. during this impact earth lost $ { \ approx } 4 $ \ % of its crust and mantle that is now is found on mars and the moon. at the antipode of the giant impact, $ \ approx $ 60 \ % of earth ' s crust, atmosphere, and a large amount of mantle were ejected into space forming the moon. the lost crust never reformed and became the earth ' s ocean basins. the theia impact site corresponds to indian ocean gravitational anomaly on earth and the hellas basin on mars. the dynamics of the giant impact are consistent with the rotational rates and axial tilts of both earth and mars. the giant impact removed sufficient co $ _ 2 $ from earth ' s atmosphere to avoid a runaway greenhouse effect, initiated plate tectonics, and gave life time to form near geothermal vents at the continental margins. mercury formed near venus where on a close approach it was slingshot into the sun ' s convective zone losing 94 \ % of its mass, much of which remains there today. black carbon, from co $ _ 2 $ decomposed by the intense heat, is still found on the surface of mercury. arriving at 616 km / s, mercury dramatically altered the sun ' s rotational energy, explaining both its anomalously slow rotation rate and axial tilt. these results are quantitatively supported by mass balances, the current locations of the terrestrial planets, and the orientations of their major orbital axes.
armed with an astrolabe and kepler ' s laws one can arrive at accurate estimates of the orbits of planets.
light and cold extrasolar planets such as ogle 2005 - blg - 390lb, a 5. 5 earth - mass planet detected via microlensing, could be frequent in the galaxy according to some preliminary results from microlensing experiments. these planets can be frozen rocky - or ocean - planets, situated beyond the snow line and, therefore, beyond the habitable zone of their system. they can nonetheless host a layer of liquid water, heated by radiogenic energy, underneath an ice shell surface for billions of years, before freezing completely. these results suggest that oceans under ice, like those suspected to be present on icy moons in the solar system, could be a common feature of cold low - mass extrasolar planets.
are more expensive than cell phones ; but their advantage is that, unlike a cell phone which is limited to areas covered by cell towers, satphones can be used over most or all of the geographical area of the earth. in order for the phone to communicate with a satellite using a small omnidirectional antenna, first - generation systems use satellites in low earth orbit, about 400 β 700 miles ( 640 β 1, 100 km ) above the surface. with an orbital period of about 100 minutes, a satellite can only be in view of a phone for about 4 β 15 minutes, so the call is " handed off " to another satellite when one passes beyond the local horizon. therefore, large numbers of satellites, about 40 to 70, are required to ensure that at least one satellite is in view continuously from each point on earth. other satphone systems use satellites in geostationary orbit in which only a few satellites are needed, but these cannot be used at high latitudes because of terrestrial interference. cordless phone β a landline telephone in which the handset is portable and communicates with the rest of the phone by a short - range full duplex radio link, instead of being attached by a cord. both the handset and the base station have low - power radio transceivers that handle the short - range bidirectional radio link. as of 2022, cordless phones in most nations use the dect transmission standard. land mobile radio system β short - range mobile or portable half - duplex radio transceivers operating in the vhf or uhf band that can be used without a license. they are often installed in vehicles, with the mobile units communicating with a dispatcher at a fixed base station. special systems with reserved frequencies are used by first responder services ; police, fire, ambulance, and emergency services, and other government services. other systems are made for use by commercial firms such as taxi and delivery services. vhf systems use channels in the range 30 β 50 mhz and 150 β 172 mhz. uhf systems use the 450 β 470 mhz band and in some areas the 470 β 512 mhz range. in general, vhf systems have a longer range than uhf but require longer antennas. am or fm modulation is mainly used, but digital systems such as dmr are being introduced. the radiated power is typically limited to 4 watts. these systems have a fairly limited range, usually 3 to 20 miles ( 4. 8 to 32 km ) depending on terrain. repeaters installed on tall buildings, hills,
a 4mj planet with a 15. 8day orbital period has been detected from very precise radial velocity measurements with the coralie echelle spectrograph. a second remote and more massive companion has also been detected. all the planetary companions so far detected in orbit closer than 0. 08 au have a parent star with a statistically higher metal content compared to the metallicity distribution of other stars with planets. different processes occuring during their formation may provide a possible explanation for this observation.
planetary systems can evolve dynamically even after the full growth of the planets themselves. there is actually circumstantial evidence that most planetary systems become unstable after the disappearance of gas from the protoplanetary disk. these instabilities can be due to the original system being too crowded and too closely packed or to external perturbations such as tides, planetesimal scattering, or torques from distant stellar companions. the solar system was not exceptional in this sense. in its inner part, a crowded system of planetary embryos became unstable, leading to a series of mutual impacts that built the terrestrial planets on a timescale of ~ 100 my. in its outer part, the giant planets became temporarily unstable and their orbital configuration expanded under the effect of mutual encounters. a planet might have been ejected in this phase. thus, the orbital distributions of planetary systems that we observe today, both solar and extrasolar ones, can be different from the those emerging from the formation process and it is important to consider possible long - term evolutionary effects to connect the two.
Question: Which object in our solar system reflects light and is a satellite that orbits around one planet?
A) Earth
B) Mercury
C) the Sun
D) the Moon
|
D) the Moon
|
Context:
which applies a forces that results in fracturing ), and impact ( which employs a milling medium or the particles themselves to cause fracturing ). attrition milling equipment includes the wet scrubber ( also called the planetary mill or wet attrition mill ), which has paddles in water creating vortexes in which the material collides and break up. compression mills include the jaw crusher, roller crusher and cone crusher. impact mills include the ball mill, which has media that tumble and fracture the material, or the resonantacoustic mixer. shaft impactors cause particle - to particle attrition and compression. batching is the process of weighing the oxides according to recipes, and preparing them for mixing and drying. mixing occurs after batching and is performed with various machines, such as dry mixing ribbon mixers ( a type of cement mixer ), resonantacoustic mixers, mueller mixers, and pug mills. wet mixing generally involves the same equipment. forming is making the mixed material into shapes, ranging from toilet bowls to spark plug insulators. forming can involve : ( 1 ) extrusion, such as extruding " slugs " to make bricks, ( 2 ) pressing to make shaped parts, ( 3 ) slip casting, as in making toilet bowls, wash basins and ornamentals like ceramic statues. forming produces a " green " part, ready for drying. green parts are soft, pliable, and over time will lose shape. handling the green product will change its shape. for example, a green brick can be " squeezed ", and after squeezing it will stay that way. drying is removing the water or binder from the formed material. spray drying is widely used to prepare powder for pressing operations. other dryers are tunnel dryers and periodic dryers. controlled heat is applied in this two - stage process. first, heat removes water. this step needs careful control, as rapid heating causes cracks and surface defects. the dried part is smaller than the green part, and is brittle, necessitating careful handling, since a small impact will cause crumbling and breaking. sintering is where the dried parts pass through a controlled heating process, and the oxides are chemically changed to cause bonding and densification. the fired part will be smaller than the dried part. = = forming methods = = ceramic forming techniques include throwing, slipcasting, tape casting, freeze - casting, injection molding, dry pressing, isostatic pressing, hot isostatic pressing
##olithic / mesolithic. the mesolithic technology included the use of microliths as composite stone tools, along with wood, bone, and antler tools. the later stone age, during which the rudiments of agricultural technology were developed, is called the neolithic period. during this period, polished stone tools were made from a variety of hard rocks such as flint, jade, jadeite, and greenstone, largely by working exposures as quarries, but later the valuable rocks were pursued by tunneling underground, the first steps in mining technology. the polished axes were used for forest clearance and the establishment of crop farming and were so effective as to remain in use when bronze and iron appeared. these stone axes were used alongside a continued use of stone tools such as a range of projectiles, knives, and scrapers, as well as tools, made from organic materials such as wood, bone, and antler. stone age cultures developed music and engaged in organized warfare. stone age humans developed ocean - worthy outrigger canoe technology, leading to migration across the malay archipelago, across the indian ocean to madagascar and also across the pacific ocean, which required knowledge of the ocean currents, weather patterns, sailing, and celestial navigation. although paleolithic cultures left no written records, the shift from nomadic life to settlement and agriculture can be inferred from a range of archaeological evidence. such evidence includes ancient tools, cave paintings, and other prehistoric art, such as the venus of willendorf. human remains also provide direct evidence, both through the examination of bones, and the study of mummies. scientists and historians have been able to form significant inferences about the lifestyle and culture of various prehistoric peoples, and especially their technology. = = = ancient = = = = = = = copper and bronze ages = = = = metallic copper occurs on the surface of weathered copper ore deposits and copper was used before copper smelting was known. copper smelting is believed to have originated when the technology of pottery kilns allowed sufficiently high temperatures. the concentration of various elements such as arsenic increase with depth in copper ore deposits and smelting of these ores yields arsenical bronze, which can be sufficiently work hardened to be suitable for making tools. bronze is an alloy of copper with tin ; the latter being found in relatively few deposits globally caused a long time to elapse before true tin bronze became widespread. ( see : tin sources and trade in ancient times ) bronze was a major advancement over stone as a material for
wrought, which itself is the original past passive participle of the word work, now superseded by the weak verb forms worker and worked respectively. ) blacksmithing and the various related smithing and metal - crafts. folk music played on acoustic instruments. mathematics ( particularly, pure mathematics ) organic farming and animal husbandry ( i. e. ; agriculture as practiced by all american farmers prior to world war ii ). milling in the sense of operating hand - constructed equipment with the intent to either grind grain, or the reduction of timber to lumber as practiced in a saw - mill. fulling, felting, drop spindle spinning, hand knitting, crochet, & similar textile preparation. the production of charcoal by the collier, for use in home heating, foundry operations, smelting, the various smithing trades, and for brushing ones teeth as in colonial america. glass - blowing. various subskills of food preservation : smoking salting pickling drying note : home canning is a counter example of a low technology since some of the supplies needed to pursue this skill rely on a global trade network and an existing manufacturing infrastructure. the production of various alcoholic beverages : wine : poorly preserved fruit juice. beer : a way to preserve the calories of grain products from decay. whiskey : an improved ( distilled ) form of beer. flint - knapping masonry as used in castles, cathedrals, and root cellars. = = = domestic or consumer = = = ( non exhaustive ) list of low - tech in a westerner ' s everyday life : getting around by bike, and repairing it with second - hand materials using a cargo bike to carry loads ( rather than a gasoline vehicle ) drying clothes on a clothesline or on a drying rack washing clothes by hand, or in a human - powered washing machine cooling one ' s home with a fan or an air expander ( rather than electrical appliances such as air conditioners ) using a bell as door bell a cellar, " desert fridge ", or icebox ( rather than a fridge or freezer ) long - distance travel by sailing boat ( rather than by plane ) a wicker bag or a tote bag ( rather than a plastic bag ) to carry things swedish lighter ( rather than disposable lighter or matches ) a hand drill, instead of an electric one lighting with sunlight or candles hemp textiles to water plants with drip irrigation paper sheets for note - taking to clean with a broom ( rather than a vacuum cleaner ) to find one ' s way with map
superdielectric behavior was observed in pastes made of high surface area alumina filled to the level of incipient wetness with water containing dissolved sodium chloride ( table salt ). in some cases the dielectric constants were greater than 10 ^ 10.
power to watermills and water - raising machines. many of these technologies were transferred to medieval europe. wind - powered machines used to grind grain and pump water, the windmill and wind pump, first appeared in what are now iran, afghanistan and pakistan by the 9th century. they were used to grind grains and draw up water, and used in the gristmilling and sugarcane industries. sugar mills first appeared in the medieval islamic world. they were first driven by watermills, and then windmills from the 9th and 10th centuries in what are today afghanistan, pakistan and iran. crops such as almonds and citrus fruit were brought to europe through al - andalus, and sugar cultivation was gradually adopted across europe. arab merchants dominated trade in the indian ocean until the arrival of the portuguese in the 16th century. the muslim world adopted papermaking from china. the earliest paper mills appeared in abbasid - era baghdad during 794 β 795. the knowledge of gunpowder was also transmitted from china via predominantly islamic countries, where formulas for pure potassium nitrate were developed. the spinning wheel was invented in the islamic world by the early 11th century. it was later widely adopted in europe, where it was adapted into the spinning jenny, a key device during the industrial revolution. the crankshaft was invented by al - jazari in 1206, and is central to modern machinery such as the steam engine, internal combustion engine and automatic controls. the camshaft was also first described by al - jazari in 1206. early programmable machines were also invented in the muslim world. the first music sequencer, a programmable musical instrument, was an automated flute player invented by the banu musa brothers, described in their book of ingenious devices, in the 9th century. in 1206, al - jazari invented programmable automata / robots. he described four automaton musicians, including two drummers operated by a programmable drum machine, where the drummer could be made to play different rhythms and different drum patterns. the castle clock, a hydropowered mechanical astronomical clock invented by al - jazari, was an early programmable analog computer. in the ottoman empire, a practical impulse steam turbine was invented in 1551 by taqi ad - din muhammad ibn ma ' ruf in ottoman egypt. he described a method for rotating a spit by means of a jet of steam playing on rotary vanes around the periphery of a wheel. known as a steam jack, a similar device for rotating a spit was also later described by john
years ago, saw the introduction of pressure flaking, where a wood, bone, or antler punch could be used to shape a stone very finely. the end of the last ice age about 10, 000 years ago is taken as the end point of the upper paleolithic and the beginning of the epipaleolithic / mesolithic. the mesolithic technology included the use of microliths as composite stone tools, along with wood, bone, and antler tools. the later stone age, during which the rudiments of agricultural technology were developed, is called the neolithic period. during this period, polished stone tools were made from a variety of hard rocks such as flint, jade, jadeite, and greenstone, largely by working exposures as quarries, but later the valuable rocks were pursued by tunneling underground, the first steps in mining technology. the polished axes were used for forest clearance and the establishment of crop farming and were so effective as to remain in use when bronze and iron appeared. these stone axes were used alongside a continued use of stone tools such as a range of projectiles, knives, and scrapers, as well as tools, made from organic materials such as wood, bone, and antler. stone age cultures developed music and engaged in organized warfare. stone age humans developed ocean - worthy outrigger canoe technology, leading to migration across the malay archipelago, across the indian ocean to madagascar and also across the pacific ocean, which required knowledge of the ocean currents, weather patterns, sailing, and celestial navigation. although paleolithic cultures left no written records, the shift from nomadic life to settlement and agriculture can be inferred from a range of archaeological evidence. such evidence includes ancient tools, cave paintings, and other prehistoric art, such as the venus of willendorf. human remains also provide direct evidence, both through the examination of bones, and the study of mummies. scientists and historians have been able to form significant inferences about the lifestyle and culture of various prehistoric peoples, and especially their technology. = = = ancient = = = = = = = copper and bronze ages = = = = metallic copper occurs on the surface of weathered copper ore deposits and copper was used before copper smelting was known. copper smelting is believed to have originated when the technology of pottery kilns allowed sufficiently high temperatures. the concentration of various elements such as arsenic increase with depth in copper ore deposits and smelting of these ores yields arsenical bronze, which can be sufficiently
##osity and regular pore size. = = = = solvent casting and particulate leaching = = = = solvent casting and particulate leaching ( scpl ) allows for the preparation of structures with regular porosity, but with limited thickness. first, the polymer is dissolved into a suitable organic solvent ( e. g. polylactic acid could be dissolved into dichloromethane ), then the solution is cast into a mold filled with porogen particles. such porogen can be an inorganic salt like sodium chloride, crystals of saccharose, gelatin spheres or paraffin spheres. the size of the porogen particles will affect the size of the scaffold pores, while the polymer to porogen ratio is directly correlated to the amount of porosity of the final structure. after the polymer solution has been cast the solvent is allowed to fully evaporate, then the composite structure in the mold is immersed in a bath of a liquid suitable for dissolving the porogen : water in the case of sodium chloride, saccharose and gelatin or an aliphatic solvent like hexane for use with paraffin. once the porogen has been fully dissolved, a porous structure is obtained. other than the small thickness range that can be obtained, another drawback of scpl lies in its use of organic solvents which must be fully removed to avoid any possible damage to the cells seeded on the scaffold. = = = = gas foaming = = = = to overcome the need to use organic solvents and solid porogens, a technique using gas as a porogen has been developed. first, disc - shaped structures made of the desired polymer are prepared by means of compression molding using a heated mold. the discs are then placed in a chamber where they are exposed to high pressure co2 for several days. the pressure inside the chamber is gradually restored to atmospheric levels. during this procedure the pores are formed by the carbon dioxide molecules that abandon the polymer, resulting in a sponge - like structure. the main problems resulting from such a technique are caused by the excessive heat used during compression molding ( which prohibits the incorporation of any temperature labile material into the polymer matrix ) and by the fact that the pores do not form an interconnected structure. = = = = emulsification freeze - drying = = = = this technique does not require the use of a solid porogen like scpl. first, a synthetic polymer is dissolved into a suitable solvent ( e. g. poly
in the present - day universe, it appears that most, and perhaps all, massive stars are born in star clusters. it also appears that all star clusters contain stars drawn from an approximately universal initial mass function, so that almost all rich young star clusters contain massive stars. in this review i discuss the physical processes associated with both massive star formation and with star cluster formation. first i summarize the observed properties of star - forming gas clumps, then address the following questions. how do these clumps emerge from giant molecular clouds? in these clustered environments, how do individual stars form and gain mass? can a forming star cluster be treated as an equilibrium system or is this process too rapid for equilibrium to be established? how does feedback affect the formation process?
ultramagnetized neutron stars or magnetars are magnetically powered neutron stars. their strong magnetic fields dominate the physical processes in their crusts and their surroundings. the past few years have seen several advances in our theoretical and observational understanding of these objects. in spite of a surfeit of observations, their spectra are still poorly understood. i will discuss the emission from strongly magnetized condensed matter surfaces of neutron stars, recent advances in our expectations of the surface composition of magnetars and a model for the non - thermal emission from these objects.
subsea engineering and the ability to detect, track and destroy submarines ( anti - submarine warfare ) required the parallel development of a host of marine scientific instrumentation and sensors. visible light is not transferred far underwater, so the medium for transmission of data is primarily acoustic. high - frequency sound is used to measure the depth of the ocean, determine the nature of the seafloor, and detect submerged objects. the higher the frequency, the higher the definition of the data that is returned. sound navigation and ranging or sonar was developed during the first world war to detect submarines, and has been greatly refined through to the present day. submarines similarly use sonar equipment to detect and target other submarines and surface ships, and to detect submerged obstacles such as seamounts that pose a navigational obstacle. simple echo - sounders point straight down and can give an accurate reading of ocean depth ( or look up at the underside of sea - ice ). more advanced echo sounders use a fan - shaped beam or sound, or multiple beams to derive highly detailed images of the ocean floor. high power systems can penetrate the soil and seabed rocks to give information about the geology of the seafloor, and are widely used in geophysics for the discovery of hydrocarbons, or for engineering survey. for close - range underwater communications, optical transmission is possible, mainly using blue lasers. these have a high bandwidth compared with acoustic systems, but the range is usually only a few tens of metres, and ideally at night. as well as acoustic communications and navigation, sensors have been developed to measure ocean parameters such as temperature, salinity, oxygen levels and other properties including nitrate levels, levels of trace chemicals and environmental dna. the industry trend has been towards smaller, more accurate and more affordable systems so that they can be purchased and used by university departments and small companies as well as large corporations, research organisations and governments. the sensors and instruments are fitted to autonomous and remotely - operated systems as well as ships, and are enabling these systems to take on tasks that hitherto required an expensive human - crewed platform. manufacture of marine sensors and instruments mainly takes place in asia, europe and north america. products are advertised in specialist journals, and through trade shows such as oceanology international and ocean business which help raise awareness of the products. = = = environmental engineering = = = in every coastal and offshore project, environmental sustainability is an important consideration for the preservation of ocean ecosystems and natural resources. instances in which marine engineers benefit from knowledge of environmental engineering include creation of fisheries, clean
Question: Students stir salt, sand, sawdust, and seashells into a jar filled with water. Which of the materials will form a solution?
A) salt and water
B) sand and sawdust
C) seashells and water
D) sawdust and seashells
|
A) salt and water
|
Context:
##morphology studies the origin of landscapes. structural geology studies the deformation of rocks to produce mountains and lowlands. resource geology studies how energy resources can be obtained from minerals. environmental geology studies how pollution and contaminants affect soil and rock. mineralogy is the study of minerals and includes the study of mineral formation, crystal structure, hazards associated with minerals, and the physical and chemical properties of minerals. petrology is the study of rocks, including the formation and composition of rocks. petrography is a branch of petrology that studies the typology and classification of rocks. = = earth ' s interior = = plate tectonics, mountain ranges, volcanoes, and earthquakes are geological phenomena that can be explained in terms of physical and chemical processes in the earth ' s crust. beneath the earth ' s crust lies the mantle which is heated by the radioactive decay of heavy elements. the mantle is not quite solid and consists of magma which is in a state of semi - perpetual convection. this convection process causes the lithospheric plates to move, albeit slowly. the resulting process is known as plate tectonics. areas of the crust where new crust is created are called divergent boundaries, those where it is brought back into the earth are convergent boundaries and those where plates slide past each other, but no new lithospheric material is created or destroyed, are referred to as transform ( or conservative ) boundaries. earthquakes result from the movement of the lithospheric plates, and they often occur near convergent boundaries where parts of the crust are forced into the earth as part of subduction. plate tectonics might be thought of as the process by which the earth is resurfaced. as the result of seafloor spreading, new crust and lithosphere is created by the flow of magma from the mantle to the near surface, through fissures, where it cools and solidifies. through subduction, oceanic crust and lithosphere vehemently returns to the convecting mantle. volcanoes result primarily from the melting of subducted crust material. crust material that is forced into the asthenosphere melts, and some portion of the melted material becomes light enough to rise to the surface β giving birth to volcanoes. = = atmospheric science = = atmospheric science initially developed in the late - 19th century as a means to forecast the weather through meteorology, the study of weather. atmospheric chemistry was developed in the 20th century to measure air pollution and expanded in the 1970s in response to
##ctonics, mountain ranges, volcanoes, and earthquakes are geological phenomena that can be explained in terms of physical and chemical processes in the earth ' s crust. beneath the earth ' s crust lies the mantle which is heated by the radioactive decay of heavy elements. the mantle is not quite solid and consists of magma which is in a state of semi - perpetual convection. this convection process causes the lithospheric plates to move, albeit slowly. the resulting process is known as plate tectonics. areas of the crust where new crust is created are called divergent boundaries, those where it is brought back into the earth are convergent boundaries and those where plates slide past each other, but no new lithospheric material is created or destroyed, are referred to as transform ( or conservative ) boundaries. earthquakes result from the movement of the lithospheric plates, and they often occur near convergent boundaries where parts of the crust are forced into the earth as part of subduction. plate tectonics might be thought of as the process by which the earth is resurfaced. as the result of seafloor spreading, new crust and lithosphere is created by the flow of magma from the mantle to the near surface, through fissures, where it cools and solidifies. through subduction, oceanic crust and lithosphere vehemently returns to the convecting mantle. volcanoes result primarily from the melting of subducted crust material. crust material that is forced into the asthenosphere melts, and some portion of the melted material becomes light enough to rise to the surface β giving birth to volcanoes. = = atmospheric science = = atmospheric science initially developed in the late - 19th century as a means to forecast the weather through meteorology, the study of weather. atmospheric chemistry was developed in the 20th century to measure air pollution and expanded in the 1970s in response to acid rain. climatology studies the climate and climate change. the troposphere, stratosphere, mesosphere, thermosphere, and exosphere are the five layers which make up earth ' s atmosphere. 75 % of the mass in the atmosphere is located within the troposphere, the lowest layer. in all, the atmosphere is made up of about 78. 0 % nitrogen, 20. 9 % oxygen, and 0. 92 % argon, and small amounts of other gases including co2 and water vapor. water vapor and co2 cause the earth ' s atmosphere to catch and hold the sun ' s
28 size spectra of extensive air showers from 7 different experiments are analysed consistently. they are fitted by adjusting either 4 or 5 parameters : knee position, power law exponents above and below the knee, overall intensity and, in addition, a parameter describing the smoothness of the bend. the residuals are then normalized to the same knee position and averaged. when 5 parameters are employed no systematic deviation from a single smooth knee is apparent at the 1 % level up to a factor of 4 above the knee. at larger shower sizes a moderately significant deviation can be seen whose shape and position are compatible with a second knee caused by iron group nuclei.
, lightning strikes, tornadoes, building fires, wildfires, and mass shootings disabling most of the system if not the entirety of it. geographic redundancy locations can be more than 621 miles ( 999 km ) continental, more than 62 miles apart and less than 93 miles ( 150 km ) apart, less than 62 miles apart, but not on the same campus, or different buildings that are more than 300 feet ( 91 m ) apart on the same campus. the following methods can reduce the risks of damage by a fire conflagration : large buildings at least 80 feet ( 24 m ) to 110 feet ( 34 m ) apart, but sometimes a minimum of 210 feet ( 64 m ) apart. : 9 high - rise buildings at least 82 feet ( 25 m ) apart : 12 open spaces clear of flammable vegetation within 200 feet ( 61 m ) on each side of objects different wings on the same building, in rooms that are separated by more than 300 feet ( 91 m ) different floors on the same wing of a building in rooms that are horizontally offset by a minimum of 70 feet ( 21 m ) with fire walls between the rooms that are on different floors two rooms separated by another room, leaving at least a 70 - foot gap between the two rooms there should be a minimum of two separated fire walls and on opposite sides of a corridor geographic redundancy is used by amazon web services ( aws ), google cloud platform ( gcp ), microsoft azure, netflix, dropbox, salesforce, linkedin, paypal, twitter, facebook, apple icloud, cisco meraki, and many others to provide geographic redundancy, high availability, fault tolerance and to ensure availability and reliability for their cloud services. as another example, to minimize risk of damage from severe windstorms or water damage, buildings can be located at least 2 miles ( 3. 2 km ) away from the shore, with an elevation of at least 5 feet ( 1. 5 m ) above sea level. for additional protection, they can be located at least 100 feet ( 30 m ) away from flood plain areas. = = functions of redundancy = = the two functions of redundancy are passive redundancy and active redundancy. both functions prevent performance decline from exceeding specification limits without human intervention using extra capacity. passive redundancy uses excess capacity to reduce the impact of component failures. one common form of passive redundancy is the extra strength of cabling and struts used in bridges.
, no nuclear weapons have been deployed offensively. nevertheless, they prompted an arms race to develop increasingly destructive bombs to provide a nuclear deterrent. just over four years later, on august 29, 1949, the soviet union detonated its first fission weapon. the united kingdom followed on october 2, 1952 ; france, on february 13, 1960 ; and china component to a nuclear weapon. approximately half of the deaths from hiroshima and nagasaki died two to five years afterward from radiation exposure. a radiological weapon is a type of nuclear weapon designed to distribute hazardous nuclear material in enemy areas. such a weapon would not have the explosive capability of a fission or fusion bomb, but would kill many people and contaminate a large area. a radiological weapon has never been deployed. while considered useless by a conventional military, such a weapon raises concerns over nuclear terrorism. there have been over 2, 000 nuclear tests conducted since 1945. in 1963, all nuclear and many non - nuclear states signed the limited test ban treaty, pledging to refrain from testing nuclear weapons in the atmosphere, underwater, or in outer space. the treaty permitted underground nuclear testing. france continued atmospheric testing until 1974, while china continued up until 1980. the last underground test by the united states was in 1992, the soviet union in 1990, the united kingdom in 1991, and both france and china continued testing until 1996. after signing the comprehensive test ban treaty in 1996 ( which had as of 2011 not entered into force ), all of these states have pledged to discontinue all nuclear testing. non - signatories india and pakistan last tested nuclear weapons in 1998. nuclear weapons are the most destructive weapons known - the archetypal weapons of mass destruction. throughout the cold war, the opposing powers had huge nuclear arsenals, sufficient to kill hundreds of millions of people. generations of people grew up under the shadow of nuclear devastation, portrayed in films such as dr. strangelove and the atomic cafe. however, the tremendous energy release in the detonation of a nuclear weapon also suggested the possibility of a new energy source. = = civilian uses = = = = = nuclear power = = = nuclear power is a type of nuclear technology involving the controlled use of nuclear fission to release energy for work including propulsion, heat, and the generation of electricity. nuclear energy is produced by a controlled nuclear chain reaction which creates heat β and which is used to boil water, produce steam, and drive a steam turbine. the turbine is used to generate electricity and / or to do mechanical work. currently nuclear
radiation exposure. marie curie died from aplastic anemia which resulted from her high levels of exposure. two scientists, an american and canadian respectively, harry daghlian and louis slotin, died after mishandling the same plutonium mass. unlike conventional weapons, the intense light, heat, and explosive force is not the only deadly component to a nuclear weapon. approximately half of the deaths from hiroshima and nagasaki died two to five years afterward from radiation exposure. civilian nuclear and radiological accidents primarily involve nuclear power plants. most common are nuclear leaks that expose workers to hazardous material. a nuclear meltdown refers to the more serious hazard of releasing nuclear material into the surrounding environment. the most significant meltdowns occurred at three mile island in pennsylvania and chernobyl in the soviet ukraine. the earthquake and tsunami on march 11, 2011 caused serious damage to three nuclear reactors and a spent fuel storage pond at the fukushima daiichi nuclear power plant in japan. military reactors that experienced similar accidents were windscale in the united kingdom and sl - 1 in the united states. military accidents usually involve the loss or unexpected detonation of nuclear weapons. the castle bravo test in 1954 produced a larger yield than expected, which contaminated nearby islands, a japanese fishing boat ( with one fatality ), and raised concerns about contaminated fish in japan. in the 1950s through 1970s, several nuclear bombs were lost from submarines and aircraft, some of which have never been recovered. the last twenty years have seen a marked decline in such accidents. = = examples of environmental benefits = = proponents of nuclear energy note that annually, nuclear - generated electricity reduces 470 million metric tons of carbon dioxide emissions that would otherwise come from fossil fuels. additionally, the amount of comparatively low waste that nuclear energy does create is safely disposed of by the large scale nuclear energy production facilities or it is repurposed / recycled for other energy uses. proponents of nuclear energy also bring to attention the opportunity cost of utilizing other forms of electricity. for example, the environmental protection agency estimates that coal kills 30, 000 people a year, as a result of its environmental impact, while 60 people died in the chernobyl disaster. a real world example of impact provided by proponents of nuclear energy is the 650, 000 ton increase in carbon emissions in the two months following the closure of the vermont yankee nuclear plant. = = see also = = atomic age lists of nuclear disasters and radioactive incidents nuclear power debate outline of nuclear technology radiology = = references = = = = external links = = nuclear energy institute β beneficial uses
the group velocity of light has been measured at eight different wavelengths between 385 nm and 532 nm in the mediterranean sea at a depth of about 2. 2 km with the antares optical beacon systems. a parametrisation of the dependence of the refractive index on wavelength based on the salinity, pressure and temperature of the sea water at the antares site is in good agreement with these measurements.
, crystal structure, hazards associated with minerals, and the physical and chemical properties of minerals. petrology is the study of rocks, including the formation and composition of rocks. petrography is a branch of petrology that studies the typology and classification of rocks. = = earth ' s interior = = plate tectonics, mountain ranges, volcanoes, and earthquakes are geological phenomena that can be explained in terms of physical and chemical processes in the earth ' s crust. beneath the earth ' s crust lies the mantle which is heated by the radioactive decay of heavy elements. the mantle is not quite solid and consists of magma which is in a state of semi - perpetual convection. this convection process causes the lithospheric plates to move, albeit slowly. the resulting process is known as plate tectonics. areas of the crust where new crust is created are called divergent boundaries, those where it is brought back into the earth are convergent boundaries and those where plates slide past each other, but no new lithospheric material is created or destroyed, are referred to as transform ( or conservative ) boundaries. earthquakes result from the movement of the lithospheric plates, and they often occur near convergent boundaries where parts of the crust are forced into the earth as part of subduction. plate tectonics might be thought of as the process by which the earth is resurfaced. as the result of seafloor spreading, new crust and lithosphere is created by the flow of magma from the mantle to the near surface, through fissures, where it cools and solidifies. through subduction, oceanic crust and lithosphere vehemently returns to the convecting mantle. volcanoes result primarily from the melting of subducted crust material. crust material that is forced into the asthenosphere melts, and some portion of the melted material becomes light enough to rise to the surface β giving birth to volcanoes. = = atmospheric science = = atmospheric science initially developed in the late - 19th century as a means to forecast the weather through meteorology, the study of weather. atmospheric chemistry was developed in the 20th century to measure air pollution and expanded in the 1970s in response to acid rain. climatology studies the climate and climate change. the troposphere, stratosphere, mesosphere, thermosphere, and exosphere are the five layers which make up earth ' s atmosphere. 75 % of the mass in the atmosphere is located within the troposphere, the lowest
oscillations of the sun have been used to understand its interior structure. the extension of similar studies to more distant stars has raised many difficulties despite the strong efforts of the international community over the past decades. the corot ( convection rotation and planetary transits ) satellite, launched in december 2006, has now measured oscillations and the stellar granulation signature in three main sequence stars that are noticeably hotter than the sun. the oscillation amplitudes are about 1. 5 times as large as those in the sun ; the stellar granulation is up to three times as high. the stellar amplitudes are about 25 % below the theoretic values, providing a measurement of the nonadiabaticity of the process ruling the oscillations in the outer layers of the stars.
the gravitational acceleration of antimatter, $ \ bar g $, has yet to be directly measured but could change our understanding of gravity, the universe, and the possibility of a fifth force. three avenues are apparent for such a measurement : antihydrogen, positronium, and muonium, the last requiring a precision atom interferometer and benefiting from a novel muonium beam under development. the interferometer and its few - picometer alignment and calibration systems appear to be feasible. with 100 nm grating pitch, measurements of $ \ bar g $ to 10 %, 1 %, or better can be envisioned. this could constitute the first gravitational measurement of leptonic matter, of second - generation matter and, possibly, the first measurement of the gravitational acceleration of antimatter.
Question: Which scale is used to report the intensity of an earthquake?
A) Saffir-Simpson Scale
B) Modified Mercalli Scale
C) Richter Magnitude Scale
D) Moment Magnitude Scale
|
B) Modified Mercalli Scale
|
Context:
plant cells and tissues, whereas plant morphology is the study of their external form. all plants are multicellular eukaryotes, their dna stored in nuclei. the characteristic features of plant cells that distinguish them from those of animals and fungi include a primary cell wall composed of the polysaccharides cellulose, hemicellulose and pectin, larger vacuoles than in animal cells and the presence of plastids with unique photosynthetic and biosynthetic functions as in the chloroplasts. other plastids contain storage products such as starch ( amyloplasts ) or lipids ( elaioplasts ). uniquely, streptophyte cells and those of the green algal order trentepohliales divide by construction of a phragmoplast as a template for building a cell plate late in cell division. the bodies of vascular plants including clubmosses, ferns and seed plants ( gymnosperms and angiosperms ) generally have aerial and subterranean subsystems. the shoots consist of stems bearing green photosynthesising leaves and reproductive structures. the underground vascularised roots bear root hairs at their tips and generally lack chlorophyll. non - vascular plants, the liverworts, hornworts and mosses do not produce ground - penetrating vascular roots and most of the plant participates in photosynthesis. the sporophyte generation is nonphotosynthetic in liverworts but may be able to contribute part of its energy needs by photosynthesis in mosses and hornworts. the root system and the shoot system are interdependent β the usually nonphotosynthetic root system depends on the shoot system for food, and the usually photosynthetic shoot system depends on water and minerals from the root system. cells in each system are capable of creating cells of the other and producing adventitious shoots or roots. stolons and tubers are examples of shoots that can grow roots. roots that spread out close to the surface, such as those of willows, can produce shoots and ultimately new plants. in the event that one of the systems is lost, the other can often regrow it. in fact it is possible to grow an entire plant from a single leaf, as is the case with plants in streptocarpus sect. saintpaulia, or even a single cell β which can dedifferentiate into a callus ( a mass of
the internal functions and processes within plant organelles, cells, tissues, whole plants, plant populations and plant communities. at each of these levels, a botanist may be concerned with the classification ( taxonomy ), phylogeny and evolution, structure ( anatomy and morphology ), or function ( physiology ) of plant life. the strictest definition of " plant " includes only the " land plants " or embryophytes, which include seed plants ( gymnosperms, including the pines, and flowering plants ) and the free - sporing cryptogams including ferns, clubmosses, liverworts, hornworts and mosses. embryophytes are multicellular eukaryotes descended from an ancestor that obtained its energy from sunlight by photosynthesis. they have life cycles with alternating haploid and diploid phases. the sexual haploid phase of embryophytes, known as the gametophyte, nurtures the developing diploid embryo sporophyte within its tissues for at least part of its life, even in the seed plants, where the gametophyte itself is nurtured by its parent sporophyte. other groups of organisms that were previously studied by botanists include bacteria ( now studied in bacteriology ), fungi ( mycology ) β including lichen - forming fungi ( lichenology ), non - chlorophyte algae ( phycology ), and viruses ( virology ). however, attention is still given to these groups by botanists, and fungi ( including lichens ) and photosynthetic protists are usually covered in introductory botany courses. palaeobotanists study ancient plants in the fossil record to provide information about the evolutionary history of plants. cyanobacteria, the first oxygen - releasing photosynthetic organisms on earth, are thought to have given rise to the ancestor of plants by entering into an endosymbiotic relationship with an early eukaryote, ultimately becoming the chloroplasts in plant cells. the new photosynthetic plants ( along with their algal relatives ) accelerated the rise in atmospheric oxygen started by the cyanobacteria, changing the ancient oxygen - free, reducing, atmosphere to one in which free oxygen has been abundant for more than 2 billion years. among the important botanical questions of the 21st century are the role of plants as primary producers in the global cycling of life ' s basic ingredients : energy, carbon, oxygen, nitrogen and water, and ways
frits went. the first known auxin, indole - 3 - acetic acid ( iaa ), which promotes cell growth, was only isolated from plants about 50 years later. this compound mediates the tropic responses of shoots and roots towards light and gravity. the finding in 1939 that plant callus could be maintained in culture containing iaa, followed by the observation in 1947 that it could be induced to form roots and shoots by controlling the concentration of growth hormones were key steps in the development of plant biotechnology and genetic modification. cytokinins are a class of plant hormones named for their control of cell division ( especially cytokinesis ). the natural cytokinin zeatin was discovered in corn, zea mays, and is a derivative of the purine adenine. zeatin is produced in roots and transported to shoots in the xylem where it promotes cell division, bud development, and the greening of chloroplasts. the gibberelins, such as gibberelic acid are diterpenes synthesised from acetyl coa via the mevalonate pathway. they are involved in the promotion of germination and dormancy - breaking in seeds, in regulation of plant height by controlling stem elongation and the control of flowering. abscisic acid ( aba ) occurs in all land plants except liverworts, and is synthesised from carotenoids in the chloroplasts and other plastids. it inhibits cell division, promotes seed maturation, and dormancy, and promotes stomatal closure. it was so named because it was originally thought to control abscission. ethylene is a gaseous hormone that is produced in all higher plant tissues from methionine. it is now known to be the hormone that stimulates or regulates fruit ripening and abscission, and it, or the synthetic growth regulator ethephon which is rapidly metabolised to produce ethylene, are used on industrial scale to promote ripening of cotton, pineapples and other climacteric crops. another class of phytohormones is the jasmonates, first isolated from the oil of jasminum grandiflorum which regulates wound responses in plants by unblocking the expression of genes required in the systemic acquired resistance response to pathogen attack. in addition to being the primary energy source for plants, light functions as a signalling device, providing information to the plant, such as how
pectin and xyloglucan from which the land plant cell wall is constructed. vascular land plants make lignin, a polymer used to strengthen the secondary cell walls of xylem tracheids and vessels to keep them from collapsing when a plant sucks water through them under water stress. lignin is also used in other cell types like sclerenchyma fibres that provide structural support for a plant and is a major constituent of wood. sporopollenin is a chemically resistant polymer found in the outer cell walls of spores and pollen of land plants responsible for the survival of early land plant spores and the pollen of seed plants in the fossil record. it is widely regarded as a marker for the start of land plant evolution during the ordovician period. the concentration of carbon dioxide in the atmosphere today is much lower than it was when plants emerged onto land during the ordovician and silurian periods. many monocots like maize and the pineapple and some dicots like the asteraceae have since independently evolved pathways like crassulacean acid metabolism and the c4 carbon fixation pathway for photosynthesis which avoid the losses resulting from photorespiration in the more common c3 carbon fixation pathway. these biochemical strategies are unique to land plants. = = = medicine and materials = = = phytochemistry is a branch of plant biochemistry primarily concerned with the chemical substances produced by plants during secondary metabolism. some of these compounds are toxins such as the alkaloid coniine from hemlock. others, such as the essential oils peppermint oil and lemon oil are useful for their aroma, as flavourings and spices ( e. g., capsaicin ), and in medicine as pharmaceuticals as in opium from opium poppies. many medicinal and recreational drugs, such as tetrahydrocannabinol ( active ingredient in cannabis ), caffeine, morphine and nicotine come directly from plants. others are simple derivatives of botanical natural products. for example, the pain killer aspirin is the acetyl ester of salicylic acid, originally isolated from the bark of willow trees, and a wide range of opiate painkillers like heroin are obtained by chemical modification of morphine obtained from the opium poppy. popular stimulants come from plants, such as caffeine from coffee, tea and chocolate, and nicotine from tobacco. most alcoholic beverages come from fermentation of carbohy
shaping the cell. cell membranes are involved in various cellular processes such as cell adhesion, storing electrical energy, and cell signalling and serve as the attachment surface for several extracellular structures such as a cell wall, glycocalyx, and cytoskeleton. within the cytoplasm of a cell, there are many biomolecules such as proteins and nucleic acids. in addition to biomolecules, eukaryotic cells have specialized structures called organelles that have their own lipid bilayers or are spatially units. these organelles include the cell nucleus, which contains most of the cell ' s dna, or mitochondria, which generate adenosine triphosphate ( atp ) to power cellular processes. other organelles such as endoplasmic reticulum and golgi apparatus play a role in the synthesis and packaging of proteins, respectively. biomolecules such as proteins can be engulfed by lysosomes, another specialized organelle. plant cells have additional organelles that distinguish them from animal cells such as a cell wall that provides support for the plant cell, chloroplasts that harvest sunlight energy to produce sugar, and vacuoles that provide storage and structural support as well as being involved in reproduction and breakdown of plant seeds. eukaryotic cells also have cytoskeleton that is made up of microtubules, intermediate filaments, and microfilaments, all of which provide support for the cell and are involved in the movement of the cell and its organelles. in terms of their structural composition, the microtubules are made up of tubulin ( e. g., Ξ± - tubulin and Ξ² - tubulin ) whereas intermediate filaments are made up of fibrous proteins. microfilaments are made up of actin molecules that interact with other strands of proteins. = = = metabolism = = = all cells require energy to sustain cellular processes. metabolism is the set of chemical reactions in an organism. the three main purposes of metabolism are : the conversion of food to energy to run cellular processes ; the conversion of food / fuel to monomer building blocks ; and the elimination of metabolic wastes. these enzyme - catalyzed reactions allow organisms to grow and reproduce, maintain their structures, and respond to their environments. metabolic reactions may be categorized as catabolic β the breaking down of compounds ( for example, the breaking down of glucose to pyruvate by cellular respiration
, there are many biomolecules such as proteins and nucleic acids. in addition to biomolecules, eukaryotic cells have specialized structures called organelles that have their own lipid bilayers or are spatially units. these organelles include the cell nucleus, which contains most of the cell ' s dna, or mitochondria, which generate adenosine triphosphate ( atp ) to power cellular processes. other organelles such as endoplasmic reticulum and golgi apparatus play a role in the synthesis and packaging of proteins, respectively. biomolecules such as proteins can be engulfed by lysosomes, another specialized organelle. plant cells have additional organelles that distinguish them from animal cells such as a cell wall that provides support for the plant cell, chloroplasts that harvest sunlight energy to produce sugar, and vacuoles that provide storage and structural support as well as being involved in reproduction and breakdown of plant seeds. eukaryotic cells also have cytoskeleton that is made up of microtubules, intermediate filaments, and microfilaments, all of which provide support for the cell and are involved in the movement of the cell and its organelles. in terms of their structural composition, the microtubules are made up of tubulin ( e. g., Ξ± - tubulin and Ξ² - tubulin ) whereas intermediate filaments are made up of fibrous proteins. microfilaments are made up of actin molecules that interact with other strands of proteins. = = = metabolism = = = all cells require energy to sustain cellular processes. metabolism is the set of chemical reactions in an organism. the three main purposes of metabolism are : the conversion of food to energy to run cellular processes ; the conversion of food / fuel to monomer building blocks ; and the elimination of metabolic wastes. these enzyme - catalyzed reactions allow organisms to grow and reproduce, maintain their structures, and respond to their environments. metabolic reactions may be categorized as catabolic β the breaking down of compounds ( for example, the breaking down of glucose to pyruvate by cellular respiration ) ; or anabolic β the building up ( synthesis ) of compounds ( such as proteins, carbohydrates, lipids, and nucleic acids ). usually, catabolism releases energy, and anabolism consumes energy. the chemical reactions of metabolism are organized into metabolic pathways, in which
smallest genomes among flowering plants. arabidopsis was the first plant to have its genome sequenced, in 2000. the sequencing of some other relatively small genomes, of rice ( oryza sativa ) and brachypodium distachyon, has made them important model species for understanding the genetics, cellular and molecular biology of cereals, grasses and monocots generally. model plants such as arabidopsis thaliana are used for studying the molecular biology of plant cells and the chloroplast. ideally, these organisms have small genomes that are well known or completely sequenced, small stature and short generation times. corn has been used to study mechanisms of photosynthesis and phloem loading of sugar in c4 plants. the single celled green alga chlamydomonas reinhardtii, while not an embryophyte itself, contains a green - pigmented chloroplast related to that of land plants, making it useful for study. a red alga cyanidioschyzon merolae has also been used to study some basic chloroplast functions. spinach, peas, soybeans and a moss physcomitrella patens are commonly used to study plant cell biology. agrobacterium tumefaciens, a soil rhizosphere bacterium, can attach to plant cells and infect them with a callus - inducing ti plasmid by horizontal gene transfer, causing a callus infection called crown gall disease. schell and van montagu ( 1977 ) hypothesised that the ti plasmid could be a natural vector for introducing the nif gene responsible for nitrogen fixation in the root nodules of legumes and other plant species. today, genetic modification of the ti plasmid is one of the main techniques for introduction of transgenes to plants and the creation of genetically modified crops. = = = epigenetics = = = epigenetics is the study of heritable changes in gene function that cannot be explained by changes in the underlying dna sequence but cause the organism ' s genes to behave ( or " express themselves " ) differently. one example of epigenetic change is the marking of the genes by dna methylation which determines whether they will be expressed or not. gene expression can also be controlled by repressor proteins that attach to silencer regions of the dna and prevent that region of the dna code from being expressed. epigenetic marks may be added
##aggeration to say that the tip of the radicle.. acts like the brain of one of the lower animals.. directing the several movements ". about the same time, the role of auxins ( from the greek auxein, to grow ) in control of plant growth was first outlined by the dutch scientist frits went. the first known auxin, indole - 3 - acetic acid ( iaa ), which promotes cell growth, was only isolated from plants about 50 years later. this compound mediates the tropic responses of shoots and roots towards light and gravity. the finding in 1939 that plant callus could be maintained in culture containing iaa, followed by the observation in 1947 that it could be induced to form roots and shoots by controlling the concentration of growth hormones were key steps in the development of plant biotechnology and genetic modification. cytokinins are a class of plant hormones named for their control of cell division ( especially cytokinesis ). the natural cytokinin zeatin was discovered in corn, zea mays, and is a derivative of the purine adenine. zeatin is produced in roots and transported to shoots in the xylem where it promotes cell division, bud development, and the greening of chloroplasts. the gibberelins, such as gibberelic acid are diterpenes synthesised from acetyl coa via the mevalonate pathway. they are involved in the promotion of germination and dormancy - breaking in seeds, in regulation of plant height by controlling stem elongation and the control of flowering. abscisic acid ( aba ) occurs in all land plants except liverworts, and is synthesised from carotenoids in the chloroplasts and other plastids. it inhibits cell division, promotes seed maturation, and dormancy, and promotes stomatal closure. it was so named because it was originally thought to control abscission. ethylene is a gaseous hormone that is produced in all higher plant tissues from methionine. it is now known to be the hormone that stimulates or regulates fruit ripening and abscission, and it, or the synthetic growth regulator ethephon which is rapidly metabolised to produce ethylene, are used on industrial scale to promote ripening of cotton, pineapples and other climacteric crops. another class of phytohormones is the jasmonates, first isolated
much sunlight the plant receives each day. this can result in adaptive changes in a process known as photomorphogenesis. phytochromes are the photoreceptors in a plant that are sensitive to light. = = plant anatomy and morphology = = plant anatomy is the study of the structure of plant cells and tissues, whereas plant morphology is the study of their external form. all plants are multicellular eukaryotes, their dna stored in nuclei. the characteristic features of plant cells that distinguish them from those of animals and fungi include a primary cell wall composed of the polysaccharides cellulose, hemicellulose and pectin, larger vacuoles than in animal cells and the presence of plastids with unique photosynthetic and biosynthetic functions as in the chloroplasts. other plastids contain storage products such as starch ( amyloplasts ) or lipids ( elaioplasts ). uniquely, streptophyte cells and those of the green algal order trentepohliales divide by construction of a phragmoplast as a template for building a cell plate late in cell division. the bodies of vascular plants including clubmosses, ferns and seed plants ( gymnosperms and angiosperms ) generally have aerial and subterranean subsystems. the shoots consist of stems bearing green photosynthesising leaves and reproductive structures. the underground vascularised roots bear root hairs at their tips and generally lack chlorophyll. non - vascular plants, the liverworts, hornworts and mosses do not produce ground - penetrating vascular roots and most of the plant participates in photosynthesis. the sporophyte generation is nonphotosynthetic in liverworts but may be able to contribute part of its energy needs by photosynthesis in mosses and hornworts. the root system and the shoot system are interdependent β the usually nonphotosynthetic root system depends on the shoot system for food, and the usually photosynthetic shoot system depends on water and minerals from the root system. cells in each system are capable of creating cells of the other and producing adventitious shoots or roots. stolons and tubers are examples of shoots that can grow roots. roots that spread out close to the surface, such as those of willows, can produce shoots and ultimately new plants. in the event that one of the systems is lost
or lipids ( elaioplasts ). uniquely, streptophyte cells and those of the green algal order trentepohliales divide by construction of a phragmoplast as a template for building a cell plate late in cell division. the bodies of vascular plants including clubmosses, ferns and seed plants ( gymnosperms and angiosperms ) generally have aerial and subterranean subsystems. the shoots consist of stems bearing green photosynthesising leaves and reproductive structures. the underground vascularised roots bear root hairs at their tips and generally lack chlorophyll. non - vascular plants, the liverworts, hornworts and mosses do not produce ground - penetrating vascular roots and most of the plant participates in photosynthesis. the sporophyte generation is nonphotosynthetic in liverworts but may be able to contribute part of its energy needs by photosynthesis in mosses and hornworts. the root system and the shoot system are interdependent β the usually nonphotosynthetic root system depends on the shoot system for food, and the usually photosynthetic shoot system depends on water and minerals from the root system. cells in each system are capable of creating cells of the other and producing adventitious shoots or roots. stolons and tubers are examples of shoots that can grow roots. roots that spread out close to the surface, such as those of willows, can produce shoots and ultimately new plants. in the event that one of the systems is lost, the other can often regrow it. in fact it is possible to grow an entire plant from a single leaf, as is the case with plants in streptocarpus sect. saintpaulia, or even a single cell β which can dedifferentiate into a callus ( a mass of unspecialised cells ) that can grow into a new plant. in vascular plants, the xylem and phloem are the conductive tissues that transport resources between shoots and roots. roots are often adapted to store food such as sugars or starch, as in sugar beets and carrots. stems mainly provide support to the leaves and reproductive structures, but can store water in succulent plants such as cacti, food as in potato tubers, or reproduce vegetatively as in the stolons of strawberry plants or in the process of layering. leaves gather sunlight and carry out photosyn
Question: What is the outermost structure in a plant cell?
A) cell membrane
B) cytoplasm
C) cell wall
D) chloroplast
|
C) cell wall
|
Context:
eat them. plants and other photosynthetic organisms are at the base of most food chains because they use the energy from the sun and nutrients from the soil and atmosphere, converting them into a form that can be used by animals. this is what ecologists call the first trophic level. the modern forms of the major staple foods, such as hemp, teff, maize, rice, wheat and other cereal grasses, pulses, bananas and plantains, as well as hemp, flax and cotton grown for their fibres, are the outcome of prehistoric selection over thousands of years from among wild ancestral plants with the most desirable characteristics. botanists study how plants produce food and how to increase yields, for example through plant breeding, making their work important to humanity ' s ability to feed the world and provide food security for future generations. botanists also study weeds, which are a considerable problem in agriculture, and the biology and control of plant pathogens in agriculture and natural ecosystems. ethnobotany is the study of the relationships between plants and people. when applied to the investigation of historical plant β people relationships ethnobotany may be referred to as archaeobotany or palaeoethnobotany. some of the earliest plant - people relationships arose between the indigenous people of canada in identifying edible plants from inedible plants. this relationship the indigenous people had with plants was recorded by ethnobotanists. = = plant biochemistry = = plant biochemistry is the study of the chemical processes used by plants. some of these processes are used in their primary metabolism like the photosynthetic calvin cycle and crassulacean acid metabolism. others make specialised materials like the cellulose and lignin used to build their bodies, and secondary products like resins and aroma compounds. plants and various other groups of photosynthetic eukaryotes collectively known as " algae " have unique organelles known as chloroplasts. chloroplasts are thought to be descended from cyanobacteria that formed endosymbiotic relationships with ancient plant and algal ancestors. chloroplasts and cyanobacteria contain the blue - green pigment chlorophyll a. chlorophyll a ( as well as its plant and green algal - specific cousin chlorophyll b ) absorbs light in the blue - violet and orange / red parts of the spectrum while reflecting and transmitting the green light that we see as the characteristic colour
while the modern stellar imf shows a rapid decline with increasing mass, theoretical investigations suggest that very massive stars ( > 100 solar masses ) may have been abundant in the early universe. other calculations also indicate that, lacking metals, these same stars reach their late evolutionary stages without appreciable mass loss. after central helium burning, they encounter the electron - positron pair instability, collapse, and burn oxygen and silicon explosively. if sufficient energy is released by the burning, these stars explode as brilliant supernovae with energies up to 100 times that of an ordinary core collapse supernova. they also eject up to 50 solar masses of radioactive ni56. stars less massive than 140 solar masses or more massive than 260 solar masses should collapse into black holes instead of exploding, thus bounding the pair - creation supernovae with regions of stellar mass that are nucleosynthetically sterile. pair - instability supernovae might be detectable in the near infrared out to redshifts of 20 or more and their ashes should leave a distinctive nucleosynthetic pattern.
the less of it people would be prepared to buy ( other things unchanged ). as the price of a commodity falls, consumers move toward it from relatively more expensive goods ( the substitution effect ). in addition, purchasing power from the price decline increases ability to buy ( the income effect ). other factors can change demand ; for example an increase in income will shift the demand curve for a normal good outward relative to the origin, as in the figure. all determinants are predominantly taken as constant factors of demand and supply. supply is the relation between the price of a good and the quantity available for sale at that price. it may be represented as a table or graph relating price and quantity supplied. producers, for example business firms, are hypothesised to be profit maximisers, meaning that they attempt to produce and supply the amount of goods that will bring them the highest profit. supply is typically represented as a function relating price and quantity, if other factors are unchanged. that is, the higher the price at which the good can be sold, the more of it producers will supply, as in the figure. the higher price makes it profitable to increase production. just as on the demand side, the position of the supply can shift, say from a change in the price of a productive input or a technical improvement. the " law of supply " states that, in general, a rise in price leads to an expansion in supply and a fall in price leads to a contraction in supply. here as well, the determinants of supply, such as price of substitutes, cost of production, technology applied and various factors inputs of production are all taken to be constant for a specific time period of evaluation of supply. market equilibrium occurs where quantity supplied equals quantity demanded, the intersection of the supply and demand curves in the figure above. at a price below equilibrium, there is a shortage of quantity supplied compared to quantity demanded. this is posited to bid the price up. at a price above equilibrium, there is a surplus of quantity supplied compared to quantity demanded. this pushes the price down. the model of supply and demand predicts that for given supply and demand curves, price and quantity will stabilise at the price that makes quantity supplied equal to quantity demanded. similarly, demand - and - supply theory predicts a new price - quantity combination from a shift in demand ( as to the figure ), or in supply. = = = firms = = = people frequently do not trade directly on markets. instead, on the supply side, they may work
the constraints on demand ). here, utility refers to the hypothesised relation of each individual consumer for ranking different commodity bundles as more or less preferred. the law of demand states that, in general, price and quantity demanded in a given market are inversely related. that is, the higher the price of a product, the less of it people would be prepared to buy ( other things unchanged ). as the price of a commodity falls, consumers move toward it from relatively more expensive goods ( the substitution effect ). in addition, purchasing power from the price decline increases ability to buy ( the income effect ). other factors can change demand ; for example an increase in income will shift the demand curve for a normal good outward relative to the origin, as in the figure. all determinants are predominantly taken as constant factors of demand and supply. supply is the relation between the price of a good and the quantity available for sale at that price. it may be represented as a table or graph relating price and quantity supplied. producers, for example business firms, are hypothesised to be profit maximisers, meaning that they attempt to produce and supply the amount of goods that will bring them the highest profit. supply is typically represented as a function relating price and quantity, if other factors are unchanged. that is, the higher the price at which the good can be sold, the more of it producers will supply, as in the figure. the higher price makes it profitable to increase production. just as on the demand side, the position of the supply can shift, say from a change in the price of a productive input or a technical improvement. the " law of supply " states that, in general, a rise in price leads to an expansion in supply and a fall in price leads to a contraction in supply. here as well, the determinants of supply, such as price of substitutes, cost of production, technology applied and various factors inputs of production are all taken to be constant for a specific time period of evaluation of supply. market equilibrium occurs where quantity supplied equals quantity demanded, the intersection of the supply and demand curves in the figure above. at a price below equilibrium, there is a shortage of quantity supplied compared to quantity demanded. this is posited to bid the price up. at a price above equilibrium, there is a surplus of quantity supplied compared to quantity demanded. this pushes the price down. the model of supply and demand predicts that for given supply and demand curves, price and quantity will stabilise at the price that makes quantity
for example an increase in income will shift the demand curve for a normal good outward relative to the origin, as in the figure. all determinants are predominantly taken as constant factors of demand and supply. supply is the relation between the price of a good and the quantity available for sale at that price. it may be represented as a table or graph relating price and quantity supplied. producers, for example business firms, are hypothesised to be profit maximisers, meaning that they attempt to produce and supply the amount of goods that will bring them the highest profit. supply is typically represented as a function relating price and quantity, if other factors are unchanged. that is, the higher the price at which the good can be sold, the more of it producers will supply, as in the figure. the higher price makes it profitable to increase production. just as on the demand side, the position of the supply can shift, say from a change in the price of a productive input or a technical improvement. the " law of supply " states that, in general, a rise in price leads to an expansion in supply and a fall in price leads to a contraction in supply. here as well, the determinants of supply, such as price of substitutes, cost of production, technology applied and various factors inputs of production are all taken to be constant for a specific time period of evaluation of supply. market equilibrium occurs where quantity supplied equals quantity demanded, the intersection of the supply and demand curves in the figure above. at a price below equilibrium, there is a shortage of quantity supplied compared to quantity demanded. this is posited to bid the price up. at a price above equilibrium, there is a surplus of quantity supplied compared to quantity demanded. this pushes the price down. the model of supply and demand predicts that for given supply and demand curves, price and quantity will stabilise at the price that makes quantity supplied equal to quantity demanded. similarly, demand - and - supply theory predicts a new price - quantity combination from a shift in demand ( as to the figure ), or in supply. = = = firms = = = people frequently do not trade directly on markets. instead, on the supply side, they may work in and produce through firms. the most obvious kinds of firms are corporations, partnerships and trusts. according to ronald coase, people begin to organise their production in firms when the costs of doing business becomes lower than doing it on the market. firms combine labour and capital, and can achieve far greater economies of scale ( when
to block. all three types of radiation occur naturally in certain elements. it has also become clear that the ultimate source of most terrestrial energy is nuclear, either through radiation from the sun caused by stellar thermonuclear reactions or by radioactive decay of uranium within the earth, the principal source of geothermal energy. = = = nuclear fission = = = in natural nuclear radiation, the byproducts are very small compared to the nuclei from which they originate. nuclear fission is the process of splitting a nucleus into roughly equal parts, and releasing energy and neutrons in the process. if these neutrons are captured by another unstable nucleus, they can fission as well, leading to a chain reaction. the average number of neutrons released per nucleus that go on to fission another nucleus is referred to as k. values of k larger than 1 mean that the fission reaction is releasing more neutrons than it absorbs, and therefore is referred to as a self - sustaining chain reaction. a mass of fissile material large enough ( and in a suitable configuration ) to induce a self - sustaining chain reaction is called a critical mass. when a neutron is captured by a suitable nucleus, fission may occur immediately, or the nucleus may persist in an unstable state for a short time. if there are enough immediate decays to carry on the chain reaction, the mass is said to be prompt critical, and the energy release will grow rapidly and uncontrollably, usually leading to an explosion. when discovered on the eve of world war ii, this insight led multiple countries to begin programs investigating the possibility of constructing an atomic bomb β a weapon which utilized fission reactions to generate far more energy than could be created with chemical explosives. the manhattan project, run by the united states with the help of the united kingdom and canada, developed multiple fission weapons which were used against japan in 1945 at hiroshima and nagasaki. during the project, the first fission reactors were developed as well, though they were primarily for weapons manufacture and did not generate electricity. in 1951, the first nuclear fission power plant was the first to produce electricity at the experimental breeder reactor no. 1 ( ebr - 1 ), in arco, idaho, ushering in the " atomic age " of more intensive human energy use. however, if the mass is critical only when the delayed neutrons are included, then the reaction can be controlled, for example by the introduction or removal of neutron absorbers. this is what allows nuclear reactors to be built. fast neutrons are not easily captured by nuclei
einstein, when he began working on the general theory of relativity, believed that energy of any kind is the source of the gravitational field. therefore, the energy of gravity, like any energy, must be the source of the field. it was previously discovered that the energy - momentum tensor of the gravitational field is already contained in the ricci tensor. this hypothesis is used to construct a new equation of the gravitational field.
##grade, in digital television picture quality is not affected by poor reception until, at a certain point, the receiver stops working and the screen goes black. terrestrial television, over - the - air ( ota ) television, or broadcast television β the oldest television technology, is the transmission of television signals from land - based television stations to television receivers ( called televisions or tvs ) in viewer ' s homes. terrestrial television broadcasting uses the bands 41 β 88 mhz ( vhf low band or band i, carrying rf channels 1 β 6 ), 174 β 240 mhz, ( vhf high band or band iii ; carrying rf channels 7 β 13 ), and 470 β 614 mhz ( uhf band iv and band v ; carrying rf channels 14 and up ). the exact frequency boundaries vary in different countries. propagation is by line - of - sight, so reception is limited by the visual horizon. in the us, the effective radiated power ( erp ) of television transmitters is regulated according to height above average terrain. viewers closer to the television transmitter can use a simple " rabbit ears " dipole antenna on top of the tv, but viewers in fringe reception areas typically require an outdoor antenna mounted on the roof to get adequate reception. satellite television β a set - top box which receives subscription direct - broadcast satellite television, and displays it on an ordinary television. a direct broadcast satellite in geostationary orbit 22, 200 miles ( 35, 700 km ) above the earth ' s equator transmits many channels ( up to 900 ) modulated on a 12. 2 to 12. 7 ghz ku band microwave downlink signal to a rooftop satellite dish antenna on the subscriber ' s residence. the microwave signal is converted to a lower intermediate frequency at the dish and conducted into the building by a coaxial cable to a set - top box connected to the subscriber ' s tv, where it is demodulated and displayed. the subscriber pays a monthly fee. = = = = time and frequency = = = = government standard frequency and time signal services operate time radio stations which continuously broadcast extremely accurate time signals produced by atomic clocks, as a reference to synchronize other clocks. examples are bpc, dcf77, jjy, msf, rtz, tdf, wwv, and yvto. one use is in radio clocks and watches, which include an automated receiver that periodically ( usually weekly ) receives and decodes the time signal and resets the watch ' s internal quartz clock to the correct time
##fer the validity of a general statement from a number of specific instances or infer the truth of a theory from a series of successful tests. for example, a chicken observes that each morning the farmer comes and gives it food, for hundreds of days in a row. the chicken may therefore use inductive reasoning to infer that the farmer will bring food every morning. however, one morning, the farmer comes and kills the chicken. how is scientific reasoning more trustworthy than the chicken ' s reasoning? one approach is to acknowledge that induction cannot achieve certainty, but observing more instances of a general statement can at least make the general statement more probable. so the chicken would be right to conclude from all those mornings that it is likely the farmer will come with food again the next morning, even if it cannot be certain. however, there remain difficult questions about the process of interpreting any given evidence into a probability that the general statement is true. one way out of these particular difficulties is to declare that all beliefs about scientific theories are subjective, or personal, and correct reasoning is merely about how evidence should change one ' s subjective beliefs over time. some argue that what scientists do is not inductive reasoning at all but rather abductive reasoning, or inference to the best explanation. in this account, science is not about generalizing specific instances but rather about hypothesizing explanations for what is observed. as discussed in the previous section, it is not always clear what is meant by the " best explanation ". ockham ' s razor, which counsels choosing the simplest available explanation, thus plays an important role in some versions of this approach. to return to the example of the chicken, would it be simpler to suppose that the farmer cares about it and will continue taking care of it indefinitely or that the farmer is fattening it up for slaughter? philosophers have tried to make this heuristic principle more precise regarding theoretical parsimony or other measures. yet, although various measures of simplicity have been brought forward as potential candidates, it is generally accepted that there is no such thing as a theory - independent measure of simplicity. in other words, there appear to be as many different measures of simplicity as there are theories themselves, and the task of choosing between measures of simplicity appears to be every bit as problematic as the job of choosing between theories. nicholas maxwell has argued for some decades that unity rather than simplicity is the key non - empirical factor in influencing the choice of theory in science, persistent preference for unified theories in effect committing science to the
also called projection lines ) differs, as explained below. in first - angle projection, the parallel projectors originate as if radiated from behind the viewer and pass through the 3d object to project a 2d image onto the orthogonal plane behind it. the 3d object is projected into 2d " paper " space as if you were looking at a radiograph of the object : the top view is under the front view, the right view is at the left of the front view. first - angle projection is the iso standard and is primarily used in europe. in third - angle projection, the parallel projectors originate as if radiated from the far side of the object and pass through the 3d object to project a 2d image onto the orthogonal plane in front of it. the views of the 3d object are like the panels of a box that envelopes the object, and the panels pivot as they open up flat into the plane of the drawing. thus the left view is placed on the left and the top view on the top ; and the features closest to the front of the 3d object will appear closest to the front view in the drawing. third - angle projection is primarily used in the united states and canada, where it is the default projection system according to asme standard asme y14. 3m. until the late 19th century, first - angle projection was the norm in north america as well as europe ; but circa the 1890s, third - angle projection spread throughout the north american engineering and manufacturing communities to the point of becoming a widely followed convention, and it was an asa standard by the 1950s. circa world war i, british practice was frequently mixing the use of both projection methods. as shown above, the determination of what surface constitutes the front, back, top, and bottom varies depending on the projection method used. not all views are necessarily used. generally only as many views are used as are necessary to convey all needed information clearly and economically. the front, top, and right - side views are commonly considered the core group of views included by default, but any combination of views may be used depending on the needs of the particular design. in addition to the six principal views ( front, back, top, bottom, right side, left side ), any auxiliary views or sections may be included as serve the purposes of part definition and its communication. view lines or section lines ( lines with arrows marked " a - a ", " b - b ", etc. ) define the direction and location of viewing or sectioning. sometimes a note tells the reader in which zone
Question: The main source of food for a coyote population in one area is mice. Which is most likely to happen to these coyotes if a new predator begins eating the mice?
A) There will be fewer coyotes.
B) The coyotes will produce more young.
C) More coyotes will arrive from other locations.
D) Each coyote will spend less time hunting food.
|
A) There will be fewer coyotes.
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.