source
stringlengths 31
227
| text
stringlengths 9
2k
|
---|---|
https://en.wikipedia.org/wiki/Waters%20of%20March
|
"Waters of March" ( ) is a Brazilian song composed by Antônio Carlos Jobim (1927–1994) in 1972. Jobim wrote both the Portuguese and English lyrics. The lyrics, originally written in Portuguese, do not tell a story, but rather present a series of images that form a collage; nearly every line starts with "É..." ("[It] is..."). In 2001, "Águas de março" was named as the all-time best Brazilian song in a poll of more than 200 Brazilian journalists, musicians and other artists conducted by Brazil's leading daily newspaper, Folha de S.Paulo. It was also voted by the Brazilian edition of Rolling Stone as the second greatest Brazilian song.
The inspiration for "Águas de março" came from Rio de Janeiro's rainiest month. March is typically marked by sudden storms with heavy rains and strong winds that cause flooding in many places around the city. The lyrics and the music have a constant downward progression much like the water torrent from those rains flowing in the gutters, which typically would carry sticks, stones, bits of glass, and almost everything and anything.
Lyrics
In both the Portuguese and English versions of the lyrics, "it" is a stick, a stone, a sliver of glass, a scratch, a cliff, a knot in the wood, a fish, a pin, the end of the road, and many other things, although some specific references to Brazilian culture (festa da cumeeira, garrafa de cana), flora (peroba do campo), folklore and fauna (Matita Pereira) were intentionally omitted from the English version, perhaps with the goal of providing a more universal perspective. All these details swirling around the central metaphor of the cascading "waters of March" can give the impression of the passing of daily life and its continual, inevitable progression towards death, just as the rains of March mark the end of a Brazilian summer. Both sets of lyrics speak of "the promise of life," perhaps allowing for other, more life-affirming interpretations, and the English contains the additional phrases "the joy
|
https://en.wikipedia.org/wiki/A.%20O.%20L.%20Atkin
|
Arthur Oliver Lonsdale Atkin (31 July 1925 – 28 December 2008), who published under the name A. O. L. Atkin, was a British mathematician.
As an undergraduate during World War II, Atkin worked at Bletchley Park cracking German codes. He received his Ph.D. in 1952 from the University of Cambridge, where he was one of John Littlewood's research students. During 1964–1970, he worked at the Atlas Computer Laboratory at Chilton, computing modular functions. Toward the end of his life, he was Professor Emeritus of mathematics at the University of Illinois at Chicago.
Atkin, along with Noam Elkies, extended Schoof's algorithm to create the Schoof–Elkies–Atkin algorithm. Together with Daniel J. Bernstein, he developed the sieve of Atkin.
Atkin is also known for his work on properties of the integer partition function and the monster module. He was a vocal fan of using computers in mathematics, so long as the end goal was theoretical advance: "Each new generation of machines makes feasible a whole new range of computations; provided mathematicians pursue these rather than merely break old records for old sports, computation will have a significant part to play in the development of mathematics."
Atkin died of nosocomial pneumonia on 28 December 2008, in Maywood, Illinois.
Selected publications
Atkin, A. O. L. and Morain, F. "Elliptic Curves and Primality Proving." Math. Comput. 61, 29–68, 1993.
Atkin, A. O. L. and Bernstein, D. J. Prime sieves using binary quadratic forms, Math. Comp. 73 (2004), 1023–1030..
See also
Atkin–Goldwasser–Kilian–Morain certificates
Atkin–Lehner theory
Elliptic curve primality proving
|
https://en.wikipedia.org/wiki/Biophysical%20Society
|
The Biophysical Society is an international scientific society whose purpose is to lead the development and dissemination of knowledge in biophysics. Founded in 1958, the Society currently consists of over 7,500 members in academia, government, and industry. Although the Society is based in the United States, it is an international organization. Overseas members currently comprise over one third of the total.
Origins
The Biophysical Society was founded in response to the growth of the field of biophysics after World War Two, as well as concerns that the American Physiological Society had become too large to serve the community of biophysicists. Discussions between prominent biophysicists in 1955 and 1956 led to the planning of the society's first meeting in Columbus, Ohio in 1957, with about 500 attendees. Among the scientists involved in the early effort were Ernest C. Pollard, Samuel Talbot, Otto Schmitt, Kenneth Stewart Cole, W. A. Selle, Max Lauffer, Ralph Stacy, Herman P. Schwan, and Robley C. Williams. This meeting was described by Cole as "a biophysics meeting with the ulterior motive of finding out if there was such a thing as biophysics and, if so, what sort of thing this biophysics might be."
Organization
The Biophysical Society is governed by four officers: the President, President-elect, Past-President Secretary, and Treasurer, as well as by a Council of twelve members in addition to the officers. These offices are elected by the membership of the society. The Council appoints an executive officer to oversee the functions and staff of the society. The society has a number of committees that help to implement its mission. The committees are: Awards, Early Careers, Education, Finance, Member Services, Membership, Committee for Inclusion and Diversity, Nominating, Professional Opportunities for Women, Program, Public Affairs, Publications, and Thematic Meetings.
The Biophysical Society also supports subgroups focusing on smaller areas within biophy
|
https://en.wikipedia.org/wiki/Rotational%E2%80%93vibrational%20coupling
|
In physics, rotational–vibrational coupling occurs when the rotation frequency of a system is close to or identical to a natural internal vibration frequency. The animation on the right shows ideal motion, with the force exerted by the spring and the distance from the center of rotation increasing together linearly with no friction.
In rotational-vibrational coupling, angular velocity oscillates. By pulling the circling masses closer together, the spring transfers its stored strain energy into the kinetic energy of the circling masses, increasing their angular velocity. The spring cannot bring the circling masses together, since the spring's pull weakens as the circling masses approach. At some point, the increasing angular velocity of the circling masses overcomes the pull of the spring, causing the circling masses to increasingly distance themselves. This increasingly strains the spring, strengthening its pull and causing the circling masses to transfer their kinetic energy into the spring's strain energy, thereby decreasing the circling masses' angular velocity. At some point, the pull of the spring overcomes the angular velocity of the circling masses, restarting the cycle.
In helicopter design, helicopters must incorporate damping devices, because at specific angular velocities, the rotorblade vibrations can reinforced themselves by rotational-vibrational coupling, and build up catastrophically. Without damping, these vibrations would cause the rotorblades to break loose.
Energy conversions
The animation on the right provides a clearer view on the oscillation of the angular velocity. There is a close analogy with harmonic oscillation.
When a harmonic oscillation is at its midpoint then all the energy of the system is kinetic energy. When the harmonic oscillation is at the points furthest away from the midpoint all the energy of the system is potential energy. The energy of the system is oscillating back and forth between kinetic energy and potential ener
|
https://en.wikipedia.org/wiki/Copper%28II%29%20arsenate
|
Copper arsenate (Cu3(AsO4)2·4H2O, or Cu5H2(AsO4)4·2H2O), also called copper orthoarsenate, tricopper arsenate, cupric arsenate, or tricopper orthoarsenate, is a blue or bluish-green powder insoluble in water and alcohol and soluble in aqueous ammonium and dilute acids. Its CAS number is or .
Uses
Copper arsenate is an insecticide used in agriculture. It is also used as a herbicide, fungicide, and a rodenticide. It is also used as a poison in slug baits.
Copper arsenate can also be a misnomer for copper arsenite, especially when meant as a pigment.
Natural occurrences
Anhydrous copper arsenate, Cu3(AsO4)2, is found in nature as the mineral lammerite. Copper arsenate tetrahydrate, Cu3(AsO4)2·4H2O, occurs naturally as the mineral rollandite.
Related compounds
Copper arsenate hydroxide or basic copper arsenate (Cu(OH)AsO4) is a basic variant with CAS number . It is found naturally as the mineral olivenite. It is used as an insecticide, fungicide, and miticide. Its use is banned in Thailand since 2001.
See also
Lead arsenate
Calcium arsenate
Paris Green (copper acetoarsenite)
Chromated copper arsenate
Scheele's Green (copper arsenite)
|
https://en.wikipedia.org/wiki/Monosodium%20methyl%20arsonate
|
Monosodium methyl arsenate (MSMA) is an arsenic-based herbicide. It is an organo-arsenate; less toxic than the inorganic form of arsenates. However, the EPA states that all forms of arsenic are a serious risk to human health and the United States' Agency for Toxic Substances and Disease Registry ranked arsenic as number 1 in its 2001 Priority List of Hazardous Substances at Superfund sites.
Arsenic is classified as a Group-A carcinogen. The EPA states that:
Trade names include:
Target 6 Plus
Target 6.6
MSMA 6 Plus
MSMA 6.6
|
https://en.wikipedia.org/wiki/Puddle
|
A puddle is a small accumulation of liquid, usually water, on a surface. It can form either by pooling in a depression on the surface, or by surface tension upon a flat surface. Puddles are often characterized by murky water or mud due to the disturbance and dissolving of surrounding sediment, primarily due to precipitation.
A puddle is generally shallow enough to walk through, and too small to traverse with a boat or raft. Small wildlife may be attracted to puddles.
Natural puddles and wildlife
Puddles in natural landscapes and habitats, when not resulting from precipitation, can indicate the presence of a seep or spring. Small seasonal riparian plants, grasses, and wildflowers can germinate with the ephemeral "head start" of moisture provided by a puddle.
Small wildlife, such as birds and insects, can use puddles as a source of essential moisture or for bathing. Raised constructed puddles, bird baths, are a part of domestic and wildlife gardens as a garden ornament and "micro-habitat" restoration. Swallows use the damp loam which gathers in puddles as a form of cement to help to build their nests. Many butterfly species and some other insects, but particularly male butterflies, need puddles for nutrients they can contain, such as salts and amino acids. In a behaviour known as puddling they seek out the damp mud that can be found around the edge of the puddles.
For some smaller forms of life, such as tadpoles or mosquito larvae, a puddle can form an entire habitat. Puddles that do not evaporate quickly can become standing water, which can become polluted by decaying organisms and are often home to breeding mosquitos, which can act as vectors for diseases such as malaria and, of more recent concern in certain areas of the world, West Nile virus.
Puddles on roads
Puddles commonly form during rain, and can cause problems for transport. Due to the angle of the road, puddles tend to be forced by gravity to gather on the edges of the road. This can cause splashi
|
https://en.wikipedia.org/wiki/-ase
|
The suffix -ase is used in biochemistry to form names of enzymes. The most common way to name enzymes is to add this suffix onto the end of the substrate, e.g. an enzyme that breaks down peroxides may be called peroxidase; the enzyme that produces telomeres is called telomerase. Sometimes enzymes are named for the function they perform, rather than substrate, e.g. the enzyme that polymerizes (assembles) DNA into strands is called polymerase; see also reverse transcriptase.
Etymology
The -ase suffix is a libfix derived from "diastase", the first recognized enzyme. Its usage in subsequently discovered enzymes was proposed by Émile Duclaux, with the intention of honoring the first scientists to isolate diastase.
See also
Amylase
DNA polymerase
|
https://en.wikipedia.org/wiki/Asymmetric%20crying%20facies
|
Asymmetric crying facies (ACF), also called partial unilateral facial paresis and hypoplasia of depressor angula oris muscle, is a minor congenital anomaly caused by agenesis or hypoplasia of the depressor anguli oris muscle, one of the muscles that control the movements of the lower lip. This unilateral facial weakness is first noticed when the infant cries or smiles, affecting only one corner of the mouth and occurs on the left side in nearly 80% of cases.
When the hypoplasia of the depressor anguli oris muscle is associated with congenital cardiac defects, the term 'Cayler cardiofacial syndrome' is used. Cayler syndrome is part of 22q11.2 deletion syndrome. It was characterized by Cayler in 1969.
|
https://en.wikipedia.org/wiki/Active%20networking
|
Active networking is a communication pattern that allows packets flowing through a telecommunications network to dynamically modify the operation of the network.
Active network architecture is composed of execution environments (similar to a unix shell that can execute active packets), a node operating system capable of supporting one or more execution environments.
It also consists of active hardware, capable of routing or switching as well as executing code within active packets.
This differs from the traditional network architecture which seeks robustness and stability by attempting to remove complexity and the ability to change its fundamental operation from underlying network components. Network processors are one means of implementing active networking concepts. Active networks have also been implemented as overlay networks.
What does it offer?
Active networking allows the possibility of highly tailored and rapid "real-time" changes to the underlying network operation.
This enables such ideas as sending code along with packets of information allowing the data to change its form (code) to match the channel characteristics.
The smallest program that can generate a sequence of data can be found in the definition of Kolmogorov complexity.
The use of real-time genetic algorithms within the network to compose network services is also enabled by active networking.
How it relates to other networking paradigms
Active networking relates to other networking paradigms primarily based upon how computing and communication are partitioned in the architecture.
Active networking and software-defined networking
Active networking is an approach to network architecture with in-network programmability. The name derives from a comparison with network approaches advocating minimization of in-network processing, based on design advice such as the "end-to-end argument". Two major approaches were conceived: programmable network elements ("switches") and capsules, a programmabi
|
https://en.wikipedia.org/wiki/Typographical%20Number%20Theory
|
Typographical Number Theory (TNT) is a formal axiomatic system describing the natural numbers that appears in Douglas Hofstadter's book Gödel, Escher, Bach. It is an implementation of Peano arithmetic that Hofstadter uses to help explain Gödel's incompleteness theorems.
Like any system implementing the Peano axioms, TNT is capable of referring to itself (it is self-referential).
Numerals
TNT does not use a distinct symbol for each natural number. Instead it makes use of a simple, uniform way of giving a compound symbol to each natural number:
{|
| zero
| align=right | 0
|-
| one
| align=right | S0
|-
| two
| align=right | SS0
|-
| three
| align=right | SSS0
|-
| four
| align=right | SSSS0
|-
| five
| align=right | SSSSS0
|}
The symbol S can be interpreted as "the successor of", or "the number after". Since this is, however, a number theory, such interpretations are useful, but not strict. It cannot be said that because four is the successor of three that four is SSSS0, but rather that since three is the successor of two, which is the successor of one, which is the successor of zero, which has been described as 0, four can be "proved" to be SSSS0. TNT is designed such that everything must be proven before it can be said to be true.
Variables
In order to refer to unspecified terms, TNT makes use of five variables. These are
a, b, c, d, e.
More variables can be constructed by adding the prime symbol after them; for example,
a, b, c, a, a‴ are all variables.
In the more rigid version of TNT, known as "austere" TNT, only
a, a, a‴ etc. are used.
Operators
Addition and multiplication of numerals
In Typographical Number Theory, the usual symbols of "+" for additions, and "·" for multiplications are used. Thus to write "b plus c" is to write
(b + c)
and "a times d" is written as
(a·d)
The parentheses are required. Any laxness would violate TNT's formation system (although it is trivially proved this formalism is unnecessary for operations which are both c
|
https://en.wikipedia.org/wiki/Chromated%20copper%20arsenate
|
Chromated copper arsenate (CCA) is a wood preservative containing compounds of chromium, copper, and arsenic, in various proportions. It is used to impregnate timber and other wood products, especially those intended for outdoor use, in order to protect them from attack by microbes and insects. Like other copper-based wood preservatives, it imparts a greenish tint to treated timber.
CCA was invented in 1933 by Indian chemist Sonti Kamesam, and patented in Britain in 1934. It has been used for timber treatment since the mid-1930s, and is marketed under many trade names.
In 2003, the United States Environmental Protection Agency and the lumber industry agreed to discontinue the use of CCA-treated wood in most residential construction. This agreement was intended to protect the health of humans and the environment by reducing exposure to the arsenic in CCA-treated wood. As a result of this decision, CCA-treated wood can no longer be used to construct residential structures such as playground equipment, decks, picnic tables, landscaping features, fences, patios, and walkways. Acute intoxication due to mishandling of treated products, e.g. by burning, is also a serious concern. Nevertheless, CCA remains a popular and economical option to make perishable timbers, such as plantation-grown pine, viable for applications like poles, piling, retaining structures, etc.
Composition and application
The composition of CCA products is usually described in terms of the mass percentages of chromium trioxide or "chromic acid" , arsenic pentoxide , and copper(II) oxide .
The preservative is applied as a water-based mixture containing 0.6–6.0% (by weight) of chromic acid, copper oxide, and arsenic acid (USDA, 1980), with pH 1.6–2.5. The mixture is infused into wood at high pressure.
In the treated wood, arsenic is believed to be in the form of chromium (III) arsenate and/or copper(II) arsenate , or fairly stable chromium dimer-arsenic clusters.
Mechanism of action
The chromi
|
https://en.wikipedia.org/wiki/Annuities%20in%20the%20European%20Union
|
Under European Union law, an annuity is a financial contract which provides an income stream in return for an initial payment with specific parameters. It is the opposite of a settlement funding. A Swiss annuity is not considered a European annuity for tax reasons.
Immediate annuity
An immediate annuity is an annuity for which the time between the contract date and the date of the first payment is not longer than the time interval between payments. A common use for an immediate annuity is to provide a pension to a retired person or persons.
It is a financial contract which makes a series of payments with certain characteristics:
either level or fluctuating periodical payments
made annually, or at more frequent intervals
in advance or arrears
duration may be:
fixed (annuity certain)
during the lifetime or one or more persons, possibly reduced after death of one person
during the lifetime but not longer than a maximum number of years
during the lifetime but not shorter than a minimum number of years
Annuity certain
An annuity certain pays the annuitant for a number of years designated. This option is not suitable for retirement income, as the person may outlive the number of years the annuity will pay.
Life annuity
A life annuity or lifetime immediate annuity is most often used to provide an income in old age (i.e., a pension). This type of annuity may be purchased from an insurance (Ireland and the UK, Life Assurance) company.
This annuity can be compared to a loan which is made by the purchaser to the issuing company, who then pay back the original capital with interest to the annuitant on whose life the annuity is based. The assumed period of the loan is based on the life expectancy of the annuitant but life annuities are payable until the death of the last surviving annuitant. In order to guarantee that the income continues for life, the investment relies on cross-subsidy. Because an annuity population can be expected to have a distribution of lifespans
|
https://en.wikipedia.org/wiki/Blue%20diaper%20syndrome
|
Blue diaper syndrome is a rare, autosomal recessive or X linked recessive metabolic disorder characterized in infants by bluish urine-stained diapers. It is also known as Drummond's syndrome, and hypercalcemia.
It is caused by a defect in tryptophan absorption. Bacterial degradation of unabsorbed tryptophan in the intestine leads to excessive indole production and thus to indicanuria which, on oxidation to indigo blue, causes a peculiar bluish discoloration of the diaper (indoluria). Symptoms typically include digestive disturbances, fever and visual problems. Some may also develop disease due to the incomplete breakdown of tryptophan.
It was characterized in 1964, and inherited in an autosomal recessive pattern although X-linked recessive inheritance has not been completely ruled out since reported patients have been male.
Since this syndrome is X linked, the chance for a child to receive normal genes from both parents and be genetically normal for that particular trait is 25%. If an individual receives one normal gene and one gene for the disease, the person will be a carrier for the disease, but usually will not show symptoms. Carrier females usually do not display symptoms of the disorder because it is usually the X chromosome with the abnormal gene that is “turned off”. Parents can undergo genetic testing to see if their child will get this syndrome, but most do not find out until they see the symptoms mentioned below.
Signs and symptoms
The signs and symptoms of blue diaper syndrome may include irritability, constipation, poor appetite, vomiting, and poor growth. Some children experience frequent fevers and intestinal infections.
Hypercalcemia could be a potential issue in affected children. Some children with blue diaper syndrome have eye or vision issues, particularly underdeveloped portions of the eye, including the cornea and optic disc.
Genetics
Blue diaper syndrome affects males and females equally. The number of people affected in the general popu
|
https://en.wikipedia.org/wiki/CoDeeN
|
CoDeeN is a proxy server system created at Princeton University in 2003 and deployed for general use on PlanetLab.
It operates as per the following:
Users set their internet caches to a nearby high bandwidth proxy that participates in the system.
Requests to that proxy are then forwarded to an appropriate member of the system that is in charge of the file (should be caching it) and that has sent recent updates showing that it is still alive. The file is forwarded to the proxy and thence to the client.
What this means for normal users is that if they use this and a server is slow, however the content is cached on the system, then (after the first upload) requests to that file will be fast. It also means that the request will not be satisfied by the original server, equivalent to free bandwidth.
For rare files this system could be slightly slower than downloading the file itself. The system's speed is also subject to the constraint of number of participating proxies.
For the case of large files requested by many peers, it uses a kind of 'multi-cast stream' from one peer to the others, which then distribute out to their respective proxies.
CoBlitz, a CDN technology firm (2006–2009), was a take-off of this, in that files are not saved in the web cache of a single member of the proxy-system, but are instead saved piece-wise across several members, and 'gathered up' when they are requested. This allows for more sharing of disk space among proxies, and for higher fault tolerance. To access this system, URLs were prefixed with http://coblitz.codeen.org/. Verivue Inc. acquired CoBlitz in October 2010.
|
https://en.wikipedia.org/wiki/Panniculus%20adiposus
|
The panniculus adiposus is the fatty layer of the subcutaneous tissues, superficial to a deeper vestigial layer of muscle, the panniculus carnosus.
It includes structures that are considered fascia by some sources but not by others. Some examples include the fascia of Camper and the superficial cervical fascia.
A group of disorders of inflammation of this layer is called panniculitis.
|
https://en.wikipedia.org/wiki/Familial%20male-limited%20precocious%20puberty
|
Familial male-limited precocious puberty, often abbreviated as FMPP, also known as familial sexual precocity or gonadotropin-independent testotoxicosis, is a form of gonadotropin-independent precocious puberty in which boys experience early onset and progression of puberty. Signs of puberty can develop as early as an age of 1 year.
The spinal length in boys may be short due to a rapid advance in epiphyseal maturation. It is an autosomal dominant condition with a mutation of the luteinizing hormone (LH) receptor. As FMPP is a gonadotropin-independent form of precocious puberty, gonadotropin-releasing hormone agonists (GnRH agonists) are ineffective. Treatment is with drugs that suppress or block the effects of gonadal steroidogenesis, such as cyproterone acetate, ketoconazole, spironolactone, and testolactone. Alternatively, the combination of the androgen receptor antagonist bicalutamide and the aromatase inhibitor anastrozole may be used.
Robert King Stone, personal physician to American president Abraham Lincoln, described the first case of FMPP in 1852.
See also
Follicle-stimulating hormone insensitivity
Gonadotropin-releasing hormone insensitivity
Hypergonadism, hyperandrogenism, and precocious puberty
Inborn errors of steroid metabolism
Leydig cell hypoplasia (or LH insensitivity)
|
https://en.wikipedia.org/wiki/European%20Society%20of%20Gene%20and%20Cell%20Therapy
|
European Society of Gene & Cell Therapy (ESGCT) formerly European Society of Gene Therapy (ESGT) is a legally registered professional body which emerged from a small working group in 1992 that focused on human gene therapy.
The objectives of the ESGT include the following:
promote basic and clinical research in gene therapy;
facilitate education (and the exchange of information and technologies) related to gene transfer and therapy;
serve as a professional adviser to the gene therapy community and various regulatory bodies in Europe.
The official journal of the ESGT is The Journal of Gene Medicine.
Collaborations
The ESGT works with other entities in the scientific communities in the event that an adverse effect to a specific gene therapy is discovered. Investigations the ESGT has been involved with include the adverse effects discovered during the French X-SCID gene therapy trial. The ESGT hosted a forum of 500 researchers from various facilities around the world, including representatives from the Stanford University and the Sloan Kettering Cancer Research Center.
External links
Official Site
Researchers from Israel, Southern California to Present Stem Cell Symposium
Researchers team up to tackle cystic fibrosis
Genetics societies
European medical and health organizations
|
https://en.wikipedia.org/wiki/American%20Society%20of%20Gene%20and%20Cell%20Therapy
|
American Society of Gene & Cell Therapy (ASGCT) is a professional non-profit medical and scientific organization based in Milwaukee, dedicated to understanding, development and application of gene, related cell and nucleic acid therapies, as well as promotion of professional and public education on gene therapy. With more than 4,800 members in the United States and worldwide, today ASGCT is the largest association of individuals involved in gene and cell therapy research.
Molecular Therapy is the official journal of the ASGCT.
|
https://en.wikipedia.org/wiki/Blunt%20trauma
|
Blunt trauma, also known as blunt force trauma or non-penetrating trauma, describes a physical trauma due to a forceful impact without penetration of the body's surface. Blunt trauma stands in contrast with penetrating trauma, which occurs when an object pierces the skin, enters body tissue, and creates an open wound. Blunt trauma occurs due to direct physical trauma or impactful force to a body part. Such incidents often occur with road traffic collisions, assaults, sports-related injuries, and are notably common among the elderly who experience falls.
Blunt trauma can lead to a wide range of injuries including contusions, concussions, abrasions, lacerations, internal or external hemorrhages, and bone fractures. The severity of these injuries depends on factors such as the force of the impact, the area of the body affected, and underlying comorbidities of the affected individual. In some cases, blunt force trauma can be life-threatening and may require immediate medical attention. Blunt trauma to the head and/or severe blood loss are the most likely causes of death due to blunt force traumatic injury.
Classification
Blunt abdominal trauma
Blunt abdominal trauma (BAT) represents 75% of all blunt trauma and is the most common example of this injury. 75% of BAT occurs in motor vehicle crashes, in which rapid deceleration may propel the driver into the steering wheel, dashboard, or seatbelt, causing contusions in less serious cases, or rupture of internal organs from briefly increased intraluminal pressure in the more serious, depending on the force applied. Initially, there may be few indications that serious internal abdominal injury has occurred, making assessment more challenging and requiring a high degree of clinical suspicion.
There are two basic physical mechanisms at play with the potential of injury to intra-abdominal organs: compression and deceleration. The former occurs from a direct blow, such as a punch, or compression against a non-yielding object
|
https://en.wikipedia.org/wiki/Moisture%20stress
|
Moisture stress is a form of abiotic stress that occurs when the moisture of plant tissues is reduced to suboptimal levels. Water stress occurs in response to atmospheric and soil water availability when the transpiration rate exceeds the rate of water uptake by the roots and cells lose turgor pressure. Moisture stress is described by two main metrics, water potential and water content.
Moisture stress has an effect on stomatal opening, mainly causing a closure in stomata as to reduce the amount of carbon dioxide assimilation. Closing of the stomata also slows the rate of transpiration, which limits water loss and helps to prevent the wilting effects of moisture stress. This closing can be trigged by the roots sensing dry soil and in response producing the hormone ABA which when transported up the xylem into the leaves will reduce stomatal conductance and wall extensibility of growing cells. This lowers the rates of transpiration, photosynthesis and leaf expansion. ABA also increases the loosening of growing root cell walls and in turn increases root growth in an effort to find water in the soil.
Phenotypic response of plants to long-term water stress was measured in corn and showed that plants respond to water stress with both an increase in root growth both laterally and vertically. In all Droughted conditions the corn showed decrease in plant height and yield due to the decrease in water availability.
Genes induced during water-stress conditions are thought to function not only in protecting cells from water deficit by the production of important metabolic proteins but also in the regulation of genes for signal transduction in the water-stress response. There are four pathways that have been described that show the plants genetic response to moisture stress; two are ABA dependent while two are ABA independent. They all affect gene expression that increases the plants water stress tolerance.
The effects of moisture stress on photosynthesis can depend as much
|
https://en.wikipedia.org/wiki/DRMAA
|
Distributed Resource Management Application API (DRMAA) is a high-level Open Grid Forum (OGF) API specification for the submission and control of jobs to a distributed resource management (DRM) system, such as a cluster or grid computing infrastructure. The scope of the API covers all the high level functionality required for applications to submit, control, and monitor jobs on execution resources in the DRM system.
In 2007, DRMAA was one of the first two (the other one was GridRPC) specifications that reached the full recommendation status in the OGF.
In 2012 the second version of the DRMAA standard (DRMAA2) was published in an abstract interface definition language (IDL) defining the semantic of the functions in GFD 194. DRMAA2 specifies more than twice as many calls as DRMAA. It covers cluster monitoring, has a notion of queues and machines, and introduces a multi job-session concept for single applications for a better job workflow management. Later in 2012 the C API was specified as a first language binding in GF 198.
Development model
The development of this API was done through the Global Grid Forum, in the model of IETF standard development, and it was originally co-authored by:
Roger Brobst from Cadence Design Systems
Waiman Chan from IBM
Fritz Ferstl from Sun Microsystems, now Univa
Jeff Gardiner from John P. Robarts Research Institute
Andreas Haas from Sun Microsystems (Co-chair)
Bill Nitzberg from Altair Engineering
Hrabri Rajic from Intel (Maintainer & Co-chair)
John Tollefsrud from Sun Microsystems Founding (chair)
This specification was first proposed at Global Grid Forum 3 (GGF3) in Frascati, Italy, but gained most of its momentum at Global Grid Forum 4 in Toronto, Ontario. The development of the specification was first proposed with the objective to facilitate direct interfacing of applications to existing DRM systems by application's builders, portal builders, and Independent Software Vendors (ISVs). Because the API was co-authored by partic
|
https://en.wikipedia.org/wiki/Harold%20Thomas%20%28artist%29
|
Harold Joseph Thomas (born 1947), also known as Bundoo, is an Aboriginal Australian artist and former activist, known for designing and copyrighting the Australian Aboriginal Flag. He claims to have designed the flag in 1971 as a symbol of the Aboriginal land rights movement, and in 1995 it was made an official "Flag of Australia". He later asserted copyright over his design, eventually transferring that copyright to the Commonwealth of Australia and making it freely available for public use in January 2022.
Early life and education
Harold Joseph Thomas (Bundoo) was born in 1947 to a Wombai father and Luritja mother in Alice Springs, Northern Territory. He remembers drawing and painting on pieces of paper from an early age. At seven years old, he was removed from his family as part of the Stolen Generations and was raised in South Australia by a foster family. He attended an Adelaide independent school, Pulteney Grammar.
In 1966, aged 17, Thomas won a scholarship to study at the South Australian School of Art, graduating with Honours in 1969, becoming the first Aboriginal person to graduate from an Australian art school. It was while he was studying that he became involved in the civil rights movement. In 1968 he held his first exhibition of watercolour paintings in Adelaide, which was opened by then premier Don Dunstan.
Art practice and career
In 1970 he was the first Aboriginal person to be employed at the South Australian Museum, working as a survey artist. This gave him access to a huge collection of Aboriginal artefacts as well as a wide range of art.
His main artistic influences include painters Caravaggio, Francisco Goya and Eugène Delacroix.
In 1972 he moved to the "Top End", and started painting the landscape and wildlife, especially watercolours. He changed his subject matter and style dramatically in 2016, when he started painting representations of Aboriginal people's first and subsequent contact with European colonisers, including the frontier war
|
https://en.wikipedia.org/wiki/CDDLM
|
CDDLM (Configuration Description, Deployment, and Lifecycle Management Specification) is a Global Grid Forum standard for the management, deployment and configuration of Grid Service lifecycles or inter-organization resources.
Structure
The specification is based on component documents;
Document that describes functional requirements, use cases, and high-level architectures, and otherwise serves as a Foundation Document
Document outlining the development of a non-XML based Configuration, Description and Deployment Language
Document outlining the development of an XML based Configuration, Description and Deployment Language
Document outlining the development of a Configuration, Description and Deployment Component Model
Development Model
The development of this API was done through the Global Grid Forum as an open standard, in the model of IETF standard development, and it was originally edited by D. Bell, T. Kojo, P. Goldsack, S. Loughran, D. Milojicic, S. Schaefer, J. Tatemura, and P. Toft.
Significance
System administration in a distributed environment with diverse hardware, software, patch level, and imposed user requirements makes the ability to deploy, manage, and describe services and software configuration difficult. Within a grid, this difficulty is complicated further by the need to have similar service end points, possibly on heterogeneous architectures. Grid service requests may require configuration changes.
This standard provided a framework which described a language and methods that have the ability to describe system configuration, and move system, services, and software towards desired configuration endpoints. Furthermore, it served as the first real attempt to address system administration issues within a grid. CDDLM is to grids, as CFEngine for servers.
|
https://en.wikipedia.org/wiki/PlanetLab
|
PlanetLab was a group of computers available as a testbed for computer networking and distributed systems research. It was established in 2002 by Prof. Larry L. Peterson and Prof. David Culler, and as of June 2010, it was composed of 1090 nodes at 507 sites worldwide. Each research project had a "slice", or virtual machine access to a subset of the nodes.
Accounts were limited to persons affiliated with corporations and universities that hosted PlanetLab nodes. However, a number of free, public services have been deployed on PlanetLab, including CoDeeN, the Coral Content Distribution Network, and Open DHT. Open DHT was taken down on 1 July 2009.
PlanetLab was officially shut down in May 2020 but continues in Europe.
|
https://en.wikipedia.org/wiki/Hidden%20subgroup%20problem
|
The hidden subgroup problem (HSP) is a topic of research in mathematics and theoretical computer science. The framework captures problems such as factoring, discrete logarithm, graph isomorphism, and the shortest vector problem. This makes it especially important in the theory of quantum computing because Shor's algorithm for factoring in quantum computing is an instance of the hidden subgroup problem for finite abelian groups, while the other problems correspond to finite groups that are not abelian.
Problem statement
Given a group , a subgroup , and a set , we say a function hides the subgroup if for all if and only if . Equivalently, is constant on the cosets of H, while it is different between the different cosets of H.
Hidden subgroup problem: Let be a group, a finite set, and a function that hides a subgroup . The function is given via an oracle, which uses bits. Using information gained from evaluations of via its oracle, determine a generating set for .
A special case is when is a group and is a group homomorphism in which case corresponds to the kernel of .
Motivation
The hidden subgroup problem is especially important in the theory of quantum computing for the following reasons.
Shor's algorithm for factoring and for finding discrete logarithms (as well as several of its extensions) relies on the ability of quantum computers to solve the HSP for finite abelian groups.
The existence of efficient quantum algorithms for HSPs for certain non-abelian groups would imply efficient quantum algorithms for two major problems: the graph isomorphism problem and certain shortest vector problems (SVPs) in lattices. More precisely, an efficient quantum algorithm for the HSP for the symmetric group would give a quantum algorithm for the graph isomorphism. An efficient quantum algorithm for the HSP for the dihedral group would give a quantum algorithm for the unique SVP.
Algorithms
There is an efficient quantum algorithm for solving HSP over finite
|
https://en.wikipedia.org/wiki/Optical%20disc%20recording%20technologies
|
Optical disc authoring requires a number of different optical disc recorder technologies working in tandem, from the optical disc media to the firmware to the control electronics of the optical disc drive.
Types of recordable optical disc
There are numerous formats of recordable optical direct to disk on the market, all of which are based on using a laser to change the reflectivity of the digital recording medium in order to duplicate the effects of the pits and lands created when a commercial optical disc is pressed. Emerging technologies such as holographic data storage and 3D optical data storage aim to use entirely different data storage methods, but these products are in development and are not yet widely available.
The earliest form is magneto-optical, which uses a magnetic field in combination with a laser to write to the medium. Though not widely used in consumer equipment, the original NeXT cube used MO media as its standard storage device, and consumer MO technology is available in the form of Sony's MiniDisc. This form of medium is rewriteable.
The most common form of recordable optical media is write-once organic dye technology, popularized in the form of the CD-R and still used for higher-capacity media such as DVD-R. This uses the laser alone to scorch a transparent organic dye (usually cyanine, phthalocyanine, or azo compound-based) to create "pits" (i.e. dark spots) over a reflective spiral groove. Most such media are designated with an R (recordable) suffix. Such discs are often quite colorful, generally coming in shades of blue or pale yellow or green.
Rewritable, non-magnetic optical media are possible using phase change alloys, which are converted between crystalline and amorphous states (with different reflectivity) using the heat from the drive laser. Such media must be played in specially tuned drives, since the phase-change material has less of a contrast in reflectivity than dye-based media; while most modern drives support such media,
|
https://en.wikipedia.org/wiki/Variable%20%28mathematics%29
|
In mathematics, a variable (from Latin variabilis, "changeable") is a symbol that represents a mathematical object. A variable may represent a number, a vector, a matrix, a function, the argument of a function, a set, or an element of a set.
Algebraic computations with variables as if they were explicit numbers solve a range of problems in a single computation. For example, the quadratic formula solves any quadratic equation by substituting the numeric values of the coefficients of that equation for the variables that represent them in the quadratic formula. In mathematical logic, a variable is either a symbol representing an unspecified term of the theory (a meta-variable), or a basic object of the theory that is manipulated without referring to its possible intuitive interpretation.
History
In ancient works such as Euclid's Elements, single letters refer to geometric points and shapes. In the 7th century, Brahmagupta used different colours to represent the unknowns in algebraic equations in the Brāhmasphuṭasiddhānta. One section of this book is called "Equations of Several Colours".
At the end of the 16th century, François Viète introduced the idea of representing known and unknown numbers by letters, nowadays called variables, and the idea of computing with them as if they were numbers—in order to obtain the result by a simple replacement. Viète's convention was to use consonants for known values, and vowels for unknowns.
In 1637, René Descartes "invented the convention of representing unknowns in equations by x, y, and z, and knowns by a, b, and c". Contrarily to Viète's convention, Descartes' is still commonly in use. The history of the letter x in math was discussed in a 1887 Scientific American article.
Starting in the 1660s, Isaac Newton and Gottfried Wilhelm Leibniz independently developed the infinitesimal calculus, which essentially consists of studying how an infinitesimal variation of a variable quantity induces a corresponding variation of anothe
|
https://en.wikipedia.org/wiki/Optical%20disc%20recording%20modes
|
In optical disc authoring, there are multiple modes for recording, including Disc-At-Once, Track-At-Once, and Session-At-Once.
CD Disc-At-Once
Disc-At-Once (DAO) for CD-R media is a mode that masters the disc contents in one pass, rather than a track at a time as in Track At Once. DAO mode, unlike TAO mode, allows any amount of audio data (or no data at all) to be written in the "pre-gaps" between tracks.
One use of this technique, for example, is to burn track introductions to be played before each track starts. A CD player will generally display a negative time offset counting up to the next track when such pre-gap introductions play. Pre-gap audio before the first track of the CD makes it possible to burn an unnumbered, "hidden" audio track. This track can only be accessed by "rewinding" from the start of the first track, backwards into the pre-gap audio.
DAO recording is also the only way to write data to the unused R-W sub-channels. This allows for extended graphic and text features on an audio CD such as CD+G and CD-Text. It is also the only way to write audio files that link together seamlessly with no gaps, a technique often used in progressive rock, trance and other music genres.
CD Track-At-Once
Track-At-Once (TAO) is a recording mode where the recording laser stops after each track is finished and two run-out blocks are written. One link block and four run-in blocks are written when the next track is recorded. TAO discs can have both data and audio at the same time.
There are 2 TAO writing modes
Mode 1
Mode 2 XA
DVD-R Disc At Once
Disc-At-Once (DAO) recording for DVD-R media is a mode in which all data is written sequentially to the disc in one uninterrupted recording session. The on-disk contents result in a lead-in area, followed by the data, and closed by a lead-out area. The data is addressable in sectors of 2048 bytes each, with the first sector address being zero. There are no run-out blocks as in CD-R disc-at-once.
Session At Once
|
https://en.wikipedia.org/wiki/Formal%20calculation
|
In mathematical logic, a formal calculation, or formal operation, is a calculation that is systematic but without a rigorous justification. It involves manipulating symbols in an expression using a generic substitution without proving that the necessary conditions hold. Essentially, it involves the form of an expression without considering its underlying meaning. This reasoning can either serve as positive evidence that some statement is true when it is difficult or unnecessary to provide proof or as an inspiration for the creation of new (completely rigorous) definitions.
However, this interpretation of the term formal is not universally accepted, and some consider it to mean quite the opposite: a completely rigorous argument, as in formal mathematical logic.
Examples
Formal calculations can lead to results that are wrong in one context, but correct in another context. The equation
holds if q has an absolute value less than 1. Ignoring this restriction, and substituting q = 2 to leads to
Substituting q=2 into the proof of the first equation, yields a formal calculation that produces the last equation. But it is wrong over the real numbers, since the series does not converge. However, in other contexts (e.g. working with 2-adic numbers, or with integers modulo a power of 2), the series does converge. The formal calculation implies that the last equation must be valid in those contexts.
Another example is obtained by substituting q=-1. The resulting series 1-1+1-1+... is divergent (over the real and the p-adic numbers) but a value can be assigned to it with an alternative method of summation, such as Cesàro summation. The resulting value, 1/2, is the same as that obtained by the formal computation.
Formal power series
Formal power series is a concept that adopts the form of power series from real analysis. The word "formal" indicates that the series need not converge. In mathematics, and especially in algebra, a formal series is an infinite sum that is consider
|
https://en.wikipedia.org/wiki/GPnotebook
|
GPnotebook is a British medical database for general practitioners (GPs). It is an online encyclopaedia of medicine that provides an immediate reference resource for clinicians worldwide. The database consists of over 30,000 index terms and over two million words of information. GPnotebook is provided online by Oxbridge Solutions Limited.
GPnotebook website is primarily designed with the needs of general practitioners (GPs) in mind, and written by a variety of specialists, ranging from paediatrics to accident and emergency.
The original idea for the database began in the canteen of John Radcliffe Hospital in 1990 while James McMorran, a first-year Oxford University clinical student, was writing up his medical notes. Instead of writing notes in longhand, he wrote his notes in ‘mind maps’ of packets of information linking different concepts and conditions in a two-dimensional representation of clinical knowledge. James discussed with Stewart McMorran (then a medical student at Cambridge University and a talented computer programmer) this way of representing medical knowledge and between them they created the authoring software to produce linking ‘packets’ of information in a database. This first authoring software and database was the origin of what today is GPnotebook. It was, in effect, a medical ‘Wiki’ over 16 years before the first ‘Wiki’!
Initially, James used the authoring software alone to capture his own clinical learning. There was interest from other medical students at Oxford and in the end a team of six authors (mainly Oxford medical students) became the founding (and continuing) principal authors of GPnotebook. Among them was Damian Crowther who, in time, took over the role of technical lead for the project. James takes the role of editorial lead for the website. Damian developed the software for the web version of the database which was released on the worldwide web in 2001 as GPnotebook.
GPnotebook is used within consultation by general practitioner
|
https://en.wikipedia.org/wiki/Mdm2
|
Mouse double minute 2 homolog (MDM2) also known as E3 ubiquitin-protein ligase Mdm2 is a protein that in humans is encoded by the MDM2 gene. Mdm2 is an important negative regulator of the p53 tumor suppressor. Mdm2 protein functions both as an E3 ubiquitin ligase that recognizes the N-terminal trans-activation domain (TAD) of the p53 tumor suppressor and as an inhibitor of p53 transcriptional activation.
Discovery and expression in tumor cells
The murine double minute (mdm2) oncogene, which codes for the Mdm2 protein, was originally cloned, along with two other genes (mdm1 and mdm3) from the transformed mouse cell line 3T3-DM. Mdm2 overexpression, in cooperation with oncogenic Ras, promotes transformation of primary rodent fibroblasts, and mdm2 expression led to tumor formation in nude mice. The human homologue of this protein was later identified and is sometimes called Hdm2. Further supporting the role of mdm2 as an oncogene, several human tumor types have been shown to have increased levels of Mdm2, including soft tissue sarcomas and osteosarcomas as well as breast tumors. The MDM2 oncoprotein ubiquitinates and antagonizes p53 but may also carry out p53-independent functions. MDM2 supports the Polycomb-mediated repression of lineage-specific genes, independent of p53. MDM2 depletion in the absence of p53 promoted the differentiation of human mesenchymal stem cells and diminished clonogenic survival of cancer cells. Most of the MDM2-controlled genes also responded to the inactivation of the Polycomb Repressor Complex 2 (PRC2) and its catalytic component EZH2. MDM2 physically associated with EZH2 on chromatin, enhancing the trimethylation of histone 3 at lysine 27 (H3K27me3) and the ubiquitination of histone 2A at lysine 119 (H2AK119) at its target genes. Removing MDM2 simultaneously with the H2AK119 E3 ligase Ring1B/RNF2 further induced these genes and synthetically arrested cell proliferation.
An additional Mdm2 family member, Mdm4 (also called MdmX), has be
|
https://en.wikipedia.org/wiki/Rientrodolce
|
Rientrodolce is an Italian association linked to Radicali Italiani, which concerns itself with overpopulation, natural environment and energy.
Its name comes from Marco Pannella's idea of a "mild return" ("rientro dolce" in Italian) to a world with 2 billion human beings. The association considers overpopulation the primary cause of the humanitarian, environmental and energetic crisis of our planet. It aims to inform and convince the society, the mass media, the intellectuals and the politicians of the necessity, along with the other environmental measures, of a reduction of the world population, in full respect of human rights and individual liberties.
The association has a discussion group, which is not limited to the association's members but it is open to everybody.
External links
Rientrodolce Association
Environmental organisations based in Italy
Demography
Political associations of Italy
|
https://en.wikipedia.org/wiki/Fault%20coverage
|
Fault coverage refers to the percentage of some type of fault that can be detected during the test of any engineered system. High fault coverage is particularly valuable during manufacturing test, and techniques such as Design For Test (DFT) and automatic test pattern generation are used to increase it.
In electronics for example, stuck-at fault coverage is measured by sticking each pin of the hardware model at logic '0' and logic '1', respectively, and running the test vectors. If at least one of the outputs differs from what is to be expected, the fault is said to be detected. Conceptually, the total number of simulation runs is twice the number of pins (since each pin is stuck in one of two ways, and both faults should be detected). However, there are many optimizations that can reduce the needed computation. In particular, often many non-interacting faults can be simulated in one run, and each simulation can be terminated as soon as a fault is detected.
A fault coverage test passes when at least a specified percentage of all possible faults can be detected. If it does not pass, at least three options are possible. First, the designer can augment or otherwise improve the vector set, perhaps by using a more effective automatic test pattern generation tool. Second, the circuit may be re-defined for better fault detectibility (improved controllability and observability). Third, the designer may simply accept the lower coverage.
Test coverage (computing)
The term test coverage used in the context of programming / software engineering, refers to measuring how much a software program has been exercised by tests. Coverage is a means of determining the rigour with which the question underlying the test has been answered. There are many kinds of test coverage:
code coverage
feature coverage,
scenario coverage,
screen item coverage,
requirements coverage,
model coverage.
Each of these coverage types assumes that some kind of baseline exists which defin
|
https://en.wikipedia.org/wiki/501%20%28number%29
|
501 (five hundred [and] one) is the natural number following 500 and preceding 502.
501 is the sum of the first eighteen primes.
There are 501 degree-8 polynomials with integer coefficients, all of whose roots are in the unit disk.
There are 501 ways of partitioning the digits from 0 to 9 into two sets, each of which contains at least two digits, and 501 ways of partitioning a set of five elements into any number of ordered sequences.
501 is also a figurate number based on the 5-orthoplex or 5-dimensional cross polytope.
In the gematria of Eleazar of Worms, the Hebrew words "temunah" (image) and "parsuf 'adam" (human face) both had the numerological value of 501. Eleazar used this equivalence to argue that, in several Biblical passages, God appeared to His prophets in the form of a human face.
Other uses
501 is commonly used to refer to people deported from Australia under section 501 of the 1958 Migration Act.
|
https://en.wikipedia.org/wiki/Endocast
|
An endocast is the internal cast of a hollow object, often referring to the cranial vault in the study of brain development in humans and other organisms. Endocasts can be artificially made for examining the properties of a hollow, inaccessible space, or they may occur naturally through fossilization.
Cranial endocasts
Artificial casts
Endocasts of the inside of the neurocranium (braincase) are often made in paleoanthropology to study brain structures and hemispheric specialization in extinct human ancestors. While an endocast can not directly reveal brain structure, it can allow scientists to gauge the size of areas of the brain situated close to the surface, notably Wernicke's and Broca's areas, responsible for interpreting and producing speech.
Traditionally, the casting material is some form of rubber or rubber-like material. The openings to the brain cavity, except for the foramen magnum, are closed, and the liquid rubber is slushed around in the empty cranial vault and then left to set. The resulting hollow sphere can then be drained of air like a balloon and pulled out through the foramen magnum. Rubber endocasts like these were the standard practice until the end of the 20th century and are still used in some fields. However, scientists are increasingly utilizing computerized tomography scanning technology to create digital endocasts in order to avoid risking damage to valuable specimens.
Natural endocasts
Natural cranial endocasts are also known. The famous Taung Child, the first Australopithecus found, consists of a natural endocast connected to the facial portion of the skull. It was the shape of the brain that allowed Raymond Dart to conclude that the fossil was that of a human relative rather than an extinct ape.
Mammal endocasts are particularly useful, as they resemble the fresh brain with the dura mater in place. Such "fossil brains" are known from several hundred different mammal species. More than a hundred natural casts of the cranial vaul
|
https://en.wikipedia.org/wiki/Quantum%20mirage
|
In physics, a quantum mirage is a peculiar result in quantum chaos. Every system of quantum dynamical billiards will exhibit an effect called scarring, where the quantum probability density shows traces of the paths a classical billiard ball would take. For an elliptical arena, the scarring is particularly pronounced at the foci, as this is the region where many classical trajectories converge. The scars at the foci are colloquially referred to as the "quantum mirage".
The quantum mirage was first experimentally observed by Hari Manoharan, Christopher Lutz and Donald Eigler at the IBM Almaden Research Center in San Jose, California in 2000. The effect is quite remarkable but in general agreement with prior work on the quantum mechanics of dynamical billiards in elliptical arenas.
Quantum corral
The mirage occurs at the foci of a quantum corral, a ring of atoms arranged in an arbitrary shape on a substrate. The quantum corral was demonstrated in 1993 by Lutz, Eigler, and Crommie using an elliptical ring of iron atoms on a copper surface using the tip of a low-temperature scanning tunneling microscope to manipulate individual atoms. The ferromagnetic iron atoms reflected the surface electrons of the copper inside the ring into a wave pattern, as predicted by the theory of quantum mechanics.
Quantum corrals can be viewed as artificial atoms that even show similar chemical bonding properties as real atoms.
The size and shape of the corral determine its quantum states, including the energy and distribution of the electrons. To make conditions suitable for the mirage the team at Almaden chose a configuration of the corral which concentrated the electrons at the foci of the ellipse.
When scientists placed a magnetic cobalt atom at one focus of the corral, a mirage of the atom appeared at the other focus. Specifically the same electronic properties were present in the electrons surrounding both foci, even though the cobalt atom was only present at one focus. In sca
|
https://en.wikipedia.org/wiki/Shelter%20Island%20Conference
|
The first Shelter Island Conference on the Foundations of Quantum Mechanics was held from June 2–4, 1947 at the Ram's Head Inn in Shelter Island, New York. Shelter Island was the first major opportunity since Pearl Harbor and the Manhattan Project for the leaders of the American physics community to gather after the war. As Julian Schwinger would later recall, "It was the first time that people who had all this physics pent up in them for five years could talk to each other without somebody peering over their shoulders and saying, 'Is this cleared?'"
The conference, which cost $850, was followed by the Pocono Conference of 1948 and the Oldstone Conference of 1949. They were arranged with the assistance of J. Robert Oppenheimer and the National Academy of Sciences (NAS). Later Oppenheimer deemed Shelter Island the most successful scientific meeting he had ever attended; and as Richard Feynman recalled to Jagdish Mehra in April 1970: "There have been many conferences in the world since, but I've never felt any to be as important as this.... The Shelter Island Conference was my first conference with the big men.... I had never gone to one like this in peacetime."
Organization
The conference was conceived by Duncan A. MacInnes, a electrochemistry researcher at the Rockefeller Institute for Medical Research. Once the president of the New York Academy of Sciences, MacInnes had already organized a number of small scientific conferences. However, he believed that the later conferences had suffered from a bloated attendance, and over this issue, he resigned from the academy in January 1945. That fall, he approached the NAS with the idea of a series of 2–3 day conferences limited to 20–25 people. Frank Jewett, the head of the NAS, liked the idea; he envisioned a "meeting at some quiet place where the men could live together intimately", possibly "at an inn somewhere", and suggested that MacInnes focus on a couple of pilot programs. MacInnes' first choice was "The Nature of
|
https://en.wikipedia.org/wiki/New%20York%20Wine%20Tasting%20of%201973
|
The New York Wine Tasting of 1973 was organized by pioneering wine journalist Robert Lawrence Balzer. He assembled 14 leading wine experts including France's Alexis Lichine, who owned two chateaux in Bordeaux, a manager of the Four Seasons restaurant in New York City, and Sam Aaron, a prominent New York wine merchant. They evaluated 23 Chardonnays from California, New York, and France in a blind tasting before an assemblage of 250 members of the New York Food and Wine Society. California Chardonnays received the top four scores. Fifth place went to the 1969 Beaune Clos des Mouches Joseph Drouhin. Other French wines in the competition were the 1970 Corton-Charlemagne Louis Latour, the 1971 Pouilly-Fuisse Louis Jadot, and the 1970 Chassagne-Montrachet Marquis de Laguiche Joseph Drouhin.
Three years later, in the notable Paris Wine Tasting of 1976, American and French wines were once again compared in a blind tasting.
See also
Wine competitions
Globalization of wine
|
https://en.wikipedia.org/wiki/Absolute%20OpenBSD
|
Absolute OpenBSD: Unix for the Practical Paranoid is a comprehensive guide to the OpenBSD operating system by Michael W. Lucas, author of Absolute FreeBSD and Cisco Routers for the Desperate. The book assumes basic knowledge of the design, commands, and user permissions of Unix-like operating systems. The book contains troubleshooting tips, background information on the system and its commands, and examples to assist with learning.
1st edition
The first edition was released in June 2003. Some of the information in the book became outdated when OpenBSD 3.4 was released only a few months later.
2nd edition
The second edition was released in April 2013. Peter N. M. Hansteen, author of The Book of PF, was the technical reviewer.
External links
|
https://en.wikipedia.org/wiki/Oxamniquine
|
Oxamniquine, sold under the brand name Vansil among others, is a medication used to treat schistosomiasis due to Schistosoma mansoni. Praziquantel, however, is often the preferred treatment. It is given by mouth and used as a single dose.
Common side effects include sleepiness, headache, nausea, diarrhea, and reddish urine. It is typically not recommended during pregnancy, if possible. Seizures may occur and therefore caution is recommended in people with epilepsy. It works by causing paralysis of the parasitic worms. It is in the anthelmintic family of medications.
Oxamniquine was first used medically in 1972. It is on the World Health Organization's List of Essential Medicines. It is not commercially available in the United States. It is more expensive than praziquantel.
Medical uses
Oxamniquine is used for treatment of schistosomiasis. According to one systematic review, praziquantel is the standard treatment for S. mansoni infections and oxamniquine also appears effective.
Side effects
It is generally well tolerated following oral doses. Dizziness with or without drowsiness occurs in at least a third of patients, beginning up to three hours after a dose, and usually lasts for up to six hours. Headache and gastrointestinal effects, such as nausea, vomiting, and diarrhoea, are also common.
Allergic-type reactions, including urticaria, pruritic skin rashes, and fever, may occur. Liver enzyme values have been raised transiently in some patients. Epileptiform convulsions have been reported, especially in patients with a history of convulsive disorders. Hallucinations and excitement have occurred rarely.
A reddish discoloration of urine, probably due to a metabolite of oxamniquine, has been reported.
Oxamniquine is not recommended during pregnancy.
Pharmacokinetics
Peak plasma concentrations are achieved one to three hours after a dose, and the plasma half-life is 1.0 to 2.5 hours.
It is extensively metabolised to inactive metabolites, principally the 6-car
|
https://en.wikipedia.org/wiki/Program%20comprehension
|
Program comprehension (also program understanding or [source] code comprehension) is a domain of computer science concerned with the ways software engineers maintain existing source code. The cognitive and other processes involved are identified and studied. The results are used to develop tools and training.
Software maintenance tasks have five categories: adaptive maintenance, corrective maintenance, perfective maintenance, code reuse, and code leverage.
Theories of program comprehension
Titles of works on program comprehension include
Using a behavioral theory of program comprehension in software engineering
The concept assignment problem in program understanding, and
Program Comprehension During Software Maintenance and Evolution.
Computer scientists pioneering program comprehension include Ruven Brooks, Ted J. Biggerstaff, and Anneliese von Mayrhauser.
See also
Program analysis (computer science)
Program slicing
Computer programming
|
https://en.wikipedia.org/wiki/Jacques-Louis%20Lions
|
Jacques-Louis Lions (; 2 May 1928 – 17 May 2001) was a French mathematician who made contributions to the theory of partial differential equations and to stochastic control, among other areas. He received the SIAM's John von Neumann Lecture prize in 1986 and numerous other distinctions. Lions is listed as an ISI highly cited researcher.
Biography
After being part of the French Résistance in 1943 and 1944, Jacques-Louis Lions entered the École Normale Supérieure in 1947.
He was a professor of mathematics at the University of Nancy, the Faculty of Sciences of Paris, and the École Polytechnique.
In 1966 he sent an invitation to Gury Marchuk, the soviet mathematician to visit Paris. This was hand delivered by Général De Gaulle during his visit to Akademgorodok in June of that year.
He joined the prestigious Collège de France as well as the French Academy of Sciences in 1973.
In 1979, he was appointed director of the Institut National de la Recherche en Informatique et Automatique (INRIA), where he taught and promoted the use of numerical simulations using finite elements integration. Throughout his career, Lions insisted on the use of mathematics in industry, with a particular involvement in the French space program, as well as in domains such as energy and the environment.
This eventually led him to be appointed director of the Centre National d'Etudes Spatiales (CNES) from 1984 to 1992.
Lions was elected President of the International Mathematical Union in 1991 and also received the Japan Prize and the Harvey Prize that same year. In 1992, the University of Houston awarded him an honorary doctoral degree. He was elected president of the French Academy of Sciences in 1996 and was also a Foreign Member of the Royal Society (ForMemRS) and numerous other foreign academies.
He has left a considerable body of work, among this more than 400 scientific articles, 20 volumes of mathematics that were translated into English and Russian, and major contributions to several
|
https://en.wikipedia.org/wiki/Apple%20Open%20Directory
|
Apple Open Directory is the LDAP directory service model implementation from Apple Inc. A directory service is software which stores and organizes information about a computer network's users and network resources and which allows network administrators to manage users' access to the resources.
In the context of macOS Server, Open Directory describes a shared LDAPv3 directory domain and a corresponding authentication model composed of Apple Password Server and Kerberos 5 tied together using a modular Directory Services system. Apple Open Directory is a fork of OpenLDAP.
The term Open Directory can also be used to describe the entire directory services framework used by macOS and macOS Server. In this context, it describes the role of a macOS or macOS Server system when it is connected to an existing directory domain, in which context it is sometimes referred to as Directory Services.
Apple, Inc. also publishes an API called the OpenDirectory framework, permitting macOS applications to interrogate and edit the Open Directory data.
With the release of Mac OS X Leopard (10.5), Apple chose to move away from using the NetInfo directory service (originally found in NeXTSTEP and OPENSTEP), which had been used by default for all local accounts and groups in every release of Mac OS X from 10.0 to 10.4. Mac OS X 10.5 now uses Directory Services and its plugins for all directory information. Local accounts are now registered in the Local Plugin, which uses XML property list (plist) files stored in /var/db/dslocal/nodes/Default/ as its backing storage.
Implementation in macOS Server
macOS Server can host an Open Directory domain when configured as an Open Directory Master. In addition to its local directory, this OpenLDAP-based LDAPv3 domain is designed to store centralized management data, user, group, and computer accounts, which other systems can access. The directory domain is paired with the Open Directory Password Server and, optionally, a Kerberos realm. Either p
|
https://en.wikipedia.org/wiki/Roger%20Jones%20%28mathematician%29
|
Roger L. Jones is an American mathematician specializing in harmonic analysis and ergodic theory.
Biography
He obtained a B.S. in mathematics in 1971 from University at Albany, SUNY, and a Ph.D. in mathematics in 1974 from Rutgers University, with thesis Inequalities for the Ergodic Maximal Function written under the direction of Richard Floyd Gundy. He has recently retired from a professorship in mathematics at DePaul University in Chicago. There he taught everything from remedial math to graduate-level courses. During his tenure at DePaul, Roger published numerous research papers in math, was awarded an excellence in teaching award, chaired the DePaul University Mathematics Department, and was awarded National Science Foundation grants related to teaching mathematics. He has also worked with the Chicago Public Schools on improving math instruction.
Roger was honored for his research work at the International Conference on Harmonic Analysis and Ergodic theory that was held in the name of Roger and his colleague Marshall Ash.
Roger has since retired from teaching at DePaul and moved to Northern Wisconsin, where he teaches mathematics at Conserve School.
Appointments
1974-1977: DePaul University, Assistant Professor
1977-1984: DePaul University, Associate Professor
1982-1985: DePaul University, Chairman: Department of Mathematics
1984-2004: DePaul University, Professor
2004–present: DePaul University, Professor Emeritus
Professional memberships
Mathematical Association of America
American Mathematical Society
Publications
|
https://en.wikipedia.org/wiki/Pupillary%20distance
|
Pupillary distance (PD), more correctly known as interpupillary distance (IPD) is the distance in millimeters between the centers of each pupil.
Interpupillary Distance Classifications
Distance PD is the separation between the visual axes of the eyes in their primary position, as the subject fixates on an infinitely distant object.
Near PD is the separation between the visual axes of the eyes, at the plane of the spectacle lenses, as the subject fixates on a near object at the intended working distance.
Intermediate PD is at a specified plane in between distance and near.
Monocular PD refers to the distance between either the right or left visual axis to the bridge of the nose, which may be slightly different for each eye due to anatomical variations but always sums up to the binocular PD.
For people who need to wear prescription glasses, consideration of monocular PD measurement by an optician helps to ensure that the lenses will be located in the optimum position.
Whilst PD is an optometric term used to specify prescription eyewear, IPD is more critical for the design of binocular viewing systems, where both eye pupils need to be positioned within the exit pupils of the viewing system. These viewing systems include binocular microscopes, night vision devices or goggles (NVGs), and head-mounted displays (HMDs). IPD data are used in the design of such systems to specify the range of lateral adjustment of the exit optics or eyepieces. IPD is also used to describe the distance between the exit pupils or optical axes of a binocular optical system. The distinction with IPD is the importance of anthropometric databases and the design of binocular viewing devices with an IPD adjustment that will fit a targeted population of users. Because instruments such as binoculars and microscopes can be used by different people, the distance between the eye pieces is usually made adjustable to account for IPD. In some applications, when IPD is not correctly set, it can lead to an
|
https://en.wikipedia.org/wiki/Dilution%20%28equation%29
|
Dilution is the process of decreasing the concentration of a solute in a solution, usually simply by mixing with more solvent like adding more water to the solution. To dilute a solution means to add more solvent without the addition of more solute. The resulting solution is thoroughly mixed so as to ensure that all parts of the solution are identical.
The same direct relationship applies to gases and vapors diluted in air for example. Although, thorough mixing of gases and vapors may not be as easily accomplished.
For example, if there are 10 grams of salt (the solute) dissolved in 1 litre of water (the solvent), this solution has a certain salt concentration (molarity). If one adds 1 litre of water to this solution, the salt concentration is reduced. The diluted solution still contains 10 grams of salt (0.171 moles of NaCl).
Mathematically this relationship can be shown by equation:
where
c1 = initial concentration or molarity
V1 = initial volume
c2 = final concentration or molarity
V2 = final volume
....
Basic room purge equation
The basic room purge equation is used in industrial hygiene. It determines the time required to reduce a known vapor concentration existing in a closed space to a lower vapor concentration. The equation can only be applied when the purged volume of vapor or gas is replaced with "clean" air or gas. For example, the equation can be used to calculate the time required at a certain ventilation rate to reduce a high carbon monoxide concentration in a room.
Sometimes the equation is also written as:
where
Dt = time required; the unit of time used is the same as is used for Q
V = air or gas volume of the closed space or room in cubic feet, cubic metres or litres
Q = ventilation rate into or out of the room in cubic feet per minute, cubic metres per hour or litres per second
Cinitial = initial concentration of a vapor inside the room measured in ppm
Cfinal = final reduced concentration of the vapor inside the room in ppm
Diluti
|
https://en.wikipedia.org/wiki/Throat%20singing
|
Throat singing refers to several vocal practices found in different cultures worldwide. The most distinctive feature of such vocal practices is to be associated to some type of guttural voice that contrasts with the most common types of voices employed in singing, which are usually represented by chest (modal) and head (light, or falsetto) registers. Throat singing is often described as producing the sensation of more than one pitch at a time, i.e., the listener perceives two or more distinct musical notes while the singer is producing a single vocalisation.
Throat singing, therefore, consists of a wide range of singing techniques that originally belonged to particular cultures and seem to share some sounding characteristics that make them especially noticeable by other cultures and users of mainstream singing styles. The term originates from the translation of the Tuvan/Mongolian word Xhöömei/Xhöömi, that literally means throat, guttural. Ethnic groups from Russia, Mongolia, Japan, South Africa, Canada, Italy, China and India, among others, accept and normally employ the term throat singing to describe their special way of producing voice, song and music.
The term throat singing is not precise, because any singing technique involves the sound generation in the "throat", i.e., the voice produced at the level of the larynx, which includes the vocal folds and other structures. Therefore it would be, in principle, admissible to refer to classical operatic singing or pop singing as "throat singing" for instance. However, the term throat is not adopted by the official terminology of anatomy (Terminologia Anatomica) and is not technically associated with most of the singing techniques. Many authors, performers, coaches, and listeners associate throat singing with overtone singing. Throat singing and overtone singing are certainly not synonyms, contrary to what is inaccurately indicated by many dictionaries (e.g., in the definition by Britannica) but, in some cases, both
|
https://en.wikipedia.org/wiki/Karmarkar%27s%20algorithm
|
Karmarkar's algorithm is an algorithm introduced by Narendra Karmarkar in 1984 for solving linear programming problems. It was the first reasonably efficient algorithm that solves these problems in polynomial time. The ellipsoid method is also polynomial time but proved to be inefficient in practice.
Denoting as the number of variables and as the number of bits of input to the algorithm, Karmarkar's algorithm requires operations on -digit numbers, as compared to such operations for the ellipsoid algorithm. The runtime of Karmarkar's algorithm is thus
using FFT-based multiplication (see Big O notation).
Karmarkar's algorithm falls within the class of interior-point methods: the current guess for the solution does not follow the boundary of the feasible set as in the simplex method, but moves through the interior of the feasible region, improving the approximation of the optimal solution by a definite fraction with every iteration and converging to an optimal solution with rational data.
The algorithm
Consider a linear programming problem in matrix form:
Karmarkar's algorithm determines the next feasible direction toward optimality and scales back by a factor . It is described in a number of sources. Karmarkar also has extended the method to solve problems with integer constraints and non-convex problems.
Since the actual algorithm is rather complicated, researchers looked for a more intuitive version of it, and in 1985 developed affine scaling, a version of Karmarkar's algorithm that uses affine transformations where Karmarkar used projective ones, only to realize four years later that they had rediscovered an algorithm published by Soviet mathematician I. I. Dikin in 1967. The affine-scaling method can be described succinctly as follows. While applicable to small scale problems, it is not a polynomial time algorithm.
stopping criterion, .
return unbounded
end if
end do
Example
Cons
|
https://en.wikipedia.org/wiki/Outcome%20%28game%20theory%29
|
In game theory, the outcome of a game is the ultimate result of a strategic interaction with one or more people, dependant on the choices made by all participants in a certain exchange. It represents the final payoff resulting from a set of actions that individuals can take within the context of the game. Outcomes are pivotal in determining the payoffs and expected utility for parties involved. Game theorists commonly study how the outcome of a game is determined and what factors affect it.
In game theory, a strategy is a set of actions that a player can take in response to the actions of others. Each player’s strategy is based on their expectation of what the other players are likely to do, often explained in terms of probability. Outcomes are dependent on the combination of strategies chosen by involved players and can be represented in a number of ways; one common way is a payoff matrix showing the individual payoffs for each players with a combination of strategies, as seen in the payoff matrix example below. Outcomes can be expressed in terms of monetary value or utility to a specific person. Additionally, a game tree can be used to deduce the actions leading to an outcome by displaying possible sequences of actions and the outcomes associated.
A commonly used theorem in relation to outcomes is the Nash equilibrium. This theorem is a combination of strategies in which no player can improve their payoff or outcome by changing their strategy, given the strategies of the other players. In other words, a Nash equilibrium is a set of strategies in which each player is doing the best possible, assuming what the others are doing to receive the most optimal outcome for themselves. It is important to note that not all games have a unique nash equilibrium and if they do, it may not be the most desirable outcome. Additionally, the desired outcomes is greatly affected by individuals chosen strategies, and their beliefs on what they believe other players will do under the
|
https://en.wikipedia.org/wiki/Agent%20%28economics%29
|
In economics, an agent is an actor (more specifically, a decision maker) in a model of some aspect of the economy. Typically, every agent makes decisions by solving a well- or ill-defined optimization or choice problem.
For example, buyers (consumers) and sellers (producers) are two common types of agents in partial equilibrium models of a single market. Macroeconomic models, especially dynamic stochastic general equilibrium models that are explicitly based on microfoundations, often distinguish households, firms, and governments or central banks as the main types of agents in the economy. Each of these agents may play multiple roles in the economy; households, for example, might act as consumers, as workers, and as voters in the model. Some macroeconomic models distinguish even more types of agents, such as workers and shoppers or commercial banks.
The term agent is also used in relation to principal–agent models; in this case, it refers specifically to someone delegated to act on behalf of a principal.
In agent-based computational economics, corresponding agents are "computational objects modeled as interacting according to rules" over space and time, not real people. The rules are formulated to model behavior and social interactions based on stipulated incentives and information. The concept of an agent may be broadly interpreted to be any persistent individual, social, biological, or physical entity interacting with other such entities in the context of a dynamic multi-agent economic system.
Representative vs. heterogenous agents
An economic model in which all agents of a given type (such as all consumers, or all firms) are assumed to be exactly identical is called a representative agent model. A model which recognizes differences among agents is called a heterogeneous agent model. Economists often use representative agent models when they want to describe the economy in the simplest terms possible. In contrast, they may be obliged to use heterogeneous agen
|
https://en.wikipedia.org/wiki/Schwinger%20effect
|
The Schwinger effect is a predicted physical phenomenon whereby matter is created by a strong electric field. It is also referred to as the Sauter–Schwinger effect, Schwinger mechanism, or Schwinger pair production. It is a prediction of quantum electrodynamics (QED) in which electron–positron pairs are spontaneously created in the presence of an electric field, thereby causing the decay of the electric field. The effect was originally proposed by Fritz Sauter in 1931 and further important work was carried out by Werner Heisenberg and Hans Heinrich Euler in 1936, though it was not until 1951 that Julian Schwinger gave a complete theoretical description.
The Schwinger effect can be thought of as vacuum decay in the presence of an electric field. Although the notion of vacuum decay suggests that something is created out of nothing, physical conservation laws are nevertheless obeyed. To understand this, note that electrons and positrons are each other's antiparticles, with identical properties except opposite electric charge.
To conserve energy, the electric field loses energy when an electron–positron pair is created, by an amount equal to , where is the electron rest mass and is the speed of light. Electric charge is conserved because an electron–positron pair is charge neutral. Linear and angular momentum are conserved because, in each pair, the electron and positron are created with opposite velocities and spins. In fact, the electron and positron are expected to be created at (close to) rest, and then subsequently accelerated away from each other by the electric field.
Mathematical description
Schwinger pair production in a constant electric field takes place at a constant rate per unit volume, commonly referred to as . The rate was first calculated by Schwinger and at leading (one-loop) order is equal to
where is the mass of an electron, is the elementary charge, and is the electric field strength. This formula cannot be expanded in a Taylor series in ,
|
https://en.wikipedia.org/wiki/Circumflex%20in%20French
|
The circumflex (ˆ) is one of the five diacritics used in French orthography. It may appear on the vowels a, e, i, o, and u, for example â in pâté.
The circumflex, called accent circonflexe, has three primary functions in French:
It affects the pronunciation of a, e, and o. Although it is used on i and u as well, it does not affect their pronunciation.
It often indicates the historical presence of a letter, commonly s, that has become silent and fallen away in orthography over the course of linguistic evolution.
It is used, less frequently, to distinguish between two homophones. For example, sur ('on/about') versus sûr '(sure/safe'), and du ('of the') versus dû ('due')
And in certain words, it is simply an orthographic convention.
First usages
The circumflex first appeared in written French in the 16th century. It was borrowed from Ancient Greek, and combines the acute accent and the grave accent. Grammarian Jacques Dubois (known as Sylvius) is the first writer known to have used the Greek symbol in his writing (although he wrote in Latin).
Several grammarians of the French Renaissance attempted to prescribe a precise usage for the diacritic in their treatises on language. The modern usage of the circumflex accent became standardized in the 18th or 19th century.
Jacques Dubois (Sylvius)
Sylvius used the circumflex to indicate so-called "false diphthongs". Early modern French as spoken in Sylvius' time had coalesced all its true diphthongs into phonetic monophthongs; that is, a pure vowel sound, one whose articulation at both beginning and end is relatively fixed, and which does not glide up or down towards a new position of articulation. He justifies its usage in his work Iacobii Sylvii Ambiani In Linguam Gallicam Isagoge una, cum eiusdem Grammatica Latinogallica ex Hebraeis Graecis et Latinus authoribus (An Introduction to the Gallic (French) Language, And Its Grammar With Regard to Hebrew, Latin and Greek Authors) published by Robert Estienne in 153
|
https://en.wikipedia.org/wiki/Ethanol%20precipitation
|
Ethanol precipitation is a method used to purify and/or concentrate RNA, DNA, and polysaccharides such as pectin and xyloglucan from aqueous solutions by adding ethanol as an antisolvent.
DNA precipitation
Theory
DNA is polar due to its highly charged phosphate backbone. Its polarity makes it water-soluble (water is polar) according to the principle "like dissolves like".
Because of the high polarity of water, illustrated by its high dielectric constant of 80.1 (at 20 °C), electrostatic forces between charged particles are considerably lower in aqueous solution than they are in a vacuum or in air.
This relation is reflected in Coulomb's law, which can be used to calculate the force acting on two charges and separated by a distance by using the dielectric constant (also called relative static permittivity) of the medium in the denominator of the equation ( is an electric constant):
At an atomic level, the reduction in the force acting on a charge results from water molecules forming a hydration shell around it. This fact makes water a very good solvent for charged compounds like salts. Electric force which normally holds salt crystals together by way of ionic bonds is weakened in the presence of water allowing ions to separate from the crystal and spread through solution.
The same mechanism operates in the case of negatively charged phosphate groups on a DNA backbone: even though positive ions are present in solution, the relatively weak net electrostatic force prevents them from forming stable ionic bonds with phosphates and precipitating out of solution.
Ethanol is much less polar than water, with a dielectric constant of 24.3 (at 25 °C). This means that adding ethanol to solution disrupts the screening of charges by water. If enough ethanol is added, the electrical attraction between phosphate groups and any positive ions present in solution becomes strong enough to form stable ionic bonds and DNA precipitation. This usually happens when ethanol compo
|
https://en.wikipedia.org/wiki/Quantum%20neural%20network
|
Quantum neural networks are computational neural network models which are based on the principles of quantum mechanics. The first ideas on quantum neural computation were published independently in 1995 by Subhash Kak and Ron Chrisley, engaging with the theory of quantum mind, which posits that quantum effects play a role in cognitive function. However, typical research in quantum neural networks involves combining classical artificial neural network models (which are widely used in machine learning for the important task of pattern recognition) with the advantages of quantum information in order to develop more efficient algorithms. One important motivation for these investigations is the difficulty to train classical neural networks, especially in big data applications. The hope is that features of quantum computing such as quantum parallelism or the effects of interference and entanglement can be used as resources. Since the technological implementation of a quantum computer is still in a premature stage, such quantum neural network models are mostly theoretical proposals that await their full implementation in physical experiments.
Most Quantum neural networks are developed as feed-forward networks. Similar to their classical counterparts, this structure intakes input from one layer of qubits, and passes that input onto another layer of qubits. This layer of qubits evaluates this information and passes on the output to the next layer. Eventually the path leads to the final layer of qubits. The layers do not have to be of the same width, meaning they don't have to have the same number of qubits as the layer before or after it. This structure is trained on which path to take similar to classical artificial neural networks. This is discussed in a lower section. Quantum neural networks refer to three different categories: Quantum computer with classical data, classical computer with quantum data, and quantum computer with quantum data.
Examples
Quantum neural ne
|
https://en.wikipedia.org/wiki/Nameplate
|
A nameplate identifies and displays a person or product's name. Nameplates are usually shaped as rectangles but are also seen in other shapes, sometimes taking on the shape of someone's written name. Nameplates primarily serve an informative function (as in an office environment, where nameplates mounted on doors or walls identify employees' spaces) or a commercial role (as in a retail environment, where nameplates are mounted on products to identify the brand). Whereas name tags tend to be worn on uniforms or clothing, nameplates tend to be mounted onto an object (e.g. cars, amplification devices) or physical space (e.g. doors, walls, or desktops). Nameplates are also distinct from name plaques. Plaques have larger dimensions and aim to communicate more information than a name and title.
Office nameplates
Office nameplates generally are made out of plastic, wood, metals (stainless steel, brass, aluminium, zinc, copper) and usually contain one or two lines of text. The standard format for an office nameplate is to display a person's name on the first line and a person's job title on the second line. It is common for organizations to request nameplates that exclude the job title. The primary reasons for excluding job titles are to extend the longevity of a nameplate and to promote a culture of meritocracy, where the strength of one's thoughts are not connected to one's job title. Nameplates without job titles have longer lives because someone can reuse the same nameplate after changing job titles. It is rare for an office nameplate to contain three or more lines of text. Although office nameplates range in size, the most popular nameplate size is . Office nameplates typically are made out of plastic. This is because plastic is an inexpensive material relative to wood and metal. More expensive nameplates can be manufactured out of bronze. To promote consistency, organizations tend to use the same style nameplate for all employees. This helps to achiev
|
https://en.wikipedia.org/wiki/Rose%20symbolism
|
Various folk cultures and traditions assign symbolic meaning to the rose, though these are seldom understood in-depth. Examples of deeper meanings lie within the language of flowers, and how a rose may have a different meaning in arrangements. Examples of common meanings of different coloured roses are: true love (red), mystery (blue), innocence or purity (white), death (black), friendship (yellow), and passion (orange).
In religion
Greco-Roman religion
In ancient Greece, the rose was closely associated with the goddess Aphrodite. In the Iliad, Aphrodite protects the body of Hector using the "immortal oil of the rose" and the archaic Greek lyric poet Ibycus praises a beautiful youth saying that Aphrodite nursed him "among rose blossoms". The second-century AD Greek travel writer Pausanias associates the rose with the story of Adonis Book Eleven of the ancient Roman novel The Golden Ass by Apuleius contains a scene in which the goddess Isis, who is identified with Venus, instructs the main character, Lucius, who has been transformed into a donkey, to eat rose petals from a crown of roses worn by a priest as part of a religious procession in order to regain his humanity.
Judaism
In the Song of Songs 2:1-2, the Jewish people are compared with a rose, remaining beautiful amongst thorns, although some translations instead refer to a "lily among thorns." The Zohar uses a "thirteen-petalled rose" as a symbol for the thirteen attributes of Divine Mercy named in Exodus 34:6-7. The rose and rosettes were also used to symbolize royalty and Israel, and were used in wreaths for the bridegroom at weddings in Biblical times.
Christianity
Following the Christianization of the Roman Empire, the rose became identified with the Virgin Mary. The rose symbol eventually led to the creation of the rosary and other devotional prayers in Christianity. Ever since the 1400s, the Franciscans have had a Crown Rosary of the Seven Joys of the Blessed Virgin Mary. In the 1400s and 1500s, t
|
https://en.wikipedia.org/wiki/Shadowrun%20%282007%20video%20game%29
|
Shadowrun is a first-person shooter video game, developed by FASA Studio for Xbox 360 and Windows Vista. The game features a buying system which is inspired by the game Counter-Strike. The game is also inspired by the role-playing game of the same name.
Gameplay
Shadowrun'''s multiplayer consists wholly of a first person/third person deathmatch. Players choose various races with unique abilities. Additionally, a currency system dictates in-match upgrades, with each race given a different amount of starting capital. The four playable races are Human, Elf, Dwarf, and Troll. Magic is a key component to this game. Players can heal, damage, teleport, and summon to gain advantages over others. Additionally, gadgets, or "tech", are obtainable through currency. Currency also allows players to purchase new weapons.Shadowrun features no campaign mode. If a user is without online services, they can set up bot matches and hone their skills.
Plot
According to the ancient Mayan calendar, magic is cyclical, leaving the world and returning every 5000 years. Magic enters the world, grows, peaks, and eventually retreats. When magic was last at its peak, a powerful Ziggurat was constructed near what would be modern day Santos, Brazil. The purpose of this construct is shrouded in the mists of history. Even the Chancela family, who secretly maintained the ziggurat for thousands of years, did not know its purpose, nor did they know the purpose of the strange artifact somehow connected to the ziggurat. In the millennia since its construction, the ziggurat was eventually buried, hidden in the side of a mountain. Then, on December 24, 2012, magic began returning to the world, leaving change and confusion in its wake.
The years after magic's return wrought change on a global scale. RNA Global, a powerful multinational corporation, sent a research team to Santos, Brazil. Their job was to explore and research the strange energies coming from a mountainside along one edge of Santos. Armed
|
https://en.wikipedia.org/wiki/Cauchy%27s%20convergence%20test
|
The Cauchy convergence test is a method used to test infinite series for convergence. It relies on bounding sums of terms in the series. This convergence criterion is named after Augustin-Louis Cauchy who published it in his textbook Cours d'Analyse 1821.
Statement
A series is convergent if and only if for every there is a natural number N such that
holds for all n > N and all p ≥ 1.
Explanation
The test works because the space of real numbers and the space of complex numbers (with the metric given by the absolute value) are both complete. From here, the series is convergent if and only if the partial sums
are a Cauchy sequence.
Cauchy's convergence test can only be used in complete metric spaces (such as and ), which are spaces where all Cauchy sequences converge. This is because we need only show that its elements become arbitrarily close to each other after a finite progression in the sequence to prove the series converges.
Proof
We can use the results about convergence of the sequence of partial sums of the infinite series and apply them to the convergence of the infinite series itself. The Cauchy Criterion test is one such application.
For any real sequence , the above results on convergence imply that the infinite series
converges if and only if for every there is a number N, such that m ≥ n ≥ N imply
Probably the most interesting part of this theorem is that the Cauchy condition implies the existence of the limit: this is indeed related to the completeness of the real line.
The Cauchy criterion can be generalized to a variety of situations, which can all be loosely summarized as "a vanishing oscillation condition is equivalent to convergence".
|
https://en.wikipedia.org/wiki/Features%20new%20to%20Windows%20Vista
|
Compared with previous versions of Microsoft Windows, features new to Windows Vista are very numerous, covering most aspects of the operating system, including additional management features, new aspects of security and safety, new I/O technologies, new networking features, and new technical features. Windows Vista also removed some others.
Windows Shell and user interface
Windows Aero
Windows Vista introduces a redesigned user interface and visual style named Windows Aero (a backronym for Authentic, Energetic, Reflective, and Open) that is intended to be aesthetically pleasing and cleaner than previous versions of Windows, with features such as glass translucencies, light effects, live thumbnails, and window animations enabled by the new Desktop Window Manager. Windows Aero also encompasses a new default typeface (Segoe UI)—set at a larger size than the default font of previous versions of Windows—new mouse cursors and new sounds, new dialog box, pop-up notification, and wizard interfaces, and revisions to the tone and phrasing of messages throughout the operating system. Windows Aero is available in the Home Premium, Business, Enterprise, and Ultimate Windows Vista editions.
All editions of Windows Vista include a new "Windows Vista Basic" theme with updated visuals; it is equivalent to Luna of Windows XP in that it does not rely on a compositing window manager. Glass translucencies, light effects, live thumbnails, or window animations of Windows Aero are not available. Windows Vista Home Basic additionally includes a unique "Windows Vista Standard" theme, which has the same hardware requirements of Windows Aero, but it does not include glass translucency or live thumbnail features or effects.
Start menu
The Start menu has undergone a significant revision in Windows Vista, and it is updated in accordance with Windows Aero design principles, featuring glass translucencies and subtle light effects while Windows Aero is enabled. The current user's profile p
|
https://en.wikipedia.org/wiki/Stanton%20number
|
The Stanton number, St, is a dimensionless number that measures the ratio of heat transferred into a fluid to the thermal capacity of fluid. The Stanton number is named after Thomas Stanton (engineer) (1865–1931). It is used to characterize heat transfer in forced convection flows.
Formula
where
h = convection heat transfer coefficient
ρ = density of the fluid
cp = specific heat of the fluid
u = velocity of the fluid
It can also be represented in terms of the fluid's Nusselt, Reynolds, and Prandtl numbers:
where
Nu is the Nusselt number;
Re is the Reynolds number;
Pr is the Prandtl number.
The Stanton number arises in the consideration of the geometric similarity of the momentum boundary layer and the thermal boundary layer, where it can be used to express a relationship between the shear force at the wall (due to viscous drag) and the total heat transfer at the wall (due to thermal diffusivity).
Mass transfer
Using the heat-mass transfer analogy, a mass transfer St equivalent can be found using the Sherwood number and Schmidt number in place of the Nusselt number and Prandtl number, respectively.
where
is the mass Stanton number;
is the Sherwood number based on length;
is the Reynolds number based on length;
is the Schmidt number;
is defined based on a concentration difference (kg s−1 m−2);
is the velocity of the fluid
Boundary layer flow
The Stanton number is a useful measure of the rate of change of the thermal energy deficit (or excess) in the boundary layer due to heat transfer from a planar surface. If the enthalpy thickness is defined as:
Then the Stanton number is equivalent to
for boundary layer flow over a flat plate with a constant surface temperature and properties.
Correlations using Reynolds-Colburn analogy
Using the Reynolds-Colburn analogy for turbulent flow with a thermal log and viscous sub layer model, the following correlation for turbulent heat transfer for is applicable
where
See also
Strouhal number, an unrelated nu
|
https://en.wikipedia.org/wiki/Turing%27s%20proof
|
Turing's proof is a proof by Alan Turing, first published in January 1937 with the title "On Computable Numbers, with an Application to the ". It was the second proof (after Church's theorem) of the negation of Hilbert's ; that is, the conjecture that some purely mathematical yes–no questions can never be answered by computation; more technically, that some decision problems are "undecidable" in the sense that there is no single algorithm that infallibly gives a correct "yes" or "no" answer to each instance of the problem. In Turing's own words:
"what I shall prove is quite different from the well-known results of Gödel ... I shall now show that there is no general method which tells whether a given formula U is provable in K [Principia Mathematica]".
Turing followed this proof with two others. The second and third both rely on the first. All rely on his development of typewriter-like "computing machines" that obey a simple set of rules and his subsequent development of a "universal computing machine".
Summary of the proofs
In his proof that the Entscheidungsproblem can have no solution, Turing proceeded from two proofs that were to lead to his final proof. His first theorem is most relevant to the halting problem, the second is more relevant to Rice's theorem.
First proof: that no "computing machine" exists that can decide whether or not an arbitrary "computing machine" (as represented by an integer 1, 2, 3, . . .) is "circle-free" (i.e. goes on printing its number in binary ad infinitum): "...we have no general process for doing this in a finite number of steps" (p. 132, ibid.). Turing's proof, although it seems to use the "diagonal process", in fact shows that his machine (called H) cannot calculate its own number, let alone the entire diagonal number (Cantor's diagonal argument): "The fallacy in the argument lies in the assumption that B [the diagonal number] is computable" The proof does not require much mathematics.
Second proof: This one is perhaps more f
|
https://en.wikipedia.org/wiki/Lost%20Planet%3A%20Extreme%20Condition
|
Lost Planet: Extreme Condition is a third-person shooter video game developed and published by Capcom for Xbox 360, Microsoft Windows and PlayStation 3. The game was released in Japan in December 2006 and worldwide in January 2007. Originally intended to be an Xbox 360 exclusive, it was later ported and released for Microsoft Windows in June 2007 and PlayStation 3 in February 2008.
Gameplay
The game is played through a third person over-the-shoulder view. Players are allowed to switch between first-person and third-person at any moment. Players either travel on foot or ride various types of mechanized suits called Vital Suits (VSs). VSs carry heavy weapons such as chain guns and rocket launchers. They can pick up weapons lying on the ground and fire multiple weapons at once. On foot, players are able to use a grappling hook to pull themselves up to normally hard-to-reach places, or to hook onto a VS and hijack it. Driving VSs and using certain weapons requires thermal energy. Also, the planet's cold temperature causes the characters' thermal energy level to continually decrease. Players can replenish their thermal energy level by defeating enemies or activating data posts. Data posts also allow players to use their navigational radars to see incoming enemies. Each of the 11 levels is accompanied by a boss, which can be either a VS or a large Akrid.
Multiplayer
Online multiplayer versus also requires players to monitor their thermal energy level, but here, reaching zero does not cause death. Instead, the characters cannot use VSs or fire the weapons which require thermal energy. Online multiplayer versus consists of four modes, called Elimination, Team Elimination, Post Grab, and Fugitive. Players score points by killing other players and activating posts, and they lose points for being killed or committing suicide. Post grab is a mode where players on opposite teams compete to capture as many posts as possible before the set time runs out. Team Elimination is a 1
|
https://en.wikipedia.org/wiki/Comparison%20of%20operating%20system%20kernels
|
A kernel is a component of a computer operating system. A comparison of system kernels can provide insight into the design and architectural choices made by the developers of particular operating systems.
Comparison criteria
The following tables compare general and technical information for a number of widely used and currently available operating system kernels. Please see the individual products' articles for further information.
Even though there are a large number and variety of available Linux distributions, all of these kernels are grouped under a single entry in these tables, due to the differences among them being of the patch level. See comparison of Linux distributions for a detailed comparison. Linux distributions that have highly modified kernels — for example, real-time computing kernels — should be listed separately. There are also a wide variety of minor BSD operating systems, many of which can be found at comparison of BSD operating systems.
The tables specifically do not include subjective viewpoints on the merits of each kernel or operating system.
Feature overview
The major contemporary general-purpose kernels are shown in comparison. Only an overview of the technical features is detailed.
Transport protocol support
In-kernel security
In-kernel virtualization
In-kernel server support
Binary format support
A comparison of OS support for different binary formats (executables):
File system support
Physical file systems:
Networked file system support
Supported CPU instruction sets and microarchitectures
Supported GPU processors
Supported kernel execution environment
This table indicates, for each kernel, what operating systems' executable images and device drivers can be run by that kernel.
Supported cipher algorithms
This may be usable on some situations like file system encrypting.
Supported compression algorithms
This may be usable on some situations like compression file system.
Supported message digest algorithms
Support
|
https://en.wikipedia.org/wiki/Waxy%20corn
|
Waxy corn or glutinous corn is a type of field corn characterized by its sticky texture when cooked as a result of larger amounts of amylopectin. The corn was first described from a specimen from China in 1909. As this plant showed many peculiar traits, the American breeders long used it as a genetic marker to tag the existence of hidden genes in other maize breeding programs. In 1922 a researcher found that the endosperm of waxy maize contained only amylopectin and no amylose starch molecule in opposition to normal dent corn varieties that contain both. Until World War II, the main source of starch in the United States was tapioca, but when Japan severed the supply lines of the U.S., they forced processors to turn to waxy maize. Amylopectin or waxy starch is now used mainly in food products, but also in the textile, adhesive, corrugating and paper industry.
When feeding trials later on showed that waxy maize could produce more efficient feed gains than normal dent maize, interest in waxy maize suddenly expanded. Geneticists could show that waxy maize has a defect in metabolism precluding the synthesis of amylose in the endosperm. It is coded by a single recessive gene (wx). Waxy maize yield about 3.5% less than their normal dent counterparts and has to be isolated from any nearby normal maize fields by at least 200 meters.
History
The exact history of waxy maize is unknown. The first mentions of it were found in the archives of the U.S. Department of Agriculture (USDA). In 1908, the Rev. J. M. W. Farnham, a Presbyterian missionary in Shanghai, sent a sample of seeds to the U.S. Office of Foreign Seed and Plant Introduction. A note with the seeds called it: "A peculiar kind of corn. There are several colours, but they are said to be all the same variety. The corn is much more glutinous than the other varieties, so far as I know, and may be found to be of some use, perhaps as porridge." These seeds were planted on May 9, 1908, near Washington, D.C., by a botan
|
https://en.wikipedia.org/wiki/Linux%20kernel%20oops
|
In computing, an oops is a serious but non-fatal error in the Linux kernel. An oops may precede a kernel panic, but it may also allow continued operation with compromised reliability. The term does not stand for anything, other than that it is a simple mistake.
Functioning
When the kernel detects a problem, it kills any offending processes and prints an oops message, which Linux kernel engineers can use in debugging the condition that created the oops and fixing the underlying programming error. After a system has experienced an oops, some internal resources may no longer be operational. Thus, even if the system appears to work correctly, undesirable side effects may have resulted from the active task being killed. A kernel oops often leads to a kernel panic when the system attempts to use resources that have been lost. Some kernels are configured to panic when many oopses ( by default) have occurred. This oops limit is due to the potential, for example, for attackers to repeatedly trigger an oops and an associated resource leak, which eventually overflows an integer and allows further exploitation.
The official Linux kernel documentation regarding oops messages resides in the file of the kernel sources. Some logger configurations may affect the ability to collect oops messages. The kerneloops software can collect and submit kernel oopses to a repository such as the www.kerneloops.org website, which provides statistics and public access to reported oopses.
For a person not familiar with technical details of computers and operating systems, an oops message might look confusing. Unlike other operating systems such as Windows or macOS, Linux chooses to present details explaining the crash of the kernel rather than display a simplified, user-friendly message, such as the BSoD on Windows. A simplified crash screen has been proposed a few times, however currently none are in development.
See also
kdump (Linux) Linux kernel's crash dump mechanism, which internally u
|
https://en.wikipedia.org/wiki/Anticor
|
An anticor, also known as anticoeur or avant-cœur, among farriers, is a dangerous swelling or inflammation in a horse's breast, of the size and shape of an apple, just opposite the heart. The term literally means anti heart or before heart.
The swelling may appear as a hard tumor, slow to develop, or as an inflammation. A traditional remedy, in the first case, involves splitting the skin along the breadth of the tumor, allowing the matter contained to escape, and stopping the hemorrhage by using an amadou or a hot iron. This kind of operation is best done by a veterinarian. If the tumor is inflammatory, one resorts to an oil of pompillion, an ointment made of buds of black poplar, lard and sheets of poppy, belladonna, etc. If it has formed an abscess, one first applies a soft poultice.
In pre-modern medicine, this was thought to be caused by a sanguine and bilious humour. The disease has also been erroneously attributed to the heart, whence it was called by Jacques de Solleysell a swelling of the pericardium, whereas it is really an inflammation in the gullet and throat.
In humans, this is called Ludwig's angina, or squinancy.
|
https://en.wikipedia.org/wiki/Frederic%20Charles%20Dreyer
|
Admiral Sir Frederic Charles Dreyer, (8 January 1878 – 11 December 1956) was an officer of the Royal Navy. A gunnery expert, he developed a fire control system for British warships, and served as flag captain to Admiral Sir John Jellicoe at the Battle of Jutland. He retired with the rank of admiral in 1943, having served through two world wars and having already retired once.
Background and early life
Frederic Dreyer was born on 8 January 1878 in the Irish town of Parsonstown (now Birr) in King's County (now County Offaly), the second son of the Danish-born astronomer John Louis Emil Dreyer who was director of the Armagh Observatory. Educated at The Royal School, Armagh, in 1891 Dreyer joined the Royal Navy and entered the Royal Naval College, Dartmouth.
Royal Navy career
Early years
At Dartmouth Dreyer performed well in his examinations and was placed fifth in his term. He then served as a midshipman in HMS Anson (1893–1896) and HMS Barfleur (1896–1897). In nearly all his subsequent examinations for promotions he obtained Class 1 certificates—for sub-lieutenant, lieutenant (July 1898, while aboard HMS Repulse) and then gunnery lieutenant. In 1900 he authored a book called How to Get a First Class in Seamanship. He came first in his class of three in the advanced course for gunnery and torpedo lieutenants at the Royal Naval College, Greenwich in 1901, after which he was posted to the staff of the gunnery school at Sheerness. He served as gunnery officer to the cruiser HMS Scylla for annual manoeuvres during summer 1902, then was lent to the protected cruiser HMS Hawke for a trooping trip to the Mediterranean (August–September 1902). He was appointed to the battleship HMS Hood in the Mediterranean from September 1902, but the ship's rudder had been damaged and the ship proceeded home to be repaired and paid off at Plymouth. Dreyer was reappointed to the Hawke on 13 January 1903 for another trooping voyage to Malta, and when she was paid off in March, he was appo
|
https://en.wikipedia.org/wiki/Close-ratio%20transmission
|
A close-ratio transmission describes a motor vehicle transmission with a smaller than average difference between the gear ratios. They are most often used on sports cars in order to keep the engine in the power band. There is no industry standard as to what constitutes a close-ratio transmission, a transmission that one manufacturer terms close-ratio may not necessarily be considered close-ratioed by another manufacturer. Generally speaking, the more gears a transmission has, the closer they are together. A continuously variable transmission has a near infinite "number" of gear ratios, which implies an infinitely close-ratio between gears. However, with no specific gear ratios, it would not be considered a close-ratioed transmission.
Comparison with ordinary transmission
This table compares the ratios of three Porsche 911 vehicles from 1967 to 1971, the first being the standard 901/75 transmission, the second being the 901/76 transmission denoted "For hill climbs", and the third being the 901/79 transmission denoted "Nurburgring ratios".
Mathematically, this closeness can be represented by the cumulative average spacing between, or geometric average of, gears. For the above series transmission, each successive gear's ratio is on average 75% of that of the preceding gear (e.g. (0.82 / 2.64)1/4 = 0.747). The Hill Climb transmission has successive gear ratios 81% of that of the preceding gear, and the Nuerburgring transmission has successive gear ratios 77% of that of the preceding gear. Thus, the Hill Climb transmission's gears are "closer" in numerical ratio to the preceding gear than that of the standard or Nuerburgring transmission, making it a close-ratio transmission.
There is no specific figure that is used to denote whether the steps between gears constitute a normal or close-ratio transmission. Often, manufacturers use this term when offering a standard manual transmission and an optional, sportier transmission, one with closer ratios than the other, such a
|
https://en.wikipedia.org/wiki/Heinrich%20Wilhelm%20Gottfried%20von%20Waldeyer-Hartz
|
Heinrich Wilhelm Gottfried von Waldeyer-Hartz (6 October 1836 – 23 January 1921) was a German anatomist, known for summarizing neuron theory and for naming the chromosome. He is also remembered by anatomical structures of the human body which were named after him: Waldeyer's tonsillar ring (the lymphoid tissue ring of the naso- and oropharynx) and Waldeyer's glands (of the eyelids).
Contribution to neuron theory
Waldeyer's name is associated in neuroscience with the "neuron theory", and for coining the term "neuron" to describe the basic structural unit of the nervous system. Waldeyer synthesized the discoveries by neuroanatomists (and later Nobel Prize winners) Camillo Golgi (1843–1926) and Santiago Ramón y Cajal (1852–1934), who had used the silver nitrate method of staining nerve tissue (Golgi's method), to formulate widely cited reviews of the theory. Waldeyer learned Spanish in order to absorb Cajal's detailed studies using Golgi's method and became his friend, mentor and promoter in the German-dominated field of microscopic anatomy. The theory was published in a series of papers in the main medical journal of Germany, Deutsche Medizinische Wochenschrift, which became extremely influential. However, as Cajal points out, though Waldeyer "supported the theory with the prestige of his authority, he did not contribute a single personal observation. He limited himself to a short brilliant exposition of the objective proofs, adduced by His, Kölliker, Retzius, van Gehuchten and myself, and he invented the fortunate term neuron."
Cytology and embryology
Waldeyer also studied the basophilic stained filaments which had been found to be the main constituents of chromatin, the material inside the cell nucleus, by his colleague of Kiel, Walther Flemming (1843–1905). Although its significance for genetics and for cell biology was still to be discovered, these filaments were known to be involved in the phenomenon of cell division discovered by Flemming, named mitosis. as
|
https://en.wikipedia.org/wiki/Enterprise%20unified%20process
|
The enterprise unified process (EUP) is an extended variant of the unified process and was developed by Scott W. Ambler and Larry Constantine in 2000, eventually reworked in 2005 by Ambler, John Nalbone and Michael Vizdos. EUP was originally introduced to overcome some shortages of RUP, namely the lack of production and eventual retirement of a software system. So two phases and several new disciplines were added. EUP sees software development not as a standalone activity, but embedded in the lifecycle of the system (to be built or enhanced or replaced), the IT lifecycle of the enterprise and the organization/business lifecycle of the enterprise itself. It deals with software development as seen from the customer's point of view.
In 2013 work began to evolve EUP to be based on disciplined agile delivery instead of the unified process.
Phases
The unified process defines four project phases
Inception
Elaboration
Construction
Transition
To these EUP adds two additional phases
Production
Retirement
Disciplines
The rational unified process defines nine project disciplines
Business modeling
Requirements
Analysis and design
Implementation
Test
Deployment
Configuration and change management
Project management
Environment
To these EUP adds one additional project discipline
Operations and support
and seven enterprise disciplines
Enterprise business modeling
Portfolio management
Enterprise architecture
Strategic reuse
People management
Enterprise administration
Software process improvement
Best practices of EUP
The EUP provides following best practices:-
Develop iteratively
Manage requirements
Proven architecture
Modeling
Continuously verify quality.
Manage change
Collaborative development
Look beyond development.
Deliver working software on a regular basis
Manage risk
See also
Disciplined agile delivery
Rational unified process
Software development process
Extreme programming
|
https://en.wikipedia.org/wiki/Hylotheism
|
Hylotheism (from Gk. hyle, 'matter' and theos, 'God') is the belief that matter and God are the same, so in other words, defining God as matter.
The American Lutheran Church–Missouri Synod defines hylotheism is "Theory equating matter with God or merging one into the other" which it states as "Synonym for pantheism* and materialism.*".
See also
Pantheism
|
https://en.wikipedia.org/wiki/Arnold%20tongue
|
In mathematics, particularly in dynamical systems, Arnold tongues (named after Vladimir Arnold) are a pictorial phenomenon that occur when visualizing how the rotation number of a dynamical system, or other related invariant property thereof, changes according to two or more of its parameters. The regions of constant rotation number have been observed, for some dynamical systems, to form geometric shapes that resemble tongues, in which case they are called Arnold tongues.
Arnold tongues are observed in a large variety of natural phenomena that involve oscillating quantities, such as concentration of enzymes and substrates in biological processes and cardiac electric waves. Sometimes the frequency of oscillation depends on, or is constrained (i.e., phase-locked or mode-locked, in some contexts) based on some quantity, and it is often of interest to study this relation. For instance, the outset of a tumor triggers in the area a series of substance (mainly proteins) oscillations that interact with each other; simulations show that these interactions cause Arnold tongues to appear, that is, the frequency of some oscillations constrain the others, and this can be used to control tumor growth.
Other examples where Arnold tongues can be found include the inharmonicity of musical instruments, orbital resonance and tidal locking of orbiting moons, mode-locking in fiber optics and phase-locked loops and other electronic oscillators, as well as in cardiac rhythms, heart arrhythmias and cell cycle.
One of the simplest physical models that exhibits mode-locking consists of two rotating disks connected by a weak spring. One disk is allowed to spin freely, and the other is driven by a motor. Mode locking occurs when the freely-spinning disk turns at a frequency that is a rational multiple of that of the driven rotator.
The simplest mathematical model that exhibits mode-locking is the circle map, which attempts to capture the motion of the spinning disks at discrete time interv
|
https://en.wikipedia.org/wiki/Type%20%28model%20theory%29
|
In model theory and related areas of mathematics, a type is an object that describes how a (real or possible) element or finite collection of elements in a mathematical structure might behave. More precisely, it is a set of first-order formulas in a language L with free variables x1, x2,…, xn that are true of a set of n-tuples of an L-structure . Depending on the context, types can be complete or partial and they may use a fixed set of constants, A, from the structure . The question of which types represent actual elements of leads to the ideas of saturated models and omitting types.
Formal definition
Consider a structure for a language L. Let M be the universe of the structure. For every A ⊆ M, let L(A) be the language obtained from L by adding a constant ca for every a ∈ A. In other words,
A 1-type (of ) over A is a set p(x) of formulas in L(A) with at most one free variable x (therefore 1-type) such that for every finite subset p0(x) ⊆ p(x) there is some b ∈ M, depending on p0(x), with (i.e. all formulas in p0(x) are true in when x is replaced by b).
Similarly an n-type (of ) over A is defined to be a set p(x1,…,xn) = p(x) of formulas in L(A), each having its free variables occurring only among the given n free variables x1,…,xn, such that for every finite subset p0(x) ⊆ p(x) there are some elements b1,…,bn ∈ M with .
A complete type of over A is one that is maximal with respect to inclusion. Equivalently, for every either or . Any non-complete type is called a partial type.
So, the word type in general refers to any n-type, partial or complete, over any chosen set of parameters (possibly the empty set).
An n-type p(x) is said to be realized in if there is an element b ∈ Mn such that . The existence of such a realization is guaranteed for any type by the compactness theorem, although the realization might take place in some elementary extension of , rather than in itself.
If a complete type is realized by b in , then the type is typically d
|
https://en.wikipedia.org/wiki/SCSI%20connector
|
A SCSI connector ( ) is used to connect computer parts that use a system called SCSI to communicate with each other. Generally, two connectors, designated male and female, plug together to form a connection which allows two components, such as a computer and a disk drive, to communicate with each other. SCSI connectors can be electrical connectors or optical connectors. There have been a large variety of SCSI connectors in use at one time or another in the computer industry. Twenty-five years of evolution and three major revisions of the standards resulted in requirements for Parallel SCSI connectors that could handle an 8, 16 or 32 bit wide bus running at 5, 10 or 20 megatransfer/s, with conventional or differential signaling. Serial SCSI added another three transport types, each with one or more connector types. Manufacturers have frequently chosen connectors based on factors of size, cost, or convenience at the expense of compatibility.
SCSI makes use of cables to connect devices. In a typical example, a socket on a computer motherboard would have one end of a cable plugged into it, while the other end of the cable plugged into a disk drive or other device. Some cables have different types of connectors on them, and some cables can have as many as 16 connectors (allowing 16 devices to be wired together). Different types of connectors may be used for devices inside a computer cabinet, than for external devices such as scanners or external disk drives.
Nomenclature
Many connector designations consist of an abbreviation for the connector family, followed by a number indicating the number of pins. For example, "CN36" (also written "CN-36" or "CN 36") would be a 36-pin Centronics-style connector. For some connectors (such as the D-subminiature family) use of the hyphen or space is more common, for others (like the "DD50") less so.
Parallel SCSI
Parallel SCSI (SCSI Parallel Interface SPI) allows for attachment of up to 8 devices (8-bit Narrow SCSI) or 16 devices
|
https://en.wikipedia.org/wiki/Time%E2%80%93space%20compression
|
Time–space compression (also known as space–time compression and time–space distanciation) is an idea referring to the altering of the qualities of space–time and the relationship between space and time that is a consequence of the expansion of capital. It is rooted in Karl Marx's theory of the "annihilation of space by time" originally elaborated in the Grundrisse, and was later articulated by Marxist geographer David Harvey in his book The Condition of Postmodernity. A similar idea was proposed by Elmar Altvater in an article in PROKLA in 1987, translated into English as "Ecological and Economic Modalities of Time and Space" and published in Capitalism Nature Socialism in 1990.
Time–space compression occurs as a result of technological innovations driven by the global expansion of capital that condense or elide spatial and temporal distances, including technologies of communication (telegraph, telephones, fax machines, Internet) and travel (rail, cars, trains, jets), driven by the need to overcome spatial barriers, open up new markets, speed up production cycles, and reduce the turnover time of capital.
According to Paul Virilio, time-space compression is an essential facet of capitalist life, saying that "we are entering a space which is speed-space ... This new other time is that of electronic transmission, of high-tech machines, and therefore, man is present in this sort of time, not via his physical presence, but via programming" (qtd. in Decron 71). In Speed and Politics, Virilio coined the term dromology to describe the study of "speed-space". Virilio describes velocity as the hidden factor in wealth and power, where historical eras and political events are effectively speed-ratios. In his view, acceleration destroys space and compresses time in ways of perceiving reality.
Theorists generally identify two historical periods in which time–space compression occurred; the period from the mid-19th century to the beginnings of the First World War, and the end
|
https://en.wikipedia.org/wiki/Heat%20map
|
A heat map (or heatmap) is a 2-dimensional data visualization technique that represents the magnitude of individual values within a dataset as a color. The variation in color may be by hue or intensity.
"Heat map" is a relatively new term, but the practice of shading matrices has existed for over a century.
History
Heat maps originated in 2D displays of the values in a data matrix. Larger values were represented by small dark gray or black squares (pixels) and smaller values by lighter squares. Toussaint Loua (1873) used a shading matrix to visualize social statistics across the districts of Paris. Sneath (1957) displayed the results of a cluster analysis by permuting the rows and the columns of a matrix to place similar values near each other according to the clustering. Jacques Bertin used a similar representation to display data that conformed to a Guttman scale. The idea for joining cluster trees to the rows and columns of the data matrix originated with Robert Ling in 1973. Ling used overstruck printer characters to represent different shades of gray, one character-width per pixel. Leland Wilkinson developed the first computer program in 1994 (SYSTAT) to produce cluster heat maps with high-resolution color graphics. The Eisen et al. display shown in the figure is a replication of the earlier SYSTAT design.
Software designer Cormac Kinney trademarked the term 'heat map' in 1991 to describe a 2D display depicting financial market information. The company that acquired Kinney's invention in 2003 unintentionally allowed the trademark to lapse.
Types
There are two main type of heat maps: spatial, and grid.
A spatial heat map displays the magnitude of a spatial phenomena as color, usually cast over a map. In the image labeled “Spatial Heat Map Example,” temperature is displayed by color range across a map of the world. Color ranges from blue (cold) to red (hot).
A grid heat map displays magnitude as color in a two-dimensional matrix, with each dimension rep
|
https://en.wikipedia.org/wiki/Galaksija%20BASIC
|
Galaksija BASIC was the BASIC interpreter of the Galaksija build-it-yourself home computer from Yugoslavia. While being partially based on code taken from TRS-80 Level 1 BASIC, which the creator believed to have been a Microsoft BASIC, the extensive modifications of Galaksija BASIC—such as to include rudimentary array support, video generation code (as the CPU itself did it in absence of dedicated video circuitry) and generally improvements to the programming language—is said to have left not much more than flow-control and floating point code remaining from the original.
The core implementation of the interpreter was fully contained in the 4 KiB ROM "A" or "1". The computer's original mainboard had a reserved slot for an extension ROM "B" or "2" that added more commands and features such as a built-in Zilog Z80 assembler.
ROM "A"/"1" symbols and keywords
The core implementation, in ROM "A" or "1", contained 3 special symbols and 32 keywords:
begins a comment (equivalent of standard BASIC REM command)
Equivalent of standard BASIC DATA statement
prefix for hex numbers
Allocates an array of strings, like DIM, but can allocate only array with name A$
serves as PEEK when used as a function (e.g. PRINT BYTE(11123)) and POKE when used as a command (e.g. BYTE 11123,123).
Calls BASIC subroutine as GOSUB in most other BASICs (e.g. CALL 100+4*X)
converts an ASCII numeric code into a corresponding character (string)
draws (command) or inspects (function) a pixel at given coordinates (0<=x<=63, 0<=y<=47).
displays the clock or time controlled by content of Y$ variable. Not in standard ROM
causes specified program line to be edited
standard part of IF-ELSE construct (Galaksija did not use THEN)
compare alphanumeric values X$ and Y$
standard FOR loop
standard GOTO command
equivalent of standard BASIC CLS command - clears the screen
protects n characters from the top of the screen from being scrolled away
standard part of IF-EL
|
https://en.wikipedia.org/wiki/Dysan
|
Dysan Corporation was an American storage media manufacturing corporation, formed in 1973 in San Jose, California, by CEO and former president C. Norman Dion. It was instrumental in the development of the 5.25" floppy disk, which appeared in 1976.
History
In 1983, Jerry Pournelle reported in BYTE that a software-publisher friend of his "distributes all his software on Dysan disks. It costs more to begin with, but saves [the cost of replacing defective media] in the long run, or so he says". By that year Dysan was a Fortune 500 company, had over 1200 employees, and was ranked as among the top ten private sector employers within Silicon Valley by the San Jose Mercury News, in terms of number of employees. In addition, some of Dysan's administrative and disk production facilities, located within the company's Santa Clara, manufacturing campus, were regarded as architecturally remarkable. For example, some of Dysan's Santa Clara campus magnetic media manufacturing facilities included architectural features such as large indoor employee lounge atriums, incorporating glass encased ceilings and walls, live indoor lush landscaping, waterfalls, running water creeks, and ponds with live fish.
In addition to manufacturing floppies, tape drives and hard disk drives, Dysan also produced hardware and storage containers for the disks.
Dysan merged with Xidex Magnetics in the spring of 1984. In 1997, under the direction of Jerry Ticerelli, Xidex declared bankruptcy. Xidex was absorbed by Anacomp and later spun off as a wholly owned subsidiary as Dysan.
After a brief re-opening in 2003, the company closed six months later under the direction of Dylan Campbell.
Recycling service
It is possible that Dysan was one of the first tech-based companies to offer a service for recycling used products. Some Dysan packaging included the following label:
|
https://en.wikipedia.org/wiki/Cyril%20Hilsum
|
Cyril Hilsum (born 17 May 1925) is a British physicist and academic.
Hilsum was elected a member of the National Academy of Engineering in 1983 for the inventiveness and leadership in introducing III-V semiconductors into electronic technology.
Life
Hilsum entered Raine's Foundation School in 1936 as the middle of three brothers, leaving in 1943 after being accepted into University College London, where he did his BSc. In 1945, he joined the Royal Naval Scientific Service, moving in 1947 to the Admiralty Research Laboratory. In 1950, he transferred again to the Services Electronics Research Laboratory (SERL) where he remained until 1964 before again moving, this time to the Royal Radar Establishment. He won the Welker Award in 1978, was elected a Fellow of the Royal Academy of Engineering, a Fellow of the Royal Society in 1979 and an honorary member of the American National Academy of Engineering. In 1983, he was appointed Chief Scientist at GEC Hirst Research Centre. He was awarded the Max Born Prize in 1987, the 1988 Faraday Medal, and from then until 1990 served as President of the Institute of Physics. In the 1990 Queen's Birthday Honours, he was appointed a Commander of the Order of the British Empire (CBE) for "services to the Electrical and Electronics Industry". He was the subject of a photograph by Nick Sinclair in 1993 that is currently held by the National Portrait Gallery. In 1997, he was awarded the Glazebrook Medal and Prize from the Institute of Physics, and is notable as the only scientist to hold both this and the Faraday Medal together. He has served as a corporate research advisor for various entities, including Cambridge Display Technology, the European Commission and Unilever. In 2007, he was awarded the Royal Society's Royal Medal 'for his many outstanding contributions and for continuing to use his prodigious talents on behalf of industry, government and academe to this day'.
Hilsum serves as chairman of the scientific board for Peratech a
|
https://en.wikipedia.org/wiki/Ohmic%20contact
|
An ohmic contact is a non-rectifying electrical junction: a junction between two conductors that has a linear current–voltage (I–V) curve as with Ohm's law. Low-resistance ohmic contacts are used to allow charge to flow easily in both directions between the two conductors, without blocking due to rectification or excess power dissipation due to voltage thresholds.
By contrast, a junction or contact that does not demonstrate a linear I–V curve is called non-ohmic. Non-ohmic contacts come in a number of forms, such as p–n junction, Schottky barrier, rectifying heterojunction, or breakdown junction.
Generally the term "ohmic contact" implicitly refers to an ohmic contact of a metal to a semiconductor, where achieving ohmic contact resistance is possible but requires careful technique. Metal–metal ohmic contacts are relatively simpler to make, by ensuring direct contact between the metals without intervening layers of insulating contamination, excessive roughness or oxidation; various techniques are used to create ohmic metal–metal junctions (soldering, welding, crimping, deposition, electroplating, etc.). This article focuses on metal–semiconductor ohmic contacts.
Stable contacts at semiconductor interfaces, with low contact resistance and linear I–V behavior, are critical for the performance and reliability of semiconductor devices, and their preparation and characterization are major efforts in circuit fabrication. Poorly prepared junctions to semiconductors can easily show rectifying behaviour by causing depletion of the semiconductor near the junction, rendering the device useless by blocking the flow of charge between those devices and the external circuitry. Ohmic contacts to semiconductors are typically constructed by depositing thin metal films of a carefully chosen composition, possibly followed by annealing to alter the semiconductor–metal bond.
Physics of formation of metal–semiconductor ohmic contacts
Both ohmic contacts and Schottky barriers are depen
|
https://en.wikipedia.org/wiki/Gunrock
|
Gunrock is the official mascot of the UC Davis Aggies, the athletic teams that represent the University of California, Davis, and was based on Gunrock (1914–1932), an American Thoroughbred stallion, and the son of English Triple Crown winner Rock Sand. He was related to the American Thoroughbred racehorse Man O' War (1917–1947), who was out of Mahubah (1910–1931), bay Thoroughbred mare by Rock Sand out of Merry Token. Gunrock was bred by Clarence Mackay, and likely born in Kentucky in 1914. He had an unsuccessful racing career as a 2-year-old, and was retired to stud by age 6 in 1920.
In 1921, Gunrock was donated by his owner, American financier and horse racing investor August Belmont Jr., to the U.S. Army Remount Service, and was brought to the campus of University of California, Davis, which was breeding horses for the Cavalry at the time. Gunrock covered a total of 476 mares during his career at stud, some owned by the university, and the rest from varied farms across northern California. Gunrock's progeny also included Thoroughbred racehorses, including the successful racing mare Sugar Pie (1928) out of Tooters (1918), among others. He also has descendants registered as Holsteiners and Irish Sport Horses.
In 1924, Gunrock was adopted as the official mascot of the men's basketball team, and the stallion accompanied the team to games and rallies. Later, a traditional mascot was created, and named "Gunrock" by the students. That mascot persisted into the 1970s, when he was replaced by "Ollie the Mustang". However, Ollie did not last long, as a period of confusion about the school's mascot and nickname set in, lasting into the first decade of the 21st century.
In 2003, the school's official mascot was officially identified as a Mustang, and the name "Gunrock" returned after 93% of the UC Davis student body voted to return to the original name.
In 2022, the students of UC Davis voted to replace Gunrock the Mustang with a dairy cow after the "#Cow4Mascot" and "
|
https://en.wikipedia.org/wiki/Isopycnal
|
Isopycnals are layers within the ocean that are stratified based on their densities and can be shown as a line connecting points of a specific density or potential density on a graph. Isopycnals are often displayed graphically to help visualize "layers" of the water in the ocean or gases in the atmosphere in a similar manner to how contour lines are used in topographic maps to help visualize topography.
Types
Oceanography
Water masses in the ocean are characterized by their properties. Factors such as density, temperature, and salinity can all be used to identify these masses and their origins as well as where they are in the water column. Density plays a large role in stratifying the ocean into layers. In a body of water, as the depth increases, so does the density; water masses with the highest density are at the bottom and the lowest densities are at the top. Typically, warm freshwater is less dense than cold salty water, thus the colder water will sink below the warmer water. Isopycnals are used to display this vertical distribution of the water. Variations in temperature and salinity along isopycnals can be described with spiciness. This creates distinguishable layers of water with differing physical properties. This phenomenon is called stratification. The strata are held in place by the large differences in physical and chemical properties between layers that prevent mixing.Turbulence can disturb boundaries between the layers, causing them to bend, which causes the isopycnals to appear uneven. These boundaries are known as diapycnals (Talley, 162). The ways in which the isopycnals and diapycnal are transformed can be used by oceanographers to identify the force that caused the underwater disturbance.
Mixing
Isopycnal mixing and diapycnal mixing work together to mix and ventilate the entire ocean. Isopycnal mixing is when surface waters moving into the interior of the ocean typically run horizontally, along the isopycnal layers, settling into their corr
|
https://en.wikipedia.org/wiki/New%20methylene%20blue
|
is an organic compound of the thiazine class of heterocycles. It is used as a stain and as an antimicrobial agent. It is classified as an azine dye, and the chromophore is a cation, the anion is often unspecified.
Applications
NMB is a staining agent used in diagnostic cytopathology and histopathology, typically for staining immature red blood cells. It is a supravital stain. It is closely related to methylene blue, an older stain in wide use.
Safety
New methylene blue is toxic. Skin contact or inhalation should be avoided.
See also
Methylene blue
|
https://en.wikipedia.org/wiki/Continuously%20compounded%20nominal%20and%20real%20returns
|
Return rate is a corporate finance and accounting tool which calculates the gain and loss of investment over a certain period of time.
Nominal return
Let Pt be the price of a security at time t, including any cash dividends or interest, and let Pt − 1 be its price at t − 1. Let RSt be the simple rate of return on the security from t − 1 to t. Then
The continuously compounded rate of return or instantaneous rate of return RCt obtained during that period is
If this instantaneous return is received continuously for one period, then the initial value Pt-1 will grow to during that period. See also continuous compounding.
Since this analysis did not adjust for the effects of inflation on the purchasing power of Pt, RS and RC are referred to as nominal rates of return.
Real return
Let be the purchasing power of a dollar at time t (the number of bundles of consumption that can be purchased for $1). Then , where PLt is the price level at t (the dollar price of a bundle of consumption goods). The simple inflation rate ISt from t –1 to t is . Thus, continuing the above nominal example, the final value of the investment expressed in real terms is
Then the continuously compounded real rate of return is
The continuously compounded real rate of return is just the continuously compounded nominal rate of return minus the continuously compounded inflation rate.
Sources
Applied mathematics
|
https://en.wikipedia.org/wiki/GLI1
|
Zinc finger protein GLI1 also known as glioma-associated oncogene is a protein that in humans is encoded by the GLI1 gene. It was originally isolated from human glioblastoma cells.
Function
The Gli proteins are the effectors of Hedgehog (Hh) signaling and have been shown to be involved in cell fate determination, proliferation and patterning in many cell types and most organs during embryo development. In the developing spinal cord the target genes of Gli proteins, that are themselves transcription factors, are arranged into a complex gene regulatory network that translates the extracellular concentration gradient of Sonic hedgehog into different cell fates along the dorsoventral axis.
The Gli transcription factors activate/inhibit transcription by binding to Gli responsive genes and by interacting with the transcription complex. The Gli transcription factors have DNA binding zinc finger domains which bind to consensus sequences on their target genes to initiate or suppress transcription. Yoon showed that mutating the Gli zinc finger domain inhibited the proteins effect proving its role as a transcription factor. Gli proteins have an 18-amino acid region highly similar to the α-helical herpes simplex viral protein 16 activation domain. This domain contains a consensus recognition element for the human TFIID TATA box-binding protein associated factor TAFII31. Other proteins such as Missing in Metastasis (MIM/BEG4) have been shown to potentiate the effects of the Gli transcription factors on target gene transcription. Gli and MIM have been shown to act synergistically to induce epidermal growth and MIM + Gli1 overexpressing grafts show similar growth patterns to Shh grafts.
Gli family
There are three members of the family; Gli1, Gli2 and Gli3 which are all transcription factors mediating the Hh pathway. The GLI1, GLI2, and GLI3 genes encode transcription factors which all contain conserved tandem C2-H2 zinc finger domains and a consensus histidine/cysteine
|
https://en.wikipedia.org/wiki/Windows%20Native%20API
|
The Native API is a lightweight application programming interface (API) used by Windows NT and user mode applications. This API is used in the early stages of Windows NT startup process, when other components and APIs are still unavailable. Therefore, a few Windows components, such as the Client/Server Runtime Subsystem (CSRSS), are implemented using the Native API. The Native API is also used by subroutines such as those in kernel32.dll that implement the Windows API, the API based on which most of the Windows components are created.
Most of the Native API calls are implemented in ntoskrnl.exe and are exposed to user mode by ntdll.dll. The entry point of ntdll.dll is LdrInitializeThunk. Native API calls are handled by the kernel via the System Service Descriptor Table (SSDT).
Function groups
The Native API comprises many functions. They include C runtime functions that are needed for a very basic C runtime execution, such as strlen(), sprintf(), memcpy() and floor(). Other common procedures like malloc(), printf(), scanf() are missing (the first because it does not specify a heap to allocate memory from and the second and third because they use the console, accessed only via KERNEL32.DLL). The vast majority of other Native API routines, by convention, have a 2 or 3 letter prefix, which is:
Nt or Zw are system calls declared in ntdll.dll and ntoskrnl.exe. When called from ntdll.dll in user mode, these groups are almost exactly the same; they execute an interrupt into kernel mode and call the equivalent function in ntoskrnl.exe via the SSDT. When calling the functions directly in ntoskrnl.exe (only possible in kernel mode), the Zw variants ensure kernel mode, whereas the Nt variants do not. The Zw prefix does not stand for anything.
Rtl is the second largest group of ntdll calls. These comprise the (extended) C Run-Time Library, which includes many utility functions that can be used by native applications, yet don't directly involve kernel support.
Csr are cli
|
https://en.wikipedia.org/wiki/Intruder%20detection
|
In information security, intruder detection is the process of detecting intruders behind attacks as unique persons. This technique tries to identify the person behind an attack by analyzing their computational behaviour. This concept is sometimes confused with Intrusion Detection (also known as IDS) techniques which are the art of detecting intruder actions.
History
Some other earlier works reference the concept of Intruder Authentication, Intruder Verification, or Intruder Classification, but the Si6 project was one of the first projects to deal with the full scope of the concept.
Theory
Intruder Detection Systems try to detect who is attacking a system by analyzing his or her computational behaviour or biometric behaviour.
Some of the parameters used to identify a intruder
Keystroke Dynamics (aka keystroke patterns, typing pattern, typing behaviour)
Patterns using an interactive command interpreter:
Commands used
Commands sequence
Accessed directories
Character deletion
Patterns on the network usage:
IP address used
ISP
Country
City
Ports used
TTL analysis
Operating system used to attack
Protocols used
Connection times patterns
Keystroke dynamics
Keystroke dynamics is paramount in Intruder Detection techniques because it is the only parameter that has been classified as a real 'behavioural biometric pattern'.
Keystroke dynamics analyze times between keystrokes issued in a computer keyboard or cellular phone keypad searching for patterns. First techniques used statistics and probability concepts like 'standard deviations' and 'Mean', later approaches use data mining, neural networks, Support Vector Machine, etc.
Translation confusion
There is a confusion with the Spanish translation of 'Intrusion detection system', also known as IDS. Some people translate it as 'Sistemas de Detección de Intrusiones', but others translate it as 'Sistemas de Detección de Intrusos'. Only the former is correct.
See also
Intrusion Detection
Intrusion-de
|
https://en.wikipedia.org/wiki/Experimental%20software%20engineering
|
Experimental software engineering involves running experiments on the processes and procedures involved in the creation of software systems, with the intent that the data be used as the basis of theories about the processes involved in software engineering (theory backed by data is a fundamental tenet of the scientific method). A number of research groups primarily use empirical and experimental techniques.
The term empirical software engineering emphasizes the use of empirical studies of all kinds to accumulate knowledge. Methods used include experiments, case studies, surveys, and using whatever data is available.
Empirical software engineering research
In a keynote at the International Symposium on Empirical Software Engineering and Measurement Prof. Wohlin recommended ten commitments that the research community should follow to increase the relevance and impact of empirical software engineering research. However, at the same conference Dr. Ali effectively argued that solely following these will not be enough and we need to do more than just show the evidence substantiating the claimed benefits of our interventions but instead what is required for practical relevance and potential impact is the evidence for cost-effectiveness.
The International Software Engineering Research Network (ISERN) is a global community of research groups who are active in experimental software engineering. Its purpose is to advance the practice of and foster university and industry collaborations within experimental software engineering. ISERN holds annual meetings in conjunction with the International Symposium on Empirical Software Engineering and Measurement (ESEM) conference.
|
https://en.wikipedia.org/wiki/Entrance%20facility
|
In telecommunications, Entrance facility refers to the entrance to a building for both public and private network service cables (including antenna transmission lines, where applicable), including the entrance point at the building wall or floor, and continuing to the entrance room or entrance space.
Entrance facilities are the transmission facilities (typically wires or cables) that connect competitive LECs’ networks with incumbent LECs’ networks.
Computer networking
|
https://en.wikipedia.org/wiki/Torricelli%27s%20law
|
Torricelli's law, also known as Torricelli's theorem, is a theorem in fluid dynamics relating the speed of fluid flowing from an orifice to the height of fluid above the opening. The law states that the speed of efflux of a fluid through a sharp-edged hole at the bottom of the tank filled to a depth is the same as the speed that a body (in this case a drop of water) would acquire in falling freely from a height , i.e. , where is the acceleration due to gravity. This expression comes from equating the kinetic energy gained, , with the potential energy lost, , and solving for . The law was discovered (though not in this form) by the Italian scientist Evangelista Torricelli, in 1643. It was later shown to be a particular case of Bernoulli's principle.
Derivation
Under the assumptions of an incompressible fluid with negligible viscosity, Bernoulli's principle states that the hydraulic energy is constant
at any two points in the flowing liquid. Here is fluid speed, is the acceleration due to gravity, is the height above some reference point, is the pressure, and is the density.
In order to derive Torricelli's formula the first point with no index is taken at the liquid's surface, and the second just outside the opening. Since the liquid is assumed to be incompressible, is equal to and ; both can be represented by one symbol . The pressure and are typically both atmospheric pressure, so . Furthermore
is equal to the height of the liquid's surface over the opening:
The velocity of the surface can by related to the outflow velocity by the continuity equation , where is the orifice's cross section and is the (cylindrical) vessel's cross section. Renaming to (A like Aperture) gives:
Torricelli's law is obtained as a special case when the opening is very small relative to the horizontal cross-section of the container :
Torricelli's law can only be applied when viscous effects can be neglected which is the case for water flowing out through orif
|
https://en.wikipedia.org/wiki/Black%20rose%20symbolism
|
Black roses do not naturally exist but are symbols with different meanings or for various things.
Flowers
The flowers commonly called black roses do not really exist in said color, instead they actually have a dark shade, such as the "Black Magic", "Barkarole", "Black Beauty" and "Baccara" varieties. They can be artificially colored as well.
In the language of flowers, roses have many different meanings. Black roses symbolize ideas such as hatred, despair, death or rebirths.
Anarchism
Black Rose Books is the name of the Montreal anarchist publisher and small press imprint headed by the libertarian-municipalist and anarchist Dimitrios Roussopoulos. One of the two anarchist bookshops in Sydney is Black Rose Books which has existed in various guises since 1982.
The Black Rose was the title of a respected journal of anarchist ideas published in the Boston area during the 1970s, as well as the name of an anarchist lecture series addressed by notable anarchist and libertarian socialists (including Murray Bookchin and Noam Chomsky) into the 1990s.
Black Rose Labour (organisation) is the name of a factional political organisation associated with the United Kingdom Labour Party, which defines itself as Libertarian Socialist.
Black Rose Anarchist Federation is a political organization that was founded in 2014, with a few local and regional groups in the United States.
See also
Anarchist symbolism
|
https://en.wikipedia.org/wiki/Polysorbate%2020
|
Polysorbate 20 (common commercial brand names include Kolliphor PS 20, Scattics, Alkest TW 20, Tween 20, and Kotilen-20) is a polysorbate-type nonionic surfactant formed by the ethoxylation of sorbitan monolaurate. Its stability and relative nontoxicity allows it to be used as a detergent and emulsifier in a number of domestic, scientific, and pharmacological applications. As the name implies, the ethoxylation process leaves the molecule with 20 repeat units of polyethylene glycol; in practice these are distributed across 4 different chains, leading to a commercial product containing a range of chemical species.
Food applications
Polysorbate 20 is used as a wetting agent in flavored mouth drops such as Ice Drops, helping to provide a spreading feeling to other ingredients like SD alcohol and mint flavor.
The World Health Organization has suggested acceptable daily intake limits of 0–25 mg of polyoxyethylene sorbitan esters per kg body weight.
Biotechnical applications
In biological techniques and sciences, polysorbate 20 has a broad range of applications. For example, it is used:
as a washing agent in immunoassays, such as Western blots and ELISAs. It helps to prevent non-specific antibody binding. In this major application, it is dissolved in Tris-buffered saline or phosphate buffered saline at dilutions of 0.05% to 0.5% v/v. These buffers are used for washes between each immunoreaction, to remove unbound immunologicals, and eventually for incubating solutions of immunoreagents (labeled antibodies) to reduce nonspecific background.
to saturate binding sites on surfaces (i.e., to coat polystyrene microplates, generally combined with proteins such as BSA).
to stabilize proteins purified protein derivative (PPD) solution used in skin testing for tuberculosis exposure
as a solubilizing agent of membrane proteins
for lysing mammalian cells, at a concentration of 0.05% to 0.5% v/v, generally combined with other detergents, salts and additives.
Pharmaceutical a
|
https://en.wikipedia.org/wiki/Neusis%20construction
|
In geometry, the neusis (; ; plural: ) is a geometric construction method that was used in antiquity by Greek mathematicians.
Geometric construction
The neusis construction consists of fitting a line element of given length () in between two given lines ( and ), in such a way that the line element, or its extension, passes through a given point . That is, one end of the line element has to lie on , the other end on , while the line element is "inclined" towards .
Point is called the pole of the neusis, line the directrix, or guiding line, and line the catch line. Length is called the diastema ().
A neusis construction might be performed by means of a marked ruler that is rotatable around the point (this may be done by putting a pin into the point and then pressing the ruler against the pin). In the figure one end of the ruler is marked with a yellow eye with crosshairs: this is the origin of the scale division on the ruler. A second marking on the ruler (the blue eye) indicates the distance from the origin. The yellow eye is moved along line , until the blue eye coincides with line . The position of the line element thus found is shown in the figure as a dark blue bar.
Use of the neusis
Neuseis have been important because they sometimes provide a means to solve geometric problems that are not solvable by means of compass and straightedge alone. Examples are the trisection of any angle in three equal parts, and the doubling of the cube. Mathematicians such as Archimedes of Syracuse (287–212 BC) and Pappus of Alexandria (290–350 AD) freely used neuseis; Sir Isaac Newton (1642–1726) followed their line of thought, and also used neusis constructions. Nevertheless, gradually the technique dropped out of use.
Regular polygons
In 2002, A. Baragar showed that every point constructible with marked ruler and compass lies in a tower of fields over , , such that the degree of the extension at each step is no higher than 6. Of all prime-power polygons below the 12
|
https://en.wikipedia.org/wiki/MLX%20%28software%29
|
MLX is a series of machine language entry utilities published by the magazines COMPUTE! and COMPUTE!'s Gazette, as well as books from COMPUTE! Publications. These programs were designed to allow relatively easy entry of the type-in machine language listings that were often included in these publications. Versions were available for the Commodore 64, VIC-20, Atari 8-bit family, and Apple II. MLX listings were reserved for relatively long machine language programs such as SpeedScript.
First version
MLX was introduced in the December 1983 issue of COMPUTE! for the Commodore 64 and Atari 8-bit family. This was followed in the January 1984 issue of COMPUTE!'s Gazette by a version for the VIC-20 with 8K expansion, and in the March 1984 issue by Tiny MLX, a version for the unexpanded VIC-20. These use a format consisting of six data bytes in decimal format, and a seventh as a checksum. The program auto-increments the address and prints the comma delimiters every three characters. Invalid keystrokes are ignored.
In the Commodore version, beginning in the May 1984 issue of COMPUTE!, several keyboard keys are redefined to create a makeshift numeric keypad.
Improved version
A new version of MLX was introduced for the Apple II in the June 1985 issue. This version uses an 8-byte-per-line hexadecimal format. A more sophisticated algorithm was implemented to catch errors overlooked by the original.
The improved features were then backported to the Commodore 64. The new version, known on the title screen as "MLX II", but otherwise simply as "the new MLX", appeared in the December 1985 issue of COMPUTE!. It was printed in COMPUTE!'s Gazette the following month. This version of MLX was used until COMPUTE!'s Gazette switched to a disk-only format in December 1993.
See also
The Automatic Proofreader – COMPUTE!'s checksum utility for BASIC programs
|
https://en.wikipedia.org/wiki/Selective%20availability%20anti-spoofing%20module
|
A Selective Availability Anti-spoofing Module (SAASM) is used by military Global Positioning System receivers to allow decryption of precision GPS observations, while the accuracy of civilian GPS receivers may be reduced by the United States military through Selective Availability (SA) and anti-spoofing (AS). However, on May 1, 2000 it was announced that SA was being discontinued, along with a United States Presidential Directive that no future GPS programs will include it. Before the advent of L2C, AS was meant to prevent access to dual-frequency observations to civilian users.
SAASM allows satellite authentication, over-the-air rekeying, and contingency recovery. Those features are not available with the similar, but older, PPS-SM (Precise Positioning Service Security Module) system. PPS-SM systems require periodic updates with a classified "Red Key" that may only be transmitted by secure means (such as physically taking the receiver to a secure facility for rekeying or having a trusted courier deliver a paper tape with a new key to the receiver, after which that paper tape must be securely destroyed). SAASM systems can be updated with an encrypted "Black Key" that may be transmitted over unclassified channels. All military receivers newly deployed after the end of September 2006 must use SAASM.
SAASM does not provide any additional anti-jam capability, however, the higher data (chipping) rate of P(Y) code can provide a higher processing gain which will provide better tracking performance in a jamming environment. Future GPS upgrades, such as M-Code, will provide additional improvements to anti-jam capabilities.
SAASM hardware is covered with an anti-tampering coating, to deter analysis of their internal operation.
Deployment of the next generation military signal for GPS, called M-code, commenced with the launch of IIR-M and IIF satellites, beginning in 2005. A complete constellation of 18 satellites with M-code capability is planned for 2016.
See also
D
|
https://en.wikipedia.org/wiki/A15%20phases
|
The A15 phases (also known as β-W or Cr3Si structure types) are series of intermetallic compounds with the chemical formula A3B (where A is a transition metal and B can be any element) and a specific structure. The A15 phase is also one of the members in the Frank–Kasper phases family. Many of these compounds have superconductivity at around , which is comparatively high, and remain superconductive in magnetic fields of tens of teslas (hundreds of kilogauss). This kind of superconductivity (Type-II superconductivity) is an important area of study as it has several practical applications.
History
The first time that A15 structure was observed was in 1931 when an electrolytically deposited layer of tungsten was examined. Discussion of whether the β-tungsten structure is an allotrope of tungsten or the structure of a tungsten suboxide was long-standing, but since the 1950s there has been many publications showing that the material is a true allotrope of tungsten.
The first intermetallic compound discovered with typical A3B composition was chromium silicide Cr3Si, discovered in 1933. Several other compounds with A15 structure were discovered in following years. No large interest existed in research on those compounds. This changed with the discovery that vanadium silicide V3Si showed superconductivity at around 17 K in 1953. In following years, several other A3B superconductors were found. Niobium-germanium held the record for the highest temperature of 23.2 K from 1973 until the discovery of the cuprate superconductors in 1986. It took time for the method to produce wires from the very brittle A15 phase materials to be established. This method is still complicated. Though some A15 phase materials can withstand higher magnetic field intensity and have higher critical temperatures than the NbZr and NbTi alloys, NbTi is still used for most applications due to easier manufacturing.
Nb3Sn is used for some high field applications, for example high-end MRI scanners and NM
|
https://en.wikipedia.org/wiki/Wildlife%20Reserve%20in%20Al%20Wusta
|
The Wildlife Reserve in Al Wusta, formerly the Arabian Oryx Sanctuary, is a nature reserve in the Omani Central Desert and Coastal Hills. It was included in the UNESCO World Heritage list, but became the first site to be removed from the World Heritage list in 2007.
Species inhabiting the reserve include the mountain gazelle, Nubian ibex, Arabian wolf, honey badger and caracal.
On June 28, 2007, the reserve was removed from the World Heritage Site register. UNESCO cited Oman's decision to reduce the site by 90% after oil had been found at the site, and the decline of the population of Arabian oryx from 450 in 1996 to 65 in 2007 as a result of poaching and habitat destruction. At that time, only four mating pairs remained.
|
https://en.wikipedia.org/wiki/Sclerostin
|
Sclerostin is a protein that in humans is encoded by the SOST gene. It is a secreted glycoprotein with a C-terminal cysteine knot-like (CTCK) domain and sequence similarity to the DAN (differential screening-selected gene aberrative in neuroblastoma) family of bone morphogenetic protein (BMP) antagonists. Sclerostin is produced primarily by the osteocyte but is also expressed in other tissues, and has anti-anabolic effects on bone formation.
Structure
The sclerostin protein, with a length of 213 residues, has a secondary structure that has been determined by protein NMR to be 28% beta sheet (6 strands; 32 residues).
Function
Sclerostin, the product of the SOST gene, located on chromosome 17q12–q21 in humans, was originally believed to be a non-classical bone morphogenetic protein (BMP) antagonist. More recently, sclerostin has been identified as binding to LRP5/6 receptors and inhibiting the Wnt signaling pathway. The inhibition of the Wnt pathway leads to decreased bone formation. Although the underlying mechanisms are unclear, it is believed that the antagonism of BMP-induced bone formation by sclerostin is mediated by Wnt signaling, but not BMP signaling pathways. Sclerostin is expressed in osteocytes and some chondrocytes and it inhibits bone formation by osteoblasts.
Sclerostin production by osteocytes is inhibited by parathyroid hormone, mechanical loading, estrogen and cytokines including prostaglandin E2, oncostatin M, cardiotrophin-1 and leukemia inhibitory factor. Sclerostin production is increased by calcitonin. Thus, osteoblast activity is self regulated by a negative feedback system.
Clinical significance
Mutations in the gene that encodes the sclerostin protein are associated with disorders associated with high bone mass, sclerosteosis and van Buchem disease.
van Buchem disease is an autosomal recessive skeletal disease characterized by bone overgrowth. It was first described in 1955 as "hyperostosis corticalis generalisata familiaris", but w
|
https://en.wikipedia.org/wiki/Work%20behavior
|
Work behavior is the behavior one uses in employment and is normally more formal than other types of human behavior. This varies from profession to profession, as some are far more casual than others. For example, a computer programmer would usually have far more leeway in their work behavior than a lawyer.
People are usually more careful than outside work in how they behave around their colleagues, as many actions intended to be in jest can be perceived as inappropriate or even harassment in the work environment. In some cases, men may take considerably more care so as not to be perceived as being sexually harassing than they would ordinarily.
Work behavior is one of the significant aspects of Human Behavior. It is an individual's communication towards the rest of the members of the work place. It involves both verbal as well as non-verbal mode of communication. For example, trust is a non-verbal behavior which is often reflected by a verbal communication at a work place. It represents your attitude towards your team and colleagues. A positive and good work behavior of an individual leads to higher performance, productivity and great outputs by the team or an individual. From the organizational perspective it is the most important area where Human Resource managers should focus.
Counterproductive work behavior
Counterproductive work behavior is also a type of work behavior. The majority of people do not know what counterproductive work behavior is. Counterproductive work behavior is the act that employees have against the organizations that do harm or violate the work production. Some examples of Counterproductive work behavior would include passive actions such as not working to meet date line or faking incompetence. Even people do not recognize this behavior, it seems normal to them. Some examples of counterproductive behavior are:
Intimate partner violence: Intimate partner violence occurs more often in the workplace. About 36% to 75% of employed women who
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.