source
stringlengths 31
227
| text
stringlengths 9
2k
|
---|---|
https://en.wikipedia.org/wiki/Forbush%20Man
|
Forbush Man (spelled Forbush-Man in his early appearances) is a fictional character appearing in American comic books published by Marvel Comics. Originally the mascot of Marvel's Not Brand Echh, he is the alter-ego of Irving Forbush, a fictional employee of "Marble Comics" (a parody of Marvel). Forbush was devised in 1955 by Marvel editor Stan Lee to refer to an imaginary low-grade colleague who was often the butt of Lee's jokes. In his guise of Forbush-Man, he first appeared in 1967.
According to Marvel Comics' Alternate Universes 2005, Forbush Man is a native of Earth-665 as opposed to Marvel's regular Earth-616.
Publication history
Irving Forbush was introduced in Marvel's short-lived satirical comic book Snafu as a mascot. Forbush was given a line in the magazine's content page where he was credited as Snafu's founder. Another Forbush family member, Melvin, was mentioned in the letters column reference, "Losted [sic] by his cousin, Melvin Forbush". During Snafu's three-issue run, starting in November 1955, the "actual face" of Irving Forbush was often shown, though this face was of someone not named Irving Forbush.
Forbush-Man first appeared on the cover of the first issue of the satirical Not Brand Echh (cover-dated Aug. 1967), drawn by Jack Kirby and featuring Doctor Doom, the Fantastic Four and the Silver Surfer cowering in fear as Forbush Man approaches. Forbush-Man is a wannabe superhero with no superpowers who wears a costume comprising red long johns with the letter F on the front, black galoshes and a cooking pot with eye-holes on his head.
Forbush-Man's first major appearance was in the lead story of Not Brand Echh #5 (Dec. 1967): "The Origin of Forbush-Man", which was "conceived, created and cluttered-up" by Lee and Kirby. In this story, Forbush-Man's secret identity is revealed as Irving Forbush, the fictitious office gofer at Marble Comics. The character has a shrewish maiden aunt (Auntie Mayhem) who is indirectly responsible for her nephew beco
|
https://en.wikipedia.org/wiki/Ecology%20of%20Banksia
|
The ecology of Banksia is the relationships and interactions among the plant genus Banksia and its environment. Banksia has a number of adaptations that have so far enabled the genus to survive despite dry, nutrient-poor soil, low rates of seed set, high rates of seed predation and low rates of seedling survival. These adaptations include proteoid roots and lignotubers; specialised floral structures that attract nectariferous animals and ensure effective pollen transfer; and the release of seed in response to bushfire.
The arrival of Europeans in Australia has brought new ecological challenges. European colonisation of Australia has directly affected Banksia through deforestation, exploitation of flowers and changes to the fire regime. In addition, the accidental introduction and spread of plant pathogens such as Phytophthora cinnamomi (dieback) pose a serious threat to the genus's habitat and biodiversity. Various conservation measures have been put in place to mitigate these threats, but a number of taxa remain endangered.
Background
Banksia is a genus of around 170 species in the plant family Proteaceae. An iconic Australian wildflower and popular garden plant, Banksias are most commonly associated with their elongate flower spikes and fruiting "cones", although less than half of Banksia species possess this feature. They grow in forms varying from prostrate woody shrubs to trees up to 30 metres tall, and occur in all but the most arid areas of Australia.
Pollination
The pollination ecology of Banksia has been well studied, because the large showy inflorescences make it easy to conduct pollination experiments, and the pollination roles of nectariferous birds and mammals makes the genus a popular subject for zoologists.
Visits to Banksia inflorescences by western honeybees and nectarivorous birds are often observed and are obviously important to pollination. Also important are visits by nectariferous mammals, although such visits are rarely observed becaus
|
https://en.wikipedia.org/wiki/Parasitic%20plant
|
A parasitic plant is a plant that derives some or all of its nutritional requirements from another living plant. They make up about 1% of angiosperms and are found in almost every biome. All parasitic plants develop a specialized organ called the haustorium, which penetrates the host plant, connecting them to the host vasculature – either the xylem, phloem, or both. For example, plants like Striga or Rhinanthus connect only to the xylem, via xylem bridges (xylem-feeding). Alternately, plants like Cuscuta and some members of Orobanche connect to both the xylem and phloem of the host. This provides them with the ability to extract resources from the host. These resources can include water, nitrogen, carbon and/or sugars. Parasitic plants are classified depending on the location where the parasitic plant latches onto the host (root or stem), the amount of nutrients it requires, and their photosynthetic capability. Some parasitic plants can locate their host plants by detecting volatile chemicals in the air or soil given off by host shoots or roots, respectively. About 4,500 species of parasitic plants in approximately 20 families of flowering plants are known.
There is a wide range of effects that may occur to a host plant due to the presence of a parasitic plant. Often there is a pattern of stunted growth in hosts especially in hemi-parasitic cases, but may also result in higher mortality rates in host plant species following introduction of larger parasitic plant populations.
Classification
Parasitic plants occur in multiple plant families, indicating that the evolution is polyphyletic. Some families consist mostly of parasitic representatives such as Balanophoraceae, while other families have only a few representatives. One example is the North American Monotropa uniflora (Indian pipe or corpse plant) which is a member of the heath family, Ericaceae, better known for its member blueberries, cranberries, and rhododendrons.
Parasitic plants are characterized as
|
https://en.wikipedia.org/wiki/O-ring%20boss%20seal
|
An o-ring boss seal is a technique for joining two fluid-carrying pipes, hoses, or tubing. In an o-ring boss (abbreviated ORB) system, a male-threaded part is inserted into a female-threaded part, providing a mechanical seal. This system differs from others in that a nut is tightened over an o-ring in a chamfered area, creating a fluid-tight seal.
Application
This system is used frequently in hydraulics, although it has been applied to other systems including compressed air systems and vacuum pumps, such as many Robinair pumps, in which the intake tee has an o-ring boss seal on the bottom. The ORB system can be confused with other connection systems, such as NPT. While threads of different connectors sometimes fit (although often very inexactly), o-ring boss seal system connectors should never be used with any other type of connectors and vice versa, as leaks are likely. Under the high fluid pressures commonly seen in hydraulic systems, a leak or failure of the connection is quite dangerous and could lead to loss of life.
This system has the advantage of being able to be tightened mechanically before being sealed. Most threaded systems, such as NPT, have a seal provided by a taper in the thread, so it is difficult to orient both ends of the hose, pipe or tube so that it is not twisted. In the o-ring boss system, this problem is eliminated because the threads do not seal the connection and therefore can be rotated at least a full revolution before they are sealed while maintaining a proper mechanical connection. The orientation problem could also be solved with a suitable union.
|
https://en.wikipedia.org/wiki/Quasitopological%20space
|
In mathematics, a quasi-topology on a set X is a function that associates to every compact Hausdorff space C a collection of mappings from C to X satisfying certain natural conditions. A set with a quasi-topology is called a quasitopological space.
They were introduced by Spanier, who showed that there is a natural quasi-topology on the space of continuous maps from one space to another.
|
https://en.wikipedia.org/wiki/Concatemer
|
A concatemer is a long continuous DNA molecule that contains multiple copies of the same DNA sequence linked in series. These polymeric molecules are usually copies of an entire genome linked end to end and separated by cos sites (a protein binding nucleotide sequence that occurs once in each copy of the genome). Concatemers are frequently the result of rolling circle replication, and may be seen in the late stage of infection of bacteria by phages. As an example, if the genes in the phage DNA are arranged ABC, then in a concatemer the genes would be ABCABCABCABC and so on (assuming synthesis was initiated between genes C and A). They are further broken by ribozymes.
During active infection, some species of viruses have been shown to replicate their genetic material via the formation of concatemers. In the case of human herpesvirus-6, its entire genome is made over and over on a single strand. These long concatemers are subsequently cleaved between the pac-1 and pac-2 regions by ribozymes when the genome is packaged into individual virions.
Bacteriophage T4 replicating DNA was labeled with tritiated thymidine and examined by autoradiography. The observed DNA replication intermediates included circular and branched circular concatemeric structures that likely arose by rolling circle replication.
When assembling concatemers from synthetic oligonucleotides, increasing salt concentration to 200 mM was found to be a major optimizing factor due to its ability to enhance ionic strength, which hastened the formation of concatemers.
|
https://en.wikipedia.org/wiki/Communication%20physics
|
Communication physics is one of the applied branches of physics. It deals with various kinds of communication systems. These can range from basic ideas such as mobile phone communication to quantum communication via quantum entanglement. Communication physics is also a journal edition created in 2018 published by Nature Research that aims to publish research that involves a different way of thinking in the research field.
Applications
Communication physics aims to study and explain how a communication system works. This can be applied in a hard science way via Computer Communication or in the way of how people communicate.
An example of communication physics is how computers can transmit and receive data through networks. This would also deal with explaining how these devices encode and decode messages.
See also
Electronic communication
Optical communication
Computer communication
Telephone
Telegraph
Radio
Television
Mobile phone communication
Nanoscale network
|
https://en.wikipedia.org/wiki/Neighborhood%20semantics
|
Neighborhood semantics, also known as Scott–Montague semantics, is a formal semantics for modal logics. It is a generalization, developed independently by Dana Scott and Richard Montague, of the more widely known relational semantics for modal logic. Whereas a relational frame consists of a set W of worlds (or states) and an accessibility relation R intended to indicate which worlds are alternatives to (or, accessible from) others, a neighborhood frame still has a set W of worlds, but has instead of an accessibility relation a neighborhood function
that assigns to each element of W a set of subsets of W. Intuitively, each family of subsets assigned to a world are the propositions necessary at that world, where 'proposition' is defined as a subset of W (i.e. the set of worlds at which the proposition is true). Specifically, if M is a model on the frame, then
where
is the truth set of .
Neighborhood semantics is used for the classical modal logics that are strictly weaker than the normal modal logic K.
Correspondence between relational and neighborhood models
To every relational model M = (W, R, V) there corresponds an equivalent (in the sense of having pointwise-identical modal theories) neighborhood model M' = (W, N, V) defined by
The fact that the converse fails gives a precise sense to the remark that neighborhood models are a generalization of relational ones. Another (perhaps more natural) generalization of relational structures are general frames.
|
https://en.wikipedia.org/wiki/VPS/VM
|
VPS/VM (Virtual Processing System/Virtual Machine) was an operating system that ran on IBM System/370 – IBM 3090 computers at Boston University in general use from 1977 to around 1990, and in limited use until at least 1993. During the 1980s, VPS/VM was the main operating system of Boston University and often ran up to 250 users at a time when rival VM/CMS computing systems could only run 120 or so users.
Each user ran in a Virtual Machine under VM, an IBM hypervisor operating system. VM provided the virtual IBM 370 machine which the VPS operating system ran under. The VM code was modified to allow all the VPS virtual machines to share pages of storage with read and write access. VPS utilized a shared nucleus, as well as pages used to facilitate passing data from one VPS virtual machine to another. This organization is very similar to that of MVS; substituting Address Spaces for Virtual Machines.
Origins
According to Craig Estey, who worked at the Boston University Academic Computing Center between 1974 and 1977:
Description
An IBM-based operating system, and quite like some DOS/VSE time sharing options, VPS/VM provided the user an IBM 3270 full screen terminal (a green screen) and a user interface that was like VM/CMS. Each user had an 11 megabyte virtual machine (with a strange 3 megabyte memory gap in the middle) and, from 1984 onwards, could run several programs at a time.
The operating system was sparsely documented but was written first by Charles Brown, a BU doctoral student, and John H. Porter, a physics PHD, who later became the head of the VPS project (and eventually Boston University's vice president for information systems and technology). Marian Moore wrote much of the later VM code necessary to run the VPS system.
Josie Bondoc wrote some of the later VPS additions, like UNIX piping.
Many MVS/VM programs ran on VPS/VM, such as XEDIT, and compilers for Pascal, PL/1, C and Cobol. These MVS/VM programs ran under an OS simulation program that simula
|
https://en.wikipedia.org/wiki/In-game%20advertising
|
In-game advertising (IGA) is advertising in electronic games. IGA differs from advergames, which refers to games specifically made to advertise a product. The IGA industry is large and growing.
In-game advertising generated $34 million in 2004, $56 million in 2005, $80 million in 2006,
and $295 million in 2007.
In 2009, spending on IGA was estimated to reach $699 million USD, $1 billion by 2014 and according to Forbes is anticipated to grow to $7.2 billion by 2016.
The earliest known IGA was the 1978 computer game Adventureland, which inserted a self-promotional advertisement for its next game, Pirate Adventure.
IGA can be integrated into the game either through a display in the background, such as an in-game billboard or a commercial during the pause created when a game loads, or highly integrated within the game so that the advertised product is necessary to complete part of the game or is featured prominently within cutscenes. Due to the custom programming required, dynamic advertising is usually presented in the background; static advertisements can appear as either. One of the advantages of IGA over traditional advertisements is that consumers are less likely to multitask with other media while playing a game, however, some attention is still divided between the gameplay, controls, and the advertisement.
Static in-game advertising
Similar to product placement in the film industry, static IGAs cannot be changed after they are programmed directly into the game (unless it's completely online). However, unlike product placement in traditional media, IGA allows gamers to interact with the virtual product. For example, Splinter Cell has required the use of in-game Sony Ericsson phones to catch terrorists. Unlike static IGAs, dynamic IGAs are not limited to a developer and publisher determined pre-programmed size or location and allow the advertiser to customize the advertisement display.
A number of games utilize billboard-like advertisements or product pl
|
https://en.wikipedia.org/wiki/Structural%20dynamics
|
Structural dynamics is a type of structural analysis which covers the behavior of a structure subjected to dynamic (actions having high acceleration) loading. Dynamic loads include people, wind, waves, traffic, earthquakes, and blasts. Any structure can be subjected to dynamic loading. Dynamic analysis can be used to find dynamic displacements, time history, and modal analysis.
Structural analysis is mainly concerned with finding out the behavior of a physical structure when subjected to force. This action can be in the form of load due to the weight of things such as people, furniture, wind, snow, etc. or some other kind of excitation such as an earthquake, shaking of the ground due to a blast nearby, etc. In essence all these loads are dynamic, including the self-weight of the structure because at some point in time these loads were not there. The distinction is made between the dynamic and the static analysis on the basis of whether the applied action has enough acceleration in comparison to the structure's natural frequency. If a load is applied sufficiently slowly, the inertia forces (Newton's first law of motion) can be ignored and the analysis can be simplified as static analysis.
A static load is one which varies very slowly. A dynamic load is one which changes with time fairly quickly in comparison to the structure's natural frequency. If it changes slowly, the structure's response may be determined with static analysis, but if it varies quickly (relative to the structure's ability to respond), the response must be determined with a dynamic analysis.
Dynamic analysis for simple structures can be carried out manually, but for complex structures finite element analysis can be used to calculate the mode shapes and frequencies.
Displacements
A dynamic load can have a significantly larger effect than a static load of the same magnitude due to the structure's inability to respond quickly to the loading (by deflecting). The increase in the effect of a dynami
|
https://en.wikipedia.org/wiki/Autacoid
|
Autacoids or autocoids are biological factors (molecules) which act like local hormones, have a brief duration, and act near their site of biosynthesis. The word autacoid comes from the Greek words "autos" (self) and "acos" (relief; i.e., drug). The effects of autacoids are primarily local, though large quantities can be produced and moved into circulation. Autacoids may thus have systemic effects by being transported via the circulation. These regulating molecules are also metabolized locally. In sum, these compounds typically are produced locally, act locally and are metabolized locally. Autacoids can have a variety of different biological actions, including modulating the activities of smooth muscles, glands, nerves, platelets and other tissues.
Some autacoids are chiefly characterized by the effect they have on specific tissues, such as smooth muscle. With respect to vascular smooth muscle, there exist both vasoconstrictor and vasodilator autacoids. Vasodilator autacoids are released during periods of exercise. Their main effect is seen in the skin, where they facilitate heat loss.
These are local hormones; they therefore have a paracrine effect. Some notable autacoids are: eicosanoids, angiotensin, neurotensin, NO (nitric oxide), kinins, histamine, serotonin, endothelins and palmitoylethanolamide.
In 2015, a more precise definition of autacoids was proposed: "An autacoid is a locally produced modulating factor, influencing locally the function of cells and/or tissues, which is produced on demand and which subsequently is metabolized in the same cells and/or tissues".
|
https://en.wikipedia.org/wiki/Adenylate%20cyclase%20toxin
|
Adenylate cyclase toxin is a virulence factor produced by some members of the genus Bordetella. Together with the pertussis toxin it is the most important virulence factor of the causative agent of whooping cough, Bordetella pertussis. Bordetella bronchiseptica and Bordetella parapertussis, also able to cause pertussis-like symptoms, also produce adenylate cyclase toxin. It is a toxin secreted by the bacteria to influence the host immune system.
Structure
Adenylate cyclase toxin from Bordetella pertussis is a 1706 amino acid residue long protein. The protein consists of three domains: from the N-terminus up to roughly residue 400, there is an adenylate-cyclase domain; between residues 500 and 700, there is a hydrophobic domain; and from residue 1000 to the C-terminus, there are calcium binding repeats. Two acylation sites are located at lysine residues K860 and K983. The part of the toxin from residue 400 to the C-terminus, called hemolysin, is structurally related to a large family of bacterial toxins - RTX toxins. Differences between the toxins of different Bordetella species are mainly in the calcium-binding domain.
Folding and secretion
The toxin is secreted by the Type I secretion system, which spans both membranes and periplasm space, allowing the toxin to be secreted from the cytoplasm straight outside the cell. A large proportion of the toxin remains associated with the bacterium exterior proteins, mainly filamentous haemagglutinin, but these toxin molecules are not active. Besides attachment to bacterial proteins, aggregation also inactivates the toxin. This quick inactivation highlights the necessity of close contact between secreting bacterium and target cell.
RTX toxins
RTX stands for 'repeats in toxins,' but not all members of the family are toxins. Repeating aspartate and glycine rich nonapeptides (repeats 9 amino acids long) are a characteristic feature of this family of proteins, and are able to bind calcium ions. A feature of the RTX proteins
|
https://en.wikipedia.org/wiki/Physalaemin
|
Physalaemin is a tachykinin peptide obtained from the Physalaemus frog, closely related to substance P. Its structure was first elucidated in 1964.
Like all tachykinins, physalaemin is a sialagogue (increases salivation) and a potent vasodilator with hypotensive effects.
Structure
Physalaemin (PHY) is known to take on both a linear and helical three dimensional structure. Grace et al. (2010) have shown that in aqueous environments, PHY preferentially takes on the linear conformation whereas in an environment that simulates a cellular membrane, PHY takes on a helical confirmation from the Pro4 residue to the C-Terminus. This helical conformation is essential to allow the binding of PHY to neurokinin-1 (NK1) receptors. Consensus sequences between Substance P (a mammalian tachykinin and agonist of NK1) and PHY have been used to confirm that the helical confirmation is necessary for PHY to bind to NK1.
Use In Research
Not only is PHY closely related to Substance P (SP), but it also has a higher affinity for the mammalian neurokinin receptors that Substance P can bind to. Researchers can make use of this behavior of PHY to study the behavior of smooth muscle - a tissue where NK1 can be found. Shiina et al. (2010) used PHY to show that tachykinins as a whole can cause the longitudinal contraction of smooth muscle tissue in esophageal tissue.
Singh et Maji made use of PHY's similarity to SP along with its sequence similarity to Amyloid B-peptide 25-35 [AB(25-35)]. Despite its sequence similarity to SP, Singh et Maji showed that PHY had distinct amyloid forming capabilities . Under artificially elevated concentrations of tetrafluoroethylene (TFE) and a short incubation time, PHY was able to form amyloid fibrils. These fibrils originating from tackynins like PHY were also shown to reduce the neurotoxicity of other Amyloid fibers associated with amyloid induced diseases such as Alzheimer's disease.
|
https://en.wikipedia.org/wiki/Volterra%20series
|
The Volterra series is a model for non-linear behavior similar to the Taylor series. It differs from the Taylor series in its ability to capture "memory" effects. The Taylor series can be used for approximating the response of a nonlinear system to a given input if the output of the system depends strictly on the input at that particular time. In the Volterra series, the output of the nonlinear system depends on the input to the system at all other times. This provides the ability to capture the "memory" effect of devices like capacitors and inductors.
It has been applied in the fields of medicine (biomedical engineering) and biology, especially neuroscience. It is also used in electrical engineering to model intermodulation distortion in many devices, including power amplifiers and frequency mixers. Its main advantage lies in its generalizability: it can represent a wide range of systems. Thus, it is sometimes considered a non-parametric model.
In mathematics, a Volterra series denotes a functional expansion of a dynamic, nonlinear, time-invariant functional. The Volterra series are frequently used in system identification. The Volterra series, which is used to prove the Volterra theorem, is an infinite sum of multidimensional convolutional integrals.
History
The Volterra series is a modernized version of the theory of analytic functionals from the Italian mathematician Vito Volterra, in his work dating from 1887. Norbert Wiener became interested in this theory in the 1920s due to his contact with Volterra's student Paul Lévy. Wiener applied his theory of Brownian motion for the integration of Volterra analytic functionals. The use of the Volterra series for system analysis originated from a restricted 1942 wartime report of Wiener's, who was then a professor of mathematics at MIT. He used the series to make an approximate analysis of the effect of radar noise in a nonlinear receiver circuit. The report became public after the war. As a general method of anal
|
https://en.wikipedia.org/wiki/Cyclonic%20rotation
|
Cyclonic rotation, or cyclonic circulation, is atmospheric motion in the same direction as a planet's rotation, as opposed to anticyclonic rotation. In the case of Earth's rotation, the Coriolis effect causes cyclonic rotation to be in a counterclockwise direction in the Northern Hemisphere and clockwise in the Southern Hemisphere. A closed area of winds rotating cyclonically is known as a cyclone.
|
https://en.wikipedia.org/wiki/Electrical%20capacitance%20tomography
|
Electrical capacitance tomography (ECT) is a method for determination of the dielectric permittivity distribution in the interior of an object from external capacitance measurements. It is a close relative of electrical impedance tomography and is proposed as a method for industrial process monitoring.
Although capacitance sensing methods were in widespread use the idea of using capacitance measurement to form images is attributed to Maurice Beck and co-workers at UMIST in the 1980s.
Although usually called tomography, the technique differs from conventional tomographic methods, in which high resolution images are formed of slices of a material. The measurement electrodes, which are metallic plates, must be sufficiently large to give a measureable change in capacitance. This means that very few electrodes are used, typically eight to sixteen electrodes. An N-electrode system can only provide N(N−1)/2 independent measurements. This means that the technique is limited to producing very low resolution images of approximate slices. However, ECT is fast, and relatively inexpensive.
Applications
Applications of ECT include the measurement of flow of fluids in pipes and measurement of the concentration of one fluid in another, or the distribution of a solid in a fluid. ECT enables the visualization of multiphase flow, which play an important role in the technological processes of the chemical, petrochemical and food industries.
Due to its very low spatial resolution, ECT has not yet been used in medical diagnostics. Potentially, ECT may have similar medical applications to electrical impedance tomography, such as monitoring lung function or detecting ischemia or hemorrhage in the brain.
See also
Electrical capacitance volume tomography
Electrical impedance tomography
Electrical resistivity tomography
Industrial Tomography Systems
Process tomography
|
https://en.wikipedia.org/wiki/Riesz%20mean
|
In mathematics, the Riesz mean is a certain mean of the terms in a series. They were introduced by Marcel Riesz in 1911 as an improvement over the Cesàro mean. The Riesz mean should not be confused with the Bochner–Riesz mean or the Strong–Riesz mean.
Definition
Given a series , the Riesz mean of the series is defined by
Sometimes, a generalized Riesz mean is defined as
Here, the are a sequence with and with as . Other than this, the are taken as arbitrary.
Riesz means are often used to explore the summability of sequences; typical summability theorems discuss the case of for some sequence . Typically, a sequence is summable when the limit exists, or the limit exists, although the precise summability theorems in question often impose additional conditions.
Special cases
Let for all . Then
Here, one must take ; is the Gamma function and is the Riemann zeta function. The power series
can be shown to be convergent for . Note that the integral is of the form of an inverse Mellin transform.
Another interesting case connected with number theory arises by taking where is the Von Mangoldt function. Then
Again, one must take c > 1. The sum over ρ is the sum over the zeroes of the Riemann zeta function, and
is convergent for λ > 1.
The integrals that occur here are similar to the Nörlund–Rice integral; very roughly, they can be connected to that integral via Perron's formula.
|
https://en.wikipedia.org/wiki/Closed-loop%20pole
|
In systems theory, closed-loop poles are the positions of the poles (or eigenvalues) of a closed-loop transfer function in the s-plane. The open-loop transfer function is equal to the product of all transfer function blocks in the forward path in the block diagram. The closed-loop transfer function is obtained by dividing the open-loop transfer function by the sum of one and the product of all transfer function blocks throughout the negative feedback loop. The closed-loop transfer function may also be obtained by algebraic or block diagram manipulation. Once the closed-loop transfer function is obtained for the system, the closed-loop poles are obtained by solving the characteristic equation. The characteristic equation is nothing more than setting the denominator of the closed-loop transfer function to zero.
In control theory there are two main methods of analyzing feedback systems: the transfer function (or frequency domain) method and the state space method. When the transfer function method is used, attention is focused on the locations in the s-plane where the transfer function is undefined (the poles) or zero (the zeroes; see Zeroes and poles). Two different transfer functions are of interest to the designer. If the feedback loops in the system are opened (that is prevented from operating) one speaks of the open-loop transfer function, while if the feedback loops are operating normally one speaks of the closed-loop transfer function. For more on the relationship between the two, see root-locus.
Closed-loop poles in control theory
The response of a linear time-invariant system to any input can be derived from its impulse response and step response. The eigenvalues of the system determine completely the natural response (unforced response). In control theory, the response to any input is a combination of a transient response and steady-state response. Therefore, a crucial design parameter is the location of the eigenvalues, or closed-loop poles.
|
https://en.wikipedia.org/wiki/Ricker%20model
|
The Ricker model, named after Bill Ricker, is a classic discrete population model which gives the expected number N t+1 (or density) of individuals in generation t + 1 as a function of the number of individuals in the previous generation,
Here r is interpreted as an intrinsic growth rate and k as the carrying capacity of the environment. Unlike some other models like the Logistic map, the carrying capacity in the Ricker model is not a hard barrier that cannot be exceeded by the population, but it only determines the overall scale of the population. The Ricker model was introduced in 1954 by Ricker in the context of stock and recruitment in fisheries.
The model can be used to predict the number of fish that will be present in a fishery. Subsequent work has derived the model under other assumptions such as scramble competition, within-year resource limited competition or even as the outcome of source-sink Malthusian patches linked by density-dependent dispersal. The Ricker model is a limiting case of the Hassell model which takes the form
When c = 1, the Hassell model is simply the Beverton–Holt model.
See also
Population dynamics of fisheries
Notes
|
https://en.wikipedia.org/wiki/Longevity%20fruit
|
Longevity fruit may refer to the fruit of:
Siraitia grosvenorii
Lycium species:
Lycium barbarum
Lycium chinense
Arachis hypogaea
Longevity
|
https://en.wikipedia.org/wiki/Strabismus%20%28protein%29
|
Strabismus was originally identified as a Drosophila protein involved in planar cell polarity. Flies with mutated strabismus genes have altered development of ommatidia in their eyes. Vertebrates have two Strabismus-related proteins, VANGL1 and VANGL2 (an alternate name for the Drosophila "Strabismus" protein is "Van Gogh").
The amino acid sequence and localization studies for Strabismus indicate that it is a membrane protein. Prickle is another protein in the planar cell polarity signaling pathway. Prickle is recruited to the cell surface membrane by strabismus. In cells of the developing Drosophila wing, Prickle and Strabismus are concentrated at the cell surface membrane on the most proximal side of cells.
Vertebrate cell movement
VANGL2 is involved in the migration of groups of cells during vertebrate embryogenesis.
Humans
In humans, mutations in VANGL1 have been associated with neural tube defects including spina bifida, and with some forms of cancer including hepatocellular carcinoma.
|
https://en.wikipedia.org/wiki/List%20of%20Inferno%20applications
|
This is a list of Inferno programs. Most of these programs are very similar to the Plan 9 applications or UNIX programs with the same name.
System software
General user
dd – convert and copy a file
date – print the date
echo – print arguments
emu – Inferno emulator
mash – programmable shell
ns – display current namespace
– build Inferno namespace
os – interface to host OS commands (hosted Inferno only)
plumb – send message to plumber
plumber – plumber for interapplication message routing
rcmd – remote command execution
runas – run command as another user
sh – command language
tiny/sh – reduced command line interface to the Inferno system
wm/logon – log on to Inferno
wm/sh, wm/mash – Window frames for the Inferno shells
wm/wm – window manager
System Management
Processes and tasks management
time – time command execution
kill, broke – terminate processes
sleep, pause – suspend execution for an interval
ps – process (thread) status
wm/task – graphical task manager
User management and support
auth/passwd – change user password
man, wm/man, man2txt, lookman – print or find manual pages
Files and Text
Filesystem Utilities
chgrp – change file's group or owner
chmod – change file mode (permissions)
cp, fcp – copy files
du – disk usage
lc – list files in columns
ls – list files
mkdir – make a directory
mv – move files
bind, mount, unmount – change name space
pwd – working directory
rm – remove files
touch – update the modification time of one or more files
Archivers and compression
ar – archive maintainer
gettar, lstar, puttar – tar archive utilities
gzip, gunzip – compression and decompression utilities
Text Processing
cat – concatenate files
cmp – compare two files
diff – differential file comparator
fmt – simple text formatter
freq – print histogram of character frequencies
grep – pattern matching
p – paginate
read – read from standard input with optional seek
tail – deliver the last part of a file
tcs – translate character sets
tr – translate characters
w
|
https://en.wikipedia.org/wiki/Joseph%20Bernstein
|
Joseph Bernstein (sometimes spelled I. N. Bernshtein; ; ; born 18 April 1945) is a Soviet-born Israeli mathematician working at Tel Aviv University. He works in algebraic geometry, representation theory, and number theory.
Biography
Bernstein received his Ph.D. in 1972 under Israel Gelfand at Moscow State University. In 1981, he emigrated to the United States due to growing antisemitism in the Soviet Union.
Bernstein was a professor at Harvard during 1983-1993.
He was a visiting scholar at the Institute for Advanced Study in 1985-86 and again in 1997-98. In 1993, he moved to Israel to take a professorship at Tel Aviv University (emeritus since 2014).
Awards and honors
Bernstein received a gold medal at the 1962 International Mathematical Olympiad. He was elected to the Israel Academy of Sciences and Humanities in 2002 and was elected to the United States National Academy of Sciences in 2004. In 2004, Bernstein was awarded the Israel Prize for mathematics. In 1998, he was an Invited Speaker of the International Congress of Mathematicians in Berlin. In 2012, he became a fellow of the American Mathematical Society.
Publications
Publication list
Some pdf files of papers by Bernstein including Algebraic theory of D-modules and his notes on Meromorphic continuation of Eisenstein series
See also
Bernstein–Sato polynomial
Bernstein–Gelfand–Gelfand resolution
|
https://en.wikipedia.org/wiki/Fazio%E2%80%93Londe%20disease
|
Fazio–Londe disease (FLD), also called progressive bulbar palsy of childhood, is a very rare inherited motor neuron disease of children and young adults and is characterized by progressive paralysis of muscles innervated by cranial nerves. FLD, along with Brown–Vialetto–Van Laere syndrome (BVVL), are the two forms of infantile progressive bulbar palsy, a type of progressive bulbar palsy in children.
Signs and symptoms
FLD produces rapidly progressive weakness of tongue, face and pharyngeal muscles in a clinical pattern similar to myasthenia. Neuromuscular transmission may be abnormal in these muscles because of rapid denervation and immature reinnervation. Paralysis occurs secondary to degeneration of the motor neurons of the brain stem. It causes progressive bulbar paralysis due to involvement of motor neurons of the cranial nerve nuclei. The most frequent symptoms at onset of progressive bulbar paralysis of childhood has been a unilateral facial paralysis. It is followed in frequency by dysarthria due to facial weakness or by dysphagia. Palatal weakness and palpebral ptosis also have been reported in few patients. Both sexes can be affected.
Genetics
Fazio–Londe disease is linked to a genetic mutation in the SLC52A3 gene on chromosome 20 (locus: 20p13). It is allelic and phenotypically similar to Brown–Vialetto–Van Laere syndrome. The condition is inherited in an autosomal recessive manner. The gene encodes the intestinal riboflavin transporter (hRFT2).
Diagnosis
Symptoms of Fazio–Londe include bulbar palsy, hearing loss, facial weakness, and difficulty breathing. The disease is caused by mutations in the SLC52A2 gene and the SLC52A1 (GPR172B) genes which code for hRFT3 and hRFT1, human riboflavin transporters. Only muscle biopsy and examination of the transporter genes is considered to provide a definitive diagnosis. However, because the disease is so often fatal without treatment, and because the treatment is so inexpensive and with little risk, it is recomm
|
https://en.wikipedia.org/wiki/Center%20of%20curvature
|
In geometry, the center of curvature of a curve is found at a point that is at a distance from the curve equal to the radius of curvature lying on the normal vector. It is the point at infinity if the curvature is zero. The osculating circle to the curve is centered at the centre of curvature. Cauchy defined the center of curvature C as the intersection point of two infinitely close normal lines to the curve. The locus of centers of curvature for each point on the curve comprise the evolute of the curve. This term is generally used in physics regarding the study of lenses and mirrors (see radius of curvature (optics)).
It can also be defined as the spherical distance between the point at which all the rays falling on a lens or mirror either seems to converge to (in the case of convex lenses and concave mirrors) or diverge from (in the case of concave lenses or convex mirrors) and the lens/mirror itself.
See also
Curvature
Differential geometry of curves
|
https://en.wikipedia.org/wiki/Amanita%20gemmata
|
Amanita gemmata, commonly known as the gemmed amanita or the jonquil amanita, is an agaric mushroom of the family Amanitaceae and genus Amanita. The fruit body has a cap that is a dull to golden shade of yellow, and typically in diameter. The cap surface is sticky when moist, and characterized by white warts, which are easily detached. It is initially convex, and flattens out when mature. The flesh is white and does not change colour when cut. The gills are white and closely spaced. The stem is pale yellow, and measures long by thick. The partial veil that covers the young fruit body turns into the ring on the stem at maturity. The spore print is white, while the spores are roughly elliptical, and measure 8–10 by 6.5–7.5 μm.
This species is a mycorrhizal fungus, widespread in Europe. It can grow either singly, scattered, or in groups. It prefers habitats like coniferous and mixed forests and alongside paths, where it fruits in summer and fall. It is a toxic mushroom, containing ibotenic acid and muscimol, also found in many species in section Amanita of the Amanita genus including Amanita muscaria and A. pantherina. It is often confused with various other European species. A. gemmata resembles the false death cap, tawny grisette and panther cap mushrooms. Its cap is brighter in color than the former, and more yellow than the latter two.
Taxonomy and phylogeny
The species was first described scientifically by Swedish mycologist and botanist Elias Magnus Fries as Agaricus gemmatus in 1838. It was transferred to the genus Amanita in 1866 by the French statistician Louis Bertillon. The species has been transferred to several genera in its history, resulting in a number of synonyms, including Amanita muscaria var. gemmata (1886, Lucien Quélet), Amanitopsis gemmata (1887, Pier Andrea Saccardo), Amanitaria gemmata (1940, Jean-Edouard Gilbert), and Venenarius gemmatus (1948, William Murrill). Amanita authority Rodham E. Tulloss considers A. amici (published by Claude
|
https://en.wikipedia.org/wiki/Seafood%20boil
|
Seafood boil is the generic term for any number of types of social events in which shellfish, whether saltwater or freshwater, is the central element. Regional variations dictate the kinds of seafood, the accompaniments and side dishes, and the preparation techniques (boiling, steaming, baking, or raw). In some cases, a boil may be sponsored by a community organization as a fund-raiser or a mixer. In this way, seafood boils are like a fish fry, barbecue, or church potluck supper. Boils are also held by individuals for their friends and family for a weekend get-together and on the holidays of Memorial Day and Independence Day. While boils and bakes are traditionally associated with coastal regions of the United States, there are exceptions.
Louisiana
Shrimp, crab, and crawfish boils are a Louisiana Cajun tradition and can be found across Louisiana and can even now be found along the Gulf South. But it is the more popular crawfish boil that is most closely associated with Louisiana. The Breaux Bridge Crawfish Festival in Louisiana has been named one of the top 10 food events by USA Today and is a showcase for Cajun music and culture. Major crawfish boils are held by churches and other organizations as fundraisers throughout the spring. Tulane University holds an annual "Crawfest" in April, and the University of New Orleans holds an annual crawfish boil for all students at the end of the spring semester (Students unwinding on Crawfish and Unprecedented Fun—SUCAUF). Smaller events can be found in backyards and parks throughout April, May, and June. Locals traditionally eat crawfish, as well as crabs, without tools such as shell crackers or picks.
One reason for the popularity of crawfish may be price. During the height of the season (late spring) the price may be less than a $1.50/pound retail for live crawfish (2006) with crawfish prices currently being around $.99/pound. Shrimp and crab are higher valued cash crops, and can be a less affordable option for larger gr
|
https://en.wikipedia.org/wiki/Denis%20Noble
|
Denis Noble (born 16 November 1936) is a British physiologist and biologist who held the Burdon Sanderson Chair of Cardiovascular Physiology at the University of Oxford from 1984 to 2004 and was appointed Professor Emeritus and co-Director of Computational Physiology. He is one of the pioneers of systems biology and developed the first viable mathematical model of the working heart in 1960.
Education
Noble was educated at Emanuel School and University College London (UCL). In 1958 he began his investigations into the mechanisms of heartbeat. This led to two seminal papers in Nature in 1960 giving the first experimentally-based mathematical simulation of the electrical rhythm of the heart, extensively developed with Richard Tsien in 1975, and with Dario DiFrancesco in 1985. All three articles form the foundations of modern electrophysiology of the heart. The 1985 article was included in 2015 in the Royal Society's 350 year celebration of the publication of Philosophical Transactions.
From this work it became clear that there was not a single oscillator which controlled heartbeat, but rather this was an emergent property of the feedback loops involving the various ion channels. In 1961 he obtained his PhD working under the supervision of Otto Hutter at UCL.
Research
Noble's research focuses on using computer models of biological organs and organ systems to interpret function from the molecular level to the whole organism. Together with international collaborators, his team has used supercomputers to create the first virtual organ, the virtual heart.
As secretary-general of the International Union of Physiological Sciences 1993–2001, he played a major role in launching the Physiome Project, an international project to use computer simulations to create the quantitative physiological models necessary to interpret the genome, and he was elected president of the IUPS at its world congress in Kyoto in 2009.
Noble is also a philosopher of biology, with many publicati
|
https://en.wikipedia.org/wiki/Holonomic%20function
|
In mathematics, and more specifically in analysis, a holonomic function is a smooth function of several variables that is a solution of a system of linear homogeneous differential equations with polynomial coefficients and satisfies a suitable dimension condition in terms of D-modules theory. More precisely, a holonomic function is an element of a holonomic module of smooth functions. Holonomic functions can also be described as differentiably finite functions, also known as D-finite functions. When a power series in the variables is the Taylor expansion of a holonomic function, the sequence of its coefficients, in one or several indices, is also called holonomic. Holonomic sequences are also called P-recursive sequences: they are defined recursively by multivariate recurrences satisfied by the whole sequence and by suitable specializations of it. The situation simplifies in the univariate case: any univariate sequence that satisfies a linear homogeneous recurrence relation with polynomial coefficients, or equivalently a linear homogeneous difference equation with polynomial coefficients, is holonomic.
Holonomic functions and sequences in one variable
Definitions
Let be a field of characteristic 0 (for example, or ).
A function is called D-finite (or holonomic) if there exist polynomials such that
holds for all x. This can also be written as where
and is the differential operator that maps to . is called an annihilating operator of f (the annihilating operators of form an ideal in the ring , called the annihilator of ). The quantity r is called the order of the annihilating operator. By extension, the holonomic function f is said to be of order r when an annihilating operator of such order exists.
A sequence is called P-recursive (or holonomic) if there exist polynomials such that
holds for all n. This can also be written as where
and the shift operator that maps to . is called an annihilating operator of c (the annihilating operators of
|
https://en.wikipedia.org/wiki/Angular%20eccentricity
|
Angular eccentricity is one of many parameters which arise in the study of the ellipse or ellipsoid. It is denoted here by α (alpha). It may be defined in terms of the eccentricity, e, or the aspect ratio, b/a (the ratio of the semi-minor axis and the semi-major axis):
Angular eccentricity is not currently used in English language publications on mathematics, geodesy or map projections but it does appear in older literature.
Any non-dimensional parameter of the ellipse may be expressed in terms of the angular eccentricity. Such expressions are listed in the following table after the conventional definitions. in terms of the semi-axes. The notation for these parameters varies. Here we follow Rapp:
{| class="wikitable" style="border: 1px solid darkgray" cellpadding="5"
| (first) eccentricity
| style="padding-left: 0.5em"|
| style="padding-left: 1.5em"|
| style="padding-left: 1.5em"|
|
|-
| second eccentricity
| style="padding-left: 0.5em"|
| style="padding-left: 1.5em"|
| style="padding-left: 1.5em"|
|
|-
| third eccentricity
| style="padding-left: 0.5em"|
| style="padding-left: 1.5em"|
| style="padding-left: 1.5em"|
|
|-
| style="padding-left: 0.5em"| (first) flattening
| style="padding-left: 0.5em"|
| style="padding-left: 1.5em"|
| style="padding-left: 1.5em"|
|
|-
| style="padding-left: 0.5em"|second flattening
| style="padding-left: 0.5em"|
| style="padding-left: 1.5em"|
| style="padding-left: 1.5em"|
|
|-
| style="padding-left: 0.5em"| third flattening
| style="padding-left: 0.5em"|
| style="padding-left: 1.5em"|
| style="padding-left: 1.5em"|
|
|}
The alternative expressions for the flattenings would guard against large cancellations in numerical work.
|
https://en.wikipedia.org/wiki/Crab%20boil
|
A crab boil is a social event where boiled crabs are eaten, a kind of seafood boil.
Louisiana and New Orleans
Boiled seafood in southern Louisiana tends to be spicier than that found in other parts of the country. Homemade crab boil recipes call for abundant amounts of hot sauce, cayenne pepper, salt, bay leaf, lemon, and garlic. Mustard seeds, coriander seeds, and allspice are popular extra options. Many people will start with a commercial crab boil product and then supplement it with extra pepper. The leading commercial product is Zatarain's which comes in two forms. One is a mesh bag with seasonings inside that will steep into the water. The second is a liquid concentrate that can be added directly to the water. The concentrate form can also be used as a flavor enhancer for soups. Other regional crab boil companies are Tony Chachere's, and Rex Crab Boil. Note that even when boiling shrimp or crawfish, most recipes call for adding crab boil packets as a seasoning.
Carolinas
The Lowcountry boil, Tidewater boil, and Frogmore Stew are variations on the same theme in North and South Carolina.
|
https://en.wikipedia.org/wiki/Actibind
|
Actibind is an actin-binding fungal T(2)-RNase protein that is produced by the black mold Aspergillus niger, a microorganism used in biotechnology and food technology. In plants, actibind binds actin, a major component of the cytoskeleton, interfering with the plants' pollen tubes and halting cell growth. Research published in the journal Cancer on 15 May 2006 reports evidence that actibind has antiangiogenic and anticarcinogenic characteristics. In human colon cancer, breast cancer and melanoma, increasing the level of actibind was found to reduce the ability of these cells to form tumorogenic colonies. In animal models, increased actibind inhibited the growth of colon cancer-derived tumors, metastases and blood vessel formation. During the completion of the Human Genome Project, the gene encoding for RNaseT2, the human actibind-like protein, was found on chromosome 6.
Why ACTIBIND?
The reason why ACTIBIND is an enzyme of interest in biochemical laboratories is due to the fact that researchers observed that ACTIBIND inhibits the elongation of pollen tubes by interfering with the intracellular actin network of the plant cell. The specific actin network ACTIBIND inhibits are actin rich pseudopods, which are important for a variety of cellular functions including elongation of plant pollen tubes, motility of mammalian cells, and most importantly cancer cell function. In cancer cells specifically, the actin rich pseudopods help the cancer's invasion and metastasis. Because ACTIBIND was able to interfere with actin networks in plant cells, researchers aimed to find out whether ACTIBIND could also inhibit mammalian cancer cell development.
ACTIBIND as an Antiangiogenic
ACTIBIND, an extracellular ribonuclease (T2-RNase), has been researched heavily in biochemical laboratories due the enzyme's ability to efficiently hydrolyze/degrade RNA molecules. The ability to cleave off RNA molecules it what makes ACTIBIND an effective antiangiogenic. Angiogenesis is the growth of
|
https://en.wikipedia.org/wiki/BMPR2
|
Bone morphogenetic protein receptor type II or BMPR2 is a serine/threonine receptor kinase encoded by the BMPR2 gene. It binds bone morphogenetic proteins, members of the TGF beta superfamily of ligands, which are involved in paracrine signaling. BMPs are involved in a host of cellular functions including osteogenesis, cell growth and cell differentiation. Signaling in the BMP pathway begins with the binding of a BMP to the type II receptor. This causes the recruitment of a BMP type I receptor, which the type II receptor phosphorylates. The type I receptor phosphorylates an R-SMAD, a transcriptional regulator.
Function
Unlike the TGFβ type II receptor, which has a high affinity for TGF-β1, BMPR2 does not have a high affinity for BMP-2, BMP-7 and BMP-4, unless it is co-expressed with a type I BMP receptor.
On ligand binding, a receptor complex is formed, consisting of two type II and two type I transmembrane
serine/threonine kinases. Type II receptors phosphorylate and activate type I receptors which autophosphorylate, then
bind and activate SMAD transcriptional regulators. They bind to BMP-7, BMP-2 and, less efficiently, BMP-4. Binding is weak but enhanced by the presence of type I receptors for BMPs. In TGF beta signaling all of the receptors exist in homodimers before ligand binding. In the case of BMP receptors only a small fraction of the receptors exist in homomeric forms before ligand binding. Once a ligand has bound to a receptor, the amount of homomeric receptor oligomers increase, suggesting that the equilibrium shifts towards the homodimeric form. The low affinity for ligands suggests that BMPR2 may differ from other type II TGF beta receptors in that the ligand may bind the type I receptor first.
Oocyte Development
BMPR2 is expressed on both human and animal granulosa cells, and is a crucial receptor for bone morphogenetic protein 15 (BMP15) and growth differentiation factor 9 (GDF 9). These two protein signaling molecules and their BMPR2-mediated
|
https://en.wikipedia.org/wiki/J.%20Anthony%20Hall
|
J. Anthony Hall FREng is a leading British software engineer specializing in the use of formal methods, especially the Z notation.
Anthony Hall was educated at the University of Oxford with a BA in chemistry and a DPhil in theoretical chemistry. His subsequent posts have included:
ICI Research Fellow, Department of Theoretical Chemistry, University of Sheffield (1971–1973)
Principal Scientific Officer, British Museum Research Laboratory (1973–1980)
Senior Consultant, Systems Programming Limited (1980–1984)
Principal Consultant, Systems Designers (1984–1986)
Visiting Professor, Carnegie Mellon University (1994)
Principal Consultant, Praxis Critical Systems (1986–2004)
In particular, Hall has worked on software development using formal methods for the UK National Air Traffic Services (NATS). He has been an invited speaker at conferences concerned with formal methods, requirements engineering and software engineering.
Since 2004, Hall has been an independent consultant. He has also been a visiting professor at the University of York. Hall was the founding chair of ForTIA, the Formal Techniques Industry Association.
Selected publications
Anthony Hall, Seven Myths of Formal Methods, IEEE Software, September 1990, pp. 11–19.
Anthony Hall and Roderick Chapman, Correctness by Construction: Developing a Commercial Secure System, IEEE Software, January/February 2002, pp. 18–25.
|
https://en.wikipedia.org/wiki/Root%20rot
|
Root rot is a condition in which anoxic conditions in the soil or potting media around the roots of a plant cause them to rot. This occurs due to excessive standing water around the roots. It is found in both indoor and outdoor plants, although it is more common in indoor plants due to overwatering, heavy potting media, or containers with poor drainage. The leaves of plants experiencing root rot often yellow and die, and if allowed to continue, the condition can be fatal.
To avoid root rot, it is best to only water plants when the soil becomes dry, and to put the plant in a well-drained pot. Using a dense potting media such as one dug up from outdoors can also cause root rot. Plants from different environments have different tolerances for soil moisture: plants evolved for desert conditions will experience root rot at lower moisture levels than plants evolved for tropical conditions. In both indoor and outdoor plants, it can be lethal and there is no effective treatment, though some plants can be propagated so they will not be lost completely.
Many cases of root rot are caused by members of the water mold genus Phytophthora; perhaps the most aggressive is P. cinnamomi. Spores from root rot causing agents do contaminate other plants, but the rot cannot take hold unless there is adequate moisture. Spores are not only airborne, but are also carried by insects and other arthropods in the soil.
It can be controlled by drenching carbendazim.
Hydroponics
Root rot can occur in hydroponic applications, if the water is not properly aerated. This is usually accomplished by use of an air pump, air stones, air diffusers and by adjustment of the frequency and length of watering cycles where applicable. Hydroponic air pumps function in much the same way as aquarium pumps, which are used for the same purpose. Root rot and other problems associated with poor water aeration were principal reasons for the development of aeroponics.
Particular diseases
Some particular pathogens in
|
https://en.wikipedia.org/wiki/Phenoptosis
|
Phenoptosis (from pheno: showing or demonstrating; ptosis: programmed death, "falling off") is a conception of the self-programmed death of an organism proposed by Vladimir Skulachev in 1999.
In many species, including salmon and marsupial mice, under certain circumstances, especially following reproduction, an organism's genes will cause the organism to rapidly degenerate and die off. Recently this has been referred to as "fast phenoptosis" as aging is being explored as "slow phenoptosis". Phenoptosis is a common feature of living species, whose ramifications for humans is still being explored. The concept of programmed cell death was used before, by Lockshin & Williams in 1964 in relation to insect tissue development, around eight years before "apoptosis" was coined. The term 'phenoptosis' is a neologism associated with Skulachev's proposal.
Evolutionary significance
In multicellular organisms, worn-out and ineffective cells are dismantled and recycled for the greater good of the whole organism in a process called apoptosis. It is believed that phenoptosis is an evolutionary mechanism that culls out the damaged, aged, infectious, or those in direct competition with their own offspring for the good of the species. Special circumstances need to exist for the "phenoptosis" strategy to be an evolutionarily stable strategy (ESS), let alone the only ESS. Examples of "phenoptosis" given below are really examples of semelpary - a life history with a single reproduction followed by death, which evolves not "for the good of the species" but as the ESS in the conditions of high adult-to-juvenile mortality ratio. The elimination of parts detrimental to the organism or individuals detrimental to the species has been deemed "The samurai law of biology" – it is better to die than to be wrong.
Stress-induced, acute, or fast phenoptosis is the rapid deterioration of an organism induced by a life event such as breeding. Elimination of the parent provides space for fitter offspri
|
https://en.wikipedia.org/wiki/Clitocybe%20odora
|
Clitocybe odora, commonly known as the blue green anise mushroom, or aniseed toadstool, is a blue-green mushroom that grows near deciduous and coniferous trees. They can be found growing in small groups along the side of tree roots. This mushroom is edible, but a few expert mushroom hunters insist that young specimens should be avoided as they can be confused with Stropharia aeruginosa. The anise odor is due to the presence of p-anisaldehyde and a small amount of benzaldehyde. This odor can give away the mushroom's presence before it is observed by eye.
Taxonomy
First described by the French mycologist Jean Baptiste Francois Pierre Bulliard (1742–1793). The specific epithet odora is from the Latin meaning "perfumed".
Description
Young specimens have a light blue texture on the cap which fades to grey in age. The gills and stem are white with no ring.
Full grown specimens have blue-green, flowery, cup-shaped caps; the gills are creamy white, or reflect the blue-green color of the cap. The cap's surface feels rough. The stem is thick, is attached to the gills with no rings, and is textured, with a pale-yellow colour. The younger ones have a bell-shaped cap with a light blue or icy blue colour. The gills and stem are white, or bluish green. It has a strong scent and taste of aniseed, hence its name.
There is a white variety (Clitocybe odora var.alba Lange) that has the same strong odour.
Distribution and habitat
Found in both deciduous, and coniferous woods, it is widespread in the temperate zones, occurring in Asia, Europe, and North America. On the East Coast of North America it favours oak woodland, but it is often abundant in the coniferous forests of the Pacific Northwest.
Edibility
The caps can be dried, and used as a condiment, or used fresh for flavouring. Mushroom hunters should be sure to pick mature ones, mainly because the younger ones can be confused with several similar poisonous ones that grow along with this mushroom. Every part of the mushroom s
|
https://en.wikipedia.org/wiki/Pulay%20stress
|
The Pulay stress or Pulay forces (named for Peter Pulay) is an error that occurs in the stress tensor (or Jacobian matrix) obtained from self-consistent field calculations (Hartree–Fock or density functional theory) due to the incompleteness of the basis set.
A plane-wave density functional calculation on a crystal with specified lattice vectors will typically include in the basis set all plane waves with energies below the specified energy cutoff. This corresponds to all points on the reciprocal lattice that lie within a sphere whose radius is related to the energy cutoff. Consider what happens when the lattice vectors are varied, resulting in a change in the reciprocal lattice vectors. The points on the reciprocal lattice which represent the basis set will no longer correspond to a sphere, but an ellipsoid. This change in the basis set will result in errors in the calculated ground state energy change.
The Pulay stress is often nearly isotropic, and tends to result in an underestimate of the equilibrium volume. Pulay stress can be reduced by increasing the energy cutoff. Another way to mitigate the effect of Pulay stress on the equilibrium cell shape is to calculate the energy at different lattice vectors with a fixed energy cutoff.
Similarly, the error occurs in any calculation where the basis set explicitly depends on the position of atomic nuclei (which are to change during the geometry optimization). In this case, the Hellmann–Feynman theorem – which is used to avoid derivation of many-parameter wave function (expanded in a basis set) – is only valid for the complete basis set. Otherwise, the terms in theorem's expression containing derivatives of the wavefunction persist, giving rise to additional forces – the Pulay forces:
The presence of Pulay forces makes the optimized geometry parameters converge slower with increasing basis set. The way to eliminate the erroneous forces is to use nuclear-position-independent basis functions, to explicitly calculate
|
https://en.wikipedia.org/wiki/Martinotti%20cell
|
Martinotti cells are small multipolar neurons with short branching dendrites. They are scattered throughout various layers of the cerebral cortex, sending their axons up to the cortical layer I where they form axonal arborization. The arbors transgress multiple columns in layer VI and make contacts with the distal tuft dendrites of pyramidal cells.
Martinotti cells express somatostatin and sometimes calbindin, but not parvalbumin or vasoactive intestinal peptide. Furthermore, Martinotti cells in layer V have been shown to express the nicotinic acetylcholine receptor α2 subunit (Chrna2).
Martinotti cells are associated with a cortical dampening mechanism. When the pyramidal neuron, which is the most common type of neuron in the cortex, starts getting overexcited, Martinotti cells start sending inhibitory signals to the surrounding neurons.
Historically, the discovery of Martinotti cells has been mistakenly attributed to Giovanni Martinotti 1888, although it is now accepted that they were actually discovered in 1889 by Carlo Martinotti (1859–1908), a student of Camillo Golgi.
External links
News, press releases
Rare cell prevents rampant brain activity - on the discovery of potential dampening influence of Martinotti cells.
NIF Search - Martinotti Cell via the Neuroscience Information Framework
See also
List of distinct cell types in the adult human body
|
https://en.wikipedia.org/wiki/Strobilus
|
A strobilus (: strobili) is a structure present on many land plant species consisting of sporangia-bearing structures densely aggregated along a stem. Strobili are often called cones, but some botanists restrict the use of the term cone to the woody seed strobili of conifers. Strobili are characterized by a central axis (anatomically a stem) surrounded by spirally arranged or decussate structures that may be modified leaves or modified stems.
Leaves that bear sporangia are called sporophylls, while sporangia-bearing stems are called sporangiophores.
Lycophytes
Some members of both of the two modern classes of Lycopodiophyta (Lycopodiopsida and Isoetopsida) produce strobili. In all cases, the lateral organs of the strobilus are microphylls, bearing sporangia. In other lycophytes, ordinary foliage leaves can act as sporophylls, and there are no organized strobili.
Sphenophytes
The single extant genus of Equisetophyta, Equisetum, produces strobili in which the lateral organs are sporangiophores. Developmental evidence and comparison with fossil members of the group show that the sporangiophores are reduced stems, rather than leaves. Sporangia are terminal.
Seed plants
With the exception of flowering plants, seed plants produce ovules and pollen in different structures. Strobili bearing microsporangia are called microsporangiate strobili or pollen cones, and those bearing ovules are megasporangiate strobili or seed cones (or ovulate cones).
Cycads
Cycadophyta are typically dioecious (seed strobili and pollen strobili are produced on separate plants). The lateral organs of seed strobili are megasporophylls (modified leaves) that bear two to several marginal ovules. Pollen strobili consist of microsporophylls, each of which may have dozens or hundreds of abaxial microsporangia.
Ginkgos
The single living member of the Ginkgophyta, Ginkgo biloba produces pollen strobili, but the ovules are typically borne in pairs at the end of a stem, not in a strobilus. When there a
|
https://en.wikipedia.org/wiki/Shadows%20of%20the%20Mind
|
Shadows of the Mind: A Search for the Missing Science of Consciousness is a 1994 book by mathematical physicist Roger Penrose that serves as a followup to his 1989 book The Emperor's New Mind: Concerning Computers, Minds and The Laws of Physics.
Penrose hypothesizes that:
Human consciousness is non-algorithmic, and thus is not capable of being modelled by a conventional Turing machine type of digital computer.
Quantum mechanics plays an essential role in the understanding of human consciousness; specifically, he believes that microtubules within neurons support quantum superpositions.
The objective collapse of the quantum wavefunction of the microtubules is critical for consciousness.
The collapse in question is physical behaviour that is non-algorithmic and transcends the limits of computability.
The human mind has abilities that no Turing machine could possess because of this mechanism of non-computable physics.
Argument
Mathematical thought
In 1931, the mathematician and logician Kurt Gödel proved his incompleteness theorems, showing that any effectively generated theory capable of expressing elementary arithmetic cannot be both consistent and complete. Further to that, for any consistent formal theory that proves certain basic arithmetic truths, there is an arithmetical statement that is true, but not provable in the theory. The essence of Penrose's argument is that while a formal proof system cannot, because of the theorem, prove its own incompleteness, Gödel-type results are provable by human mathematicians. He takes this disparity to mean that human mathematicians are not describable as formal proof systems and are not running an algorithm, so that the computational theory of mind is false, and computational approaches to artificial general intelligence are unfounded. (The argument was first given by Penrose in The Emperor's New Mind (1989) and is developed further in Shadows of The Mind. An earlier version of the argument was given by J. R. Lucas in 19
|
https://en.wikipedia.org/wiki/Basic%20helix-loop-helix%20leucine%20zipper%20transcription%20factors
|
Basic helix-loop-helix leucine zipper transcription factors are, as their name indicates, transcription factors containing both Basic helix-loop-helix and leucine zipper motifs.
Examples include Microphthalmia-associated transcription factor and Sterol regulatory element binding protein (SREBP).
External links
Gene expression
Transcription factors
|
https://en.wikipedia.org/wiki/Ankyrin%20repeat
|
The ankyrin repeat is a 33-residue motif in proteins consisting of two alpha helices separated by loops, first discovered in signaling proteins in yeast Cdc10 and Drosophila Notch. Domains consisting of ankyrin tandem repeats mediate protein–protein interactions and are among the most common structural motifs in known proteins. They appear in bacterial, archaeal, and eukaryotic proteins, but are far more common in eukaryotes. Ankyrin repeat proteins, though absent in most viruses, are common among poxviruses. Most proteins that contain the motif have four to six repeats, although its namesake ankyrin contains 24, and the largest known number of repeats is 34, predicted in a protein expressed by Giardia lamblia.
Ankyrin repeats typically fold together to form a single, linear solenoid structure called ankyrin repeat domains. These domains are one of the most common protein–protein interaction platforms in nature. They occur in a large number of functionally diverse proteins, mainly from eukaryotes. The few known examples from prokaryotes and viruses may be the result of horizontal gene transfers. The repeat has been found in proteins of diverse function such as transcriptional initiators, cell cycle regulators, cytoskeletal, ion transporters, and signal transducers. The ankyrin fold appears to be defined by its structure rather than its function, since there is no specific sequence or structure that is universally recognised by it.
Considering the atomic structures of individual ankyrin repeats, the loop is often a type 1 beta bulge loop, while both alpha-helices commonly have a Schellman loop at their N-terminus.
Role in protein folding
The ankyrin-repeat sequence motif has been studied using multiple sequence alignment to determine conserved amino acid residues critical for folding and stability. The residues on the wide lateral surface of ankyrin repeat structures are variable, often hydrophobic, and involved mainly in mediating protein–protein interactions. A
|
https://en.wikipedia.org/wiki/Gas/oil%20ratio
|
When oil is produced to surface temperature and pressure it is usual for some natural gas to come out of solution. The gas/oil ratio (GOR) is the ratio of the volume of gas ("scf") that comes out of solution to the volume of oil — at standard conditions.
In reservoir simulation gas/oil ratio is usually abbreviated .
A point to check is whether the volume of oil is measured before or after the gas comes out of solution, since the remaining oil volume will decrease when the gas comes out.
In fact, gas dissolution and oil volume shrinkage will happen at many stages during the path of the hydrocarbon stream from reservoir through the wellbore and processing plant to export. For light oils and rich gas condensates the ultimate GOR of export streams is strongly influenced by the efficiency with which the processing plant strips liquids from the gas phase. Reported GORs may be calculated from export volumes, which may not be at standard conditions.
The GOR is a dimensionless ratio (volume per volume) in metric units, but in field units, it is usually measured in cubic feet of gas per barrel of oil or condensate.
In the states of Texas and Pennsylvania, the statutory definition of a gas well is one where the GOR is greater than 100,000 ft3/bbl or 100 Kcf/bbl.
The state of New Mexico also designates a gas well as having over 100 MCFG per barrel.
The Oklahoma Geological Survey in 2015 published a map that displays gas wells with greater than 20 MCFG per barrel of oil. They go on to display oil wells with GOR of less than 5 MCFG/BBL and oil and gas wells between these limits.
The EPA's 2016 Information Collection Request for Oil and Gas Facilities (EPA ICR No. 2548.01, OMB Control No. 2060-NEW) divided well types into five categories:
1. Heavy Oil (GOR ≤ 300 scf/bbl)
2. Light Oil (GOR 300 < GOR ≤ 100,000 scf/bbl)
3. Wet Gas (100,000 < GOR ≤1,000,000 scf/bbl)
4. Dry Gas (GOR > 1,000,000 scf/bbl)
5. Coal Bed Methane.
|
https://en.wikipedia.org/wiki/EnterpriseDB
|
EnterpriseDB (EDB), a privately held company based in Massachusetts, provides software and services based on the open-source database PostgreSQL (also known as Postgres), and is one of the largest contributors to Postgres. EDB develops and integrates performance, security, and manageability enhancements into Postgres to support enterprise-class workloads. EDB has also developed database compatibility for Oracle to facilitate the migration of workloads from Oracle to EDB Postgres and to support the operation of many Oracle workloads on EDB Postgres.
EDB provides a portfolio of databases and tools that extend Postgres for enterprise workloads. This includes fully managed Postgres in the cloud, extreme high availability for Postgres, command line migration tools, Kubernetes Operator and container images, management, monitoring and optimizing of Postgres, enterprise ready Oracle migration tools and browser-based schema migration tools
EnterpriseDB was purchased by Great Hill Partners in 2019.
In June 2022, Bain Capital Private Equity announced a majority growth investment in the company, whereafter EDB continues to operate under the leadership of Ed Boyajian, President and CEO of EDB, an open source pioneer who has led the company since 2008.
Great Hill Partners, which acquired EDB in 2019, remains a significant shareholder.
History
EDB was founded in 2004. The growing acceptance of open source software created a market opportunity and the company wanted to challenge the database incumbents with a standards based product that was compatible with other vendor solutions. EnterpriseDB sought to develop an open source-based, enterprise-class relational database to compete with established vendors at an open source price point.
EDB introduced its database, EnterpriseDB 2005, in 2005. It was named Best Database Solution at LinuxWorld that year, beating solutions from Oracle, MySQL, and IBM. EDB renamed the database EnterpriseDB Advanced Server with its March 2006 rele
|
https://en.wikipedia.org/wiki/Self-healing%20material
|
Self-healing materials are artificial or synthetically created substances that have the built-in ability to automatically repair damages to themselves without any external diagnosis of the problem or human intervention. Generally, materials will degrade over time due to fatigue, environmental conditions, or damage incurred during operation. Cracks and other types of damage on a microscopic level have been shown to change thermal, electrical, and acoustical properties of materials, and the propagation of cracks can lead to eventual failure of the material. In general, cracks are hard to detect at an early stage, and manual intervention is required for periodic inspections and repairs. In contrast, self-healing materials counter degradation through the initiation of a repair mechanism that responds to the micro-damage. Some self-healing materials are classed as smart structures, and can adapt to various environmental conditions according to their sensing and actuation properties.
Although the most common types of self-healing materials are polymers or elastomers, self-healing covers all classes of materials, including metals, ceramics, and cementitious materials. Healing mechanisms vary from an instrinsic repair of the material to the addition of a repair agent contained in a microscopic vessel. For a material to be strictly defined as autonomously self-healing, it is necessary that the healing process occurs without human intervention. Self-healing polymers may, however, activate in response to an external stimulus (light, temperature change, etc.) to initiate the healing processes.
A material that can intrinsically correct damage caused by normal usage could prevent costs incurred by material failure and lower costs of a number of different industrial processes through longer part lifetime, and reduction of inefficiency caused by degradation over time.
History
The ancient Romans used a form of lime mortar that has been found to have self-healing properties. By 2
|
https://en.wikipedia.org/wiki/Bounded%20quantifier
|
In the study of formal theories in mathematical logic, bounded quantifiers (a.k.a. restricted quantifiers) are often included in a formal language in addition to the standard quantifiers "∀" and "∃". Bounded quantifiers differ from "∀" and "∃" in that bounded quantifiers restrict the range of the quantified variable. The study of bounded quantifiers is motivated by the fact that determining whether a sentence with only bounded quantifiers is true is often not as difficult as determining whether an arbitrary sentence is true.
Examples
Examples of bounded quantifiers in the context of real analysis include:
- for all x where x is larger than 0
- there exists a y where y is less than 0
- for all x where x is a real number
- every positive number is the square of a negative number
Bounded quantifiers in arithmetic
Suppose that L is the language of Peano arithmetic (the language of second-order arithmetic or arithmetic in all finite types would work as well). There are two types of bounded quantifiers: and .
These quantifiers bind the number variable n and contain a numeric term t which may not mention n but which may have other free variables. ("Numeric terms" here means terms such as "1 + 1", "2", "2 × 3", "m + 3", etc.)
These quantifiers are defined by the following rules ( denotes formulas):
There are several motivations for these quantifiers.
In applications of the language to recursion theory, such as the arithmetical hierarchy, bounded quantifiers add no complexity. If is a decidable predicate then and are decidable as well.
In applications to the study of Peano arithmetic, the fact that a particular set can be defined with only bounded quantifiers can have consequences for the computability of the set. For example, there is a definition of primality using only bounded quantifiers: a number n is prime if and only if there are not two numbers strictly less than n whose product is n. There is no quantifier-free definition of primality in the
|
https://en.wikipedia.org/wiki/James%20Earl%20Baumgartner
|
James Earl Baumgartner (March 23, 1943 – December 28, 2011) was an American mathematician who worked in set theory, mathematical logic and foundations, and topology.
Baumgartner was born in Wichita, Kansas, began his undergraduate study at the California Institute of Technology in 1960, then transferred to the University of California, Berkeley, from which he received his PhD in 1970 from for a dissertation titled Results and Independence Proofs in Combinatorial Set Theory. His advisor was Robert Vaught. He became a professor at Dartmouth College in 1969, and spent his entire career there.
One of Baumgartner's results is the consistency of the statement that any two -dense sets of reals are order isomorphic (a set of reals is -dense if it has exactly points in every open interval). With András Hajnal he proved the Baumgartner–Hajnal theorem, which states that the partition relation holds for and . He died in 2011 of a heart attack at his home in Hanover, New Hampshire.
The mathematical context in which Baumgartner worked spans Suslin's problem, Ramsey theory, uncountable order types, disjoint refinements, almost disjoint families, cardinal arithmetics, filters, ideals, and partition relations, iterated forcing and Axiom A, proper forcing and the proper forcing axiom, chromatic number of graphs, a thin very-tall superatomic Boolean algebra, closed unbounded sets, and partition relations.
See also
Baumgartner's axiom
Selected publications
Baumgartner, James E., A new class of order types, Annals of Mathematical Logic, 9:187–222, 1976
Baumgartner, James E., Ineffability properties of cardinals I, Infinite and Finite Sets, Keszthely (Hungary) 1973, volume 10 of Colloquia Mathematica Societatis János Bolyai, pages 109–130. North-Holland, 1975
Baumgartner, James E.; Harrington, Leo; Kleinberg, Eugene, Adding a closed unbounded set, Journal of Symbolic Logic, 41(2):481–482, 1976
Baumgartner, James E., Ineffability properties of cardinals II, Robert E. Butts
|
https://en.wikipedia.org/wiki/Robbins%20algebra
|
In abstract algebra, a Robbins algebra is an algebra containing a single binary operation, usually denoted by , and a single unary operation usually denoted by satisfying the following axioms:
For all elements a, b, and c:
Associativity:
Commutativity:
Robbins equation:
For many years, it was conjectured, but unproven, that all Robbins algebras are Boolean algebras. This was proved in 1996, so the term "Robbins algebra" is now simply a synonym for "Boolean algebra".
History
In 1933, Edward Huntington proposed a new set of axioms for Boolean algebras, consisting of (1) and (2) above, plus:
Huntington's equation:
From these axioms, Huntington derived the usual axioms of Boolean algebra.
Very soon thereafter, Herbert Robbins posed the Robbins conjecture, namely that the Huntington equation could be replaced with what came to be called the Robbins equation, and the result would still be Boolean algebra. would interpret Boolean join and Boolean complement. Boolean meet and the constants 0 and 1 are easily defined from the Robbins algebra primitives. Pending verification of the conjecture, the system of Robbins was called "Robbins algebra."
Verifying the Robbins conjecture required proving Huntington's equation, or some other axiomatization of a Boolean algebra, as theorems of a Robbins algebra. Huntington, Robbins, Alfred Tarski, and others worked on the problem, but failed to find a proof or counterexample.
William McCune proved the conjecture in 1996, using the automated theorem prover EQP. For a complete proof of the Robbins conjecture in one consistent notation and following McCune closely, see Mann (2003). Dahn (1998) simplified McCune's machine proof.
See also
Algebraic structure
Minimal axioms for Boolean algebra
|
https://en.wikipedia.org/wiki/Collostructional%20analysis
|
Collostructional analysis is a family of methods developed by (in alphabetical order) Stefan Th. Gries (University of California, Santa Barbara) and Anatol Stefanowitsch (Free University of Berlin). Collostructional analysis aims at measuring the degree of attraction or repulsion that words exhibit to constructions, where the notion of construction has so far been that of Goldberg's construction grammar.
Collostructional methods
Collostructional analysis so far comprises three different methods:
collexeme analysis, to measure the degree of attraction/repulsion of a lemma to a slot in one particular construction;
distinctive collexeme analysis, to measure the preference of a lemma to one particular construction over another, functionally similar construction; multiple distinctive collexeme analysis extends this approach to more than two alternative constructions;
covarying collexeme analysis, to measure the degree of attraction of lemmas in one slot of a construction to lemmas in another slot of the same construction.
Input frequencies
Collostructional analysis requires frequencies of words and constructions and is similar to a wide variety of collocation statistics. It differs from raw frequency counts by providing not only observed co-occurrence frequencies of words and constructions, but also
(i) a comparison of the observed frequency to the one expected by chance; thus, collostructional analysis can distinguish attraction and repulsion of words and constructions;
(ii) a measure of the strength of the attraction or repulsion; this is usually the log-transformed p-value of a Fisher-Yates exact test.
Versus other collocation statistics
Collostructional analysis differs from most collocation statistics such that
(i) it measures not the association of words to words, but of words to syntactic patterns or constructions; thus, it takes syntactic structure more seriously than most collocation-based analyses;
(ii) it has so far only used the most precise statis
|
https://en.wikipedia.org/wiki/Somatomedin
|
Somatomedins are a group of proteins produced predominantly by the liver when growth hormones act on target tissue. Somatomedins inhibit the release of growth hormones by acting directly on anterior pituitary and by stimulating the secretion of somatostatin from the hypothalamus.
Somatomedins are a group of proteins that promote cell growth and division in response to stimulation by growth hormone (GH), also known as somatotropin (STH).
Somatomedins have similar biological effects to somatotropin.
In addition to their actions that stimulate growth, somatomedins also stimulate production of somatostatin, which suppresses growth hormone release. Thus, levels of somatomedins are controlled via negative feedback through the intermediates of somatostatin and growth hormone. Somatomedins are produced in many tissues and have autocrine and paracrine actions in addition to their endocrine action. The liver is thought to be the predominant source of circulating somatomedins.
Three forms include:
Somatomedin A, which is another name for insulin-like growth factor 2 (IGF-2)
Somatomedin B, which is derived from vitronectin
Somatomedin C, which is another name for insulin-like growth factor 1 (IGF-1)
|
https://en.wikipedia.org/wiki/Malicious%20Software%20Removal%20Tool
|
Microsoft Windows Malicious Software Removal Tool (MSRT) is a freeware second-opinion malware scanner that Microsoft's Windows Update downloads and runs on Windows computers each month, independent of the install antivirus software. First released on January 13, 2005, MSRT does not offer real-time protection. It scans its host computer for specific, widespread malware, and tries to eliminate the infection. Outside its monthly deployment schedule, it can be separately downloaded from Microsoft.
Availability
Since its January 13, 2005, Microsoft releases the updated tool every second Tuesday of every month (commonly called "Patch Tuesday") through Windows Update, at which point it runs once automatically in the background and reports if malicious software is found. The tool is also available as a standalone download.
Since support for Windows 2000 ended on July 13, 2010, Microsoft stopped distributing the tool to Windows 2000 users via Windows Update. The last version of the tool that could run on Windows 2000 was 4.20, released on May 14, 2013. Starting with version 5.1, released on June 11, 2013, support for Windows 2000 was dropped altogether. Although Windows XP support ended on April 8, 2014, updates for the Windows XP version of the Malicious Software Removal Tool would be provided until August, 2016; version 5.39. The latest version of MSRT for Windows Vista is 5.47, released on 11 April 2017.
Despite Microsoft ending general support for the Windows 7 operating system in 2020, updates are still provided to Windows 7 users via the standard Windows Update delivery mechanism.
Operation
MSRT does not install a shortcut in the Start menu. Hence, users must manually execute %windir%\system32\mrt.exe. The tool records its results in a log file located at %windir%\debug\mrt.log.
The tool reports anonymized data about any detected infections to Microsoft. MSRT's EULA discloses this reporting behavior and explains how to disable it.
Impact
In a June 2006 Micro
|
https://en.wikipedia.org/wiki/Alexander%20Ostrowski
|
Alexander Markowich Ostrowski (; ; 25 September 1893, in Kiev, Russian Empire – 20 November 1986, in Montagnola, Lugano, Switzerland) was a mathematician.
His father Mark having been a merchant, Alexander Ostrowski attended the Kiev College of Commerce, not a high school, and thus had an insufficient qualification to be admitted to university. However, his talent did not remain undetected: Ostrowski's mentor, Dmitry Grave, wrote to Landau and Hensel for help.
Subsequently, Ostrowski began to study mathematics at Marburg University under Hensel's supervision in 1912. During World War I he was interned, but thanks to the intervention of Hensel, the restrictions on his movements were eased somewhat, and he was allowed to use the university library.
After the war ended Ostrowski moved to Göttingen where he wrote his doctoral dissertation and was influenced by Hilbert, Klein and Landau. In 1920, after having obtained his doctorate, Ostrowski moved to Hamburg where he worked as Hecke's assistant and finished his habilitation in 1922. In 1923 he returned to Göttingen, and in 1928 became Professor of Mathematics at Basel, until retirement in 1958. In 1950 Ostrowski obtained Swiss citizenship. After retirement he still published scientific papers until his late eighties.
Selected publications
Vorlesungen über Differential- und Integralrechnung, 3 vols., Birkhäuser; vol. 1, 1945; vol. 1, 2nd edition, 1960; vol. 2, 1951; vol. 3, 1954;
Solution of equations and systems of equations. Academic Press, New York 1960; 2nd edition 1965; 2016 pbk reprint of 2nd edition
Aufgabensammlung zur Infinitesimalrechnung. several vols., Birkhäuser, Basel (1st edition 1964; 2nd edition 1972) pbk reprint vol. 1; vol. 2 A; vol. 2 B; vol. 3
Collected mathematical papers. 6 vols., Birkhäuser, Basel 1983–1984. vol. 1; vol. 2; vol. 3; vol. 4; vol. 5; vol. 6
See also
Ostrowski's theorem
Ostrowski–Hadamard gap theorem
Ostrowski numeration
Ostrowski Prize
|
https://en.wikipedia.org/wiki/Remission%20%28medicine%29
|
Remission is either the reduction or disappearance of the signs and symptoms of a disease. The term may also be used to refer to the period during which this reduction occurs. A remission may be considered a partial remission or a complete remission. Each disease, type of disorder, or clinical trial can have its own definition of a partial remission. For example, a partial remission for cancer may be defined as a 50% or greater reduction in the measurable parameters of tumor growth as may be found on physical examination, radiologic study, or by biomarker levels from a blood or urine test.
A complete remission, also called a full remission, is a total disappearance of the signs and symptoms of a disease. A person whose condition is in complete remission may be considered cured or recovered. Relapse is a term to describe returning symptoms of the disease after a period of remission. In cancer-treatment, doctors usually avoid the term "cured" and instead prefer the term "no evidence of disease" (NED) to refer to a complete remission of cancer, which does not rule out the possibility of relapse.
In mental disorders, there is generally no distinction between partial remission and complete remission. For example, a person diagnosed with a personality disorder must initially fit a set or subset of criteria from a predefined list, and remission in this context is defined as no longer meeting the criteria required for diagnosis. In this case it is still possible for the person to be demonstrating some symptoms, but they are at a subclinical severity or frequency that does not merit re-diagnosis.
For some diseases featuring remission, especially for those with no known cure such as multiple sclerosis, remission is implied to always be partial.
See also
Relapsing-remitting
Spontaneous remission
|
https://en.wikipedia.org/wiki/Diffuse%20reflectance%20spectroscopy
|
Diffuse reflectance spectroscopy, or diffuse reflection spectroscopy, is a subset of absorption spectroscopy. It is sometimes called remission spectroscopy. Remission is the reflection or back-scattering of light by a material, while transmission is the passage of light through a material. The word remission implies a direction of scatter, independent of the scattering process. Remission includes both specular and diffusely back-scattered light. The word reflection often implies a particular physical process, such as specular reflection.
The use of the term remission spectroscopy is relatively recent, and found first use in applications related to medicine and biochemistry. While the term is becoming more common in certain areas of absorption spectroscopy, the term diffuse reflectance is firmly entrenched, as in diffuse reflectance infrared Fourier transform spectroscopy (DRIFTS) and diffuse-reflectance ultraviolet–visible spectroscopy.
Mathematical treatments related to diffuse reflectance and transmittance
The mathematical treatments of absorption spectroscopy for scattering materials were originally largely borrowed from other fields. The most successful treatments use the concept of dividing a sample into layers, called plane parallel layers. They are generally those consistent with a two-flux or two-stream approximation. Some of the treatments require all the scattered light, both remitted and transmitted light, to be measured. Others apply only to remitted light, with the assumption that the sample is "infinitely thick" and transmits no light. These are special cases of the more general treatments.
There are several general treatments, all of which are compatible with each other, related to the mathematics of plane parallel layers. They are the Stokes formulas, equations of Benford, Hecht finite difference formula, and the Dahm equation. For the special case of infinitesimal layers, the Kubelka–Munk and Schuster–Kortüm treatments also give compat
|
https://en.wikipedia.org/wiki/Erythropoietin%20receptor
|
The erythropoietin receptor (EpoR) is a protein that in humans is encoded by the EPOR gene. EpoR is a 52kDa peptide with a single carbohydrate chain resulting in an approximately 56-57 kDa protein found on the surface of EPO responding cells. It is a member of the cytokine receptor family. EpoR pre-exists as dimers. These dimers were originally thought to be formed by extracellular domain interactions, however, it is now assumed that it is formed by interactions of the transmembrane domain and that the original structure of the extracellular interaction site was due to crystallisation conditions and does not depict the native conformation. Binding of a 30 kDa ligand erythropoietin (Epo), changes the receptor's conformational change, resulting in the autophosphorylation of Jak2 kinases that are pre-associated with the receptor (i.e., EpoR does not possess intrinsic kinase activity and depends on Jak2 activity). At present, the best-established function of EpoR is to promote proliferation and rescue of erythroid (red blood cell) progenitors from apoptosis.
Function and mechanism of action
The cytoplasmic domains of the EpoR contain a number of phosphotyrosines that are phosphorylated by Jak2 and serve as docking sites for a variety of intracellular pathway activators and Stats (such as Stat5). In addition to activating Ras/AKT and ERK/MAP kinase, phosphatidylinositol 3-kinase/AKT pathway and STAT transcription factors, phosphotyrosines also serve as docking sites for phosphatases that negatively affect EpoR signaling in order to prevent overactivation that may lead to such disorders as erythrocytosis. In general, the defects in the erythropoietin receptor may produce erythroleukemia and familial erythrocytosis. Mutations in Jak2 kinases associated with EpoR can also lead to polycythemia vera.
Erythroid survival
Primary role of EpoR is to promote proliferation of erythroid progenitor cells and rescue erythroid progenitors from cell death. EpoR induced Jak2-Stat5
|
https://en.wikipedia.org/wiki/Somatomedin%20receptor
|
A somatomedin receptor is a receptor which binds the somatomedins (IGFs). Somatomedin is abbreviated to IGF, in reference to insulin-like growth factor.
There are two types:
Insulin-like growth factor 1 receptor (IGF-1R)
Insulin-like growth factor 2 receptor (IGF-2R)
External links
Receptors
|
https://en.wikipedia.org/wiki/List%20of%20gravitationally%20rounded%20objects%20of%20the%20Solar%20System
|
This is a list of most likely gravitationally rounded objects of the Solar System, which are objects that have a rounded, ellipsoidal shape due to their own gravity (but are not necessarily in hydrostatic equilibrium). Apart from the Sun itself, these objects qualify as planets according to common geophysical definitions of that term. The sizes of these objects range over three orders of magnitude in radius, from planetary-mass objects like dwarf planets and some moons to the planets and the Sun. This list does not include small Solar System bodies, but it does include a sample of possible planetary-mass objects whose shapes have yet to be determined. The Sun's orbital characteristics are listed in relation to the Galactic Center, while all other objects are listed in order of their distance from the Sun.
Star
The Sun is a G-type main-sequence star. It contains almost 99.9% of all the mass in the Solar System.
Planets
In 2006, the International Astronomical Union (IAU) defined a planet as a body in orbit around the Sun that was large enough to have achieved hydrostatic equilibrium and to have "cleared the neighbourhood around its orbit". The practical meaning of "cleared the neighborhood" is that a planet is comparatively massive enough for its gravitation to control the orbits of all objects in its vicinity. In practice, the term "hydrostatic equilibrium" is interpreted loosely. Mercury is round but not actually in hydrostatic equilibrium, but it is universally regarded as a planet nonetheless.
According to the IAU's explicit count, there are eight planets in the Solar System; four terrestrial planets (Mercury, Venus, Earth, and Mars) and four giant planets, which can be divided further into two gas giants (Jupiter and Saturn) and two ice giants (Uranus and Neptune). When excluding the Sun, the four giant planets account for more than 99% of the mass of the Solar System.
Dwarf planets
Dwarf planets are bodies orbiting the Sun that are massive and warm eno
|
https://en.wikipedia.org/wiki/Galanin
|
Galanin is a neuropeptide encoded by the GAL gene, that is widely expressed in the brain, spinal cord, and gut of humans as well as other mammals. Galanin signaling occurs through three G protein-coupled receptors.
Much of galanin's functional role is still undiscovered. Galanin is closely involved in the modulation and inhibition of action potentials in neurons. Galanin has been implicated in many biologically diverse functions, including: nociception, waking and sleep regulation, cognition, feeding, regulation of mood, regulation of blood pressure, it also has roles in development as well as acting as a trophic factor. Galanin neurons in the medial preoptic area of the hypothalamus may govern parental behaviour. Galanin is linked to a number of diseases including Alzheimer's disease, epilepsy as well as depression, eating disorders, cancer, and addiction. Galanin appears to have neuroprotective activity as its biosynthesis is increased 2-10 fold upon axotomy in the peripheral nervous system as well as when seizure activity occurs in the brain. It may also promote neurogenesis.
Galanin is predominantly an inhibitory, hyperpolarizing neuropeptide and as such inhibits neurotransmitter release. Galanin is often co-localized with classical neurotransmitters such as acetylcholine, serotonin, and norepinephrine, and also with other neuromodulators such as neuropeptide Y, substance P, and vasoactive intestinal peptide.
Discovery
Galanin was first identified from porcine intestinal extracts in 1978 by Professor Viktor Mutt and colleagues at the Karolinska Institute, Sweden using a chemical assay technique that detects peptides according to its C-terminal alanine amide structure. Galanin is so-called because it contains an N-terminal glycine residue and a C-terminal alanine. The structure of galanin was determined in 1983 by the same team, and the cDNA of galanin was cloned from a rat anterior pituitary library in 1987.
Tissue distribution
Galanin is located predominant
|
https://en.wikipedia.org/wiki/Heegner%20point
|
In mathematics, a Heegner point is a point on a modular curve that is the image of a quadratic imaginary point of the upper half-plane. They were defined by Bryan Birch and named after Kurt Heegner, who used similar ideas to prove Gauss's conjecture on imaginary quadratic fields of class number one.
Gross–Zagier theorem
The Gross–Zagier theorem describes the height of Heegner points in terms of a derivative of the L-function of the elliptic curve at the point s = 1. In particular if the elliptic curve has (analytic) rank 1, then the Heegner points can be used to construct a rational point on the curve of infinite order (so the Mordell–Weil group has rank at least 1). More generally, showed that Heegner points could be used to construct rational points on the curve for each positive integer n, and the heights of these points were the coefficients of a modular form of weight 3/2. Shou-Wu Zhang generalized the Gross–Zagier theorem from elliptic curves to the case of modular abelian varieties (, ).
Birch and Swinnerton-Dyer conjecture
Kolyvagin later used Heegner points to construct Euler systems, and used this to prove much of the Birch–Swinnerton-Dyer conjecture for rank 1 elliptic curves. Brown proved the Birch–Swinnerton-Dyer conjecture for most rank 1 elliptic curves over global fields of positive characteristic .
Computation
Heegner points can be used to compute very large rational points on rank 1 elliptic curves (see for a survey) that could not be found by naive methods. Implementations of the algorithm are available in Magma, PARI/GP, and Sage.
|
https://en.wikipedia.org/wiki/Rigid%20analytic%20space
|
In mathematics, a rigid analytic space is an analogue of a complex analytic space over a nonarchimedean field. Such spaces were introduced by John Tate in 1962, as an outgrowth of his work on uniformizing p-adic elliptic curves with bad reduction using the multiplicative group. In contrast to the classical theory of p-adic analytic manifolds, rigid analytic spaces admit meaningful notions of analytic continuation and connectedness.
Definitions
The basic rigid analytic object is the n-dimensional unit polydisc, whose ring of functions is the Tate algebra , made of power series in n variables whose coefficients approach zero in some complete nonarchimedean field k. The Tate algebra is the completion of the polynomial ring in n variables under the Gauss norm (taking the supremum of coefficients), and the polydisc plays a role analogous to that of affine n-space in algebraic geometry. Points on the polydisc are defined to be maximal ideals in the Tate algebra, and if k is algebraically closed, these correspond to points in whose coordinates have norm at most one.
An affinoid algebra is a k-Banach algebra that is isomorphic to a quotient of the Tate algebra by an ideal. An affinoid is then the subset of the unit polydisc on which the elements of this ideal vanish, i.e., it is the set of maximal ideals containing the ideal in question. The topology on affinoids is subtle, using notions of affinoid subdomains (which satisfy a universality property with respect to maps of affinoid algebras) and admissible open sets (which satisfy a finiteness condition for covers by affinoid subdomains). In fact, the admissible opens in an affinoid do not in general endow it with the structure of a topological space, but they do form a Grothendieck topology (called the G-topology), and this allows one to define good notions of sheaves and gluing of spaces.
A rigid analytic space over k is a pair describing a locally ringed G-topologized space with a sheaf of k-algebras, such tha
|
https://en.wikipedia.org/wiki/Configurable%20modularity
|
Configurable modularity is a term coined by Raoul de Campo of IBM Research and later expanded on by Nate Edwards of the same organization, denoting the ability to reuse independent components by changing their interconnections, but not their internals. In Edwards' view this characterizes all successful reuse systems, and indeed all systems which can be described as "engineered".
See also
Flow-Based Programming
|
https://en.wikipedia.org/wiki/Hyperostosis
|
Hyperostosis is an excessive growth of bone. It may lead to exostosis. It occurs in many musculoskeletal disorders.
See also
Diffuse idiopathic skeletal hyperostosis
Hyperostosis frontalis interna
Infantile cortical hyperostosis
Porotic hyperostosis
SAPHO syndrome
|
https://en.wikipedia.org/wiki/Tekno%20%28toy%20manufacturer%29
|
Tekno is a Danish manufacturing company (as "Dansk Legetojs Industri") of scale model trucks and other vehicles, currently headquartered in De Lier, Netherlands. Originally established and based in Copenhagen, Tekno began manufacturing construction toys in 1928 and model vehicles immediately after World War II, selling 1 million a year during its peak.
While begun as a toy company, the focus later shifted to promotional truck models as adult collectibles and the company's headquarters was moved to The Netherlands. In the past, Tekno also manufactured model cars and airplanes.
History
Established in 1928, Tekno was the only Scandinavian company to become an accomplished die-cast toy producer, somewhat similar to the German Schuco Modell, Märklin, or Gama Toys. All of these were model firms that existed before World War II. The founder was A. Siegumfeldt, a plumber from Copenhagen, who after many years of producing wooden toys, began producing die-cast vehicles post-war perhaps late 1945. Through the 1950s, Tekno gradually became a main competitor, even to Dinky Toys.
Brochures show that the company regally celebrated its 25th anniversary in 1953. By this time a wide variety of trucks, tractors and cars were being produced along with toy miniature pots, pans and other kitchenware (Tekno Dansk 1953). Wooden houses, cranes, trucks and other toys - even small electric motors - were also made. Also in the 1950s, sometimes toys were produced in association with the Swedish company Brio.
Diecasters involved
Tekno die-cast cars and airplanes were manufactured previous to world war 2, and then after, in the main factory in Kirkebjerg, west of central Copenhagen. The complete story of toy manufacturing in the company is quite complex. From 1949 through about 1969, many Tekno models were also made by H. Langes Legetøj (Langes Toys), a diecaster in Copenhagen with which Tekno apparently had an informal association.
Langes Legetøj also cooperated with a company called Te
|
https://en.wikipedia.org/wiki/Blewit
|
Blewit refers to two closely related species of edible agarics in the genus Clitocybe, the wood blewit (Clitocybe nuda) and the field blewit or blue-leg (C.saeva). Both species are treated by some authorities as belonging to the genus Lepista.
Classification
Both species have been treated by many authorities as belonging to the Clitocybe segregate genus Lepista. Recent molecular research suggests the genus Lepista is nested within Clitocybe.
Edibility
Both wood blewits and field blewits are generally regarded as edible, but they are known to cause allergic reactions in sensitive individuals. This is particularly likely if the mushroom is consumed raw, though allergic reactions are known even from cooked blewits. Wood blewits contain the sugar trehalose, which is edible for most people.
Field blewits are often infested with fly larvae and do not store very well; they should therefore be used soon after picking. They are also very porous, so they are best picked on a dry day.
In most mycologists' opinion, the blewits are considered excellent mushrooms, despite their coloration. Blewits can be eaten as a cream sauce or sautéed in butter, but it is important not to eat them raw, which could lead to indigestion. They can also be cooked like tripe or as omelette filling, and wood blewits also make good stewing mushrooms.
Footnotes
External links
"Mushroom-Collecting.com - The Blewit"
All that Rain Promises and More - Blewit
Edible fungi
Clitocybe
Fungus common names
|
https://en.wikipedia.org/wiki/Texas%20Institute%20for%20Genomic%20Medicine
|
The Texas A&M Institute for Genomic Medicine (TIGM) is a research institute of Texas A&M AgriLife Research. It was founded in 2005 under a $50 million award from the Texas Enterprise Fund to accelerate the pace of medical discoveries and foster the development of the biotechnology industry in Texas.
TIGM helps researchers gain faster access to the genetically engineered knockout mice used in medical research. TIGM owns and maintains the world's largest library of embryonic stem cells for C57BL/6 mice. In addition, TIGM has contracted access to the world's largest library of genetically modified 129 mouse cells. The Institute headquarters and laboratory facilities are based on the main campus of Texas A&M University in College Station, Texas.
|
https://en.wikipedia.org/wiki/Valis%3A%20The%20Fantasm%20Soldier
|
is a 1986 action-platform video game originally developed by Wolf Team and published by Telenet Japan for the MSX, PC-8801, X1, FM-7, and PC-9801 home computers. It is the first entry in the Valis series. It stars Yuko Asou, a Japanese teenage schoolgirl chosen as the Valis warrior and wielder of the mystical Valis sword to protect the Earth, the land of spirits, and the dream world Vecanti from demon lord Rogles. Through the journey, the player explores and search for items and power-ups, while fighting enemies and defeating bosses to increase Yuko's attributes.
Programmers Masahiro Akishino and Osamu Ikegame began planning on a side-scrolling action game featuring a customed delinquent heroine, an idea originated from Sukeban Deka to compete in a contest sponsored by Japanese computer magazine LOGiN, being kept secret within Telenet until they approved development to continue when the company learned of its existence. After a Telenet superior expressed disliking towards its graphics, writer Hiroki Hayashi was ordered to take action and fix it, leading to the conception of Valis. Akishino and Hayashi used Ikegame's work as basis to introduce their own story and character ideas, which were based on an unfinished personal novel Hayashi wrote prior to the game's production.
Valis sold well and was listed as one of the best-selling games in 1987 rankings. An almost completely reworked version was also released for the Family Computer, followed by remakes for the Sega Mega Drive/Genesis and PC Engine Super CD-ROM², and a version for mobile phones as well. The game was supplemented with manga adaptations, an anime short by Sunrise, albums from King Records and Wave Master, and doujinshi books. Critical reception has varied depending on the version; the original MSX version garnered mixed reviews while the Genesis remake carried average sentiments, however the enhanced PC Engine remake was received more favorably. It was followed by Valis II (1989).
Gameplay and prem
|
https://en.wikipedia.org/wiki/Kallidin
|
Kallidin is a bioactive kinin formed in response to injury from kininogen precursors through the action of kallikreins.
Kallidin is a decapeptide whose sequence is H-Lys-Arg-Pro-Pro-Gly-Phe-Ser-Pro-Phe-Arg-OH. It can be converted to bradykinin by the aminopeptidase enzyme.
It can be a substrate for carboxypeptidase M and N.
Kallidin is identical to bradykinin with an additional lysine residue added at the N-terminal end and signals through the bradykinin receptor.
|
https://en.wikipedia.org/wiki/Claranet
|
Claranet provides network, hosting and managed application services in the UK, France, Germany, The Netherlands (Benelux), Portugal, Spain, Italy and Brazil.
History
Charles Nasser founded the ISP in 1996 and by 1999 had 150,000 subscribers.
Claranet has grown its business through a number of acquisitions, including Netscalibur in 2003, via net.works uk in 2004 and in 2005 Amen Group, via net.works Europe and Artful. In 2012 Claranet acquired Star Technology.
The company has annualised revenues of circa £375 million, over 6,500 customers and over 2,200 employees. On a constant currency basis, revenues have increased four times in under five years. Claranet was recognised as a ‘Leader’ in Gartner’s Magic Quadrant for Managed Hybrid Cloud Hosting, Europe (2016) for the fourth consecutive year and holds Premier Partner status with Amazon Web Services and Google Cloud.
In 2017 Claranet acquired French company Oxalide and ITEN Solutions, revolutionising the IT market in Portugal.
On 5 July 2018, Claranet acquired NotSoSecure.
|
https://en.wikipedia.org/wiki/Transport%20Phenomena%20%28book%29
|
Transport Phenomena is the first textbook about transport phenomena. It is specifically designed for chemical engineering students. The first edition was published in 1960, two years after having been preliminarily published under the title Notes on Transport Phenomena based on mimeographed notes prepared for a chemical engineering course taught at the University of Wisconsin–Madison during the academic year 1957-1958. The second edition was published in August 2001. A revised second edition was published in 2007. This text is often known simply as BSL after its authors' initials.
History
As the chemical engineering profession developed in the first half of the 20th century, the concept of "unit operations" arose as being needed in the education of undergraduate chemical engineers. The theories of mass, momentum and energy transfer were being taught at that time only to the extent necessary for a narrow range of applications. As chemical engineers began moving into a number of new areas, problem definitions and solutions required a deeper knowledge of the fundamentals of transport phenomena than those provided in the textbooks then available on unit operations.
In the 1950s, R. Byron Bird, Warren E. Stewart and Edwin N. Lightfoot stepped forward to develop an undergraduate course at the University of Wisconsin–Madison to integrate the teaching of fluid flow, heat transfer, and diffusion. From this beginning, they prepared their landmark textbook Transport Phenomena.
Subjects covered in the book
The book is divided into three basic sections, named Momentum Transport, Energy Transport and Mass Transport:
Momentum Transport
Viscosity and the Mechanisms of Momentum Transport
Momentum Balances and Velocity Distributions in Laminar Flow
The Equations of Change for Isothermal Systems
Velocity Distributions in Turbulent Flow
Interphase Transport in Isothermal Systems
Macroscopic Balances for Isothermal Flow Systems
Energy Transport
Thermal Conductivity and the Me
|
https://en.wikipedia.org/wiki/Abel%20equation
|
The Abel equation, named after Niels Henrik Abel, is a type of functional equation of the form
or
.
The forms are equivalent when is invertible. or control the iteration of .
Equivalence
The second equation can be written
Taking , the equation can be written
For a known function , a problem is to solve the functional equation for the function , possibly satisfying additional requirements, such as .
The change of variables , for a real parameter , brings Abel's equation into the celebrated Schröder's equation, .
The further change into Böttcher's equation, .
The Abel equation is a special case of (and easily generalizes to) the translation equation,
e.g., for ,
. (Observe .)
The Abel function further provides the canonical coordinate for Lie advective flows (one parameter Lie groups).
History
Initially, the equation in the more general form
was reported. Even in the case of a single variable, the equation is non-trivial, and admits special analysis.
In the case of a linear transfer function, the solution is expressible compactly.
Special cases
The equation of tetration is a special case of Abel's equation, with .
In the case of an integer argument, the equation encodes a recurrent procedure, e.g.,
and so on,
Solutions
The Abel equation has at least one solution on if and only if for all and all , , where , is the function iterated times.
Analytic solutions (Fatou coordinates) can be approximated by asymptotic expansion of a function defined by power series in the sectors around a parabolic fixed point. The analytic solution is unique up to a constant.
See also
Functional equation
Schröder's equation
Böttcher's equation
Infinite compositions of analytic functions
Iterated function
Shift operator
Superfunction
|
https://en.wikipedia.org/wiki/Schr%C3%B6der%27s%20equation
|
Schröder's equation, named after Ernst Schröder, is a functional equation with one independent variable: given the function , find the function such that
Schröder's equation is an eigenvalue equation for the composition operator that sends a function to .
If is a fixed point of , meaning , then either (or ) or . Thus, provided that is finite and does not vanish or diverge, the eigenvalue is given by .
Functional significance
For , if is analytic on the unit disk, fixes , and , then Gabriel Koenigs showed in 1884 that there is an analytic (non-trivial) satisfying Schröder's equation. This is one of the first steps in a long line of theorems fruitful for understanding composition operators on analytic function spaces, cf. Koenigs function.
Equations such as Schröder's are suitable to encoding self-similarity, and have thus been extensively utilized in studies of nonlinear dynamics (often referred to colloquially as chaos theory). It is also used in studies of turbulence, as well as the renormalization group.
An equivalent transpose form of Schröder's equation for the inverse of Schröder's conjugacy function is . The change of variables (the Abel function) further converts Schröder's equation to the older Abel equation, . Similarly, the change of variables converts Schröder's equation to Böttcher's equation, .
Moreover, for the velocity, , Julia's equation, , holds.
The -th power of a solution of Schröder's equation provides a solution of Schröder's equation with eigenvalue , instead. In the same vein, for an invertible solution of Schröder's equation, the (non-invertible) function is also a solution, for any periodic function with period . All solutions of Schröder's equation are related in this manner.
Solutions
Schröder's equation was solved analytically if is an attracting (but not superattracting)
fixed point, that is by Gabriel Koenigs (1884).
In the case of a superattracting fixed point, , Schröder's equation is unwieldy, and
|
https://en.wikipedia.org/wiki/Area-to-area%20Lee%20model
|
The Lee model for area-to-area mode is a radio propagation model that operates around 900 MHz. Built as two different modes, this model includes an adjustment factor that can be adjusted to make the model more flexible to different regions of propagation.
Applicable to/under conditions
This model is suitable for using in data collected. The model predicts the behaviour of all links that has ends in specific areas.
Coverage
Frequency: 900 MHz band
Mathematical formulation
The model
The Lee model is formally expressed as:
where,
L = The median path loss. Unit: decibel (dB)
L0 = The reference path loss along 1 km. Unit: decibel (dB)
= The slope of the path loss curve. Unit: decibels per decade
d = The distance on which the path loss is to be calculated.
FA = Adjustment factor.
Calculation of reference path loss
The reference path loss is usually computed along a 1 km or 1 mile link. Any other suitable length of path can be chosen based on the applications.
where,
GB = Base station antenna gain. Unit: decibel with respect to isotropic antenna (dBi)
= Wavelength. Unit: meter (m).
GM = Mobile station antenna gain. Unit: decibel with respect to isotropic antenna (dBi).
Calculation of adjustment factors
The adjustment factor is calculated as:
where,
FBH = Base station antenna height correction factor.
FBG = Base station antenna gain correction factor.
FMH = Mobile station antenna height correction factor.
FMG = Mobile station antenna gain correction factor.
FF = Frequency correction factor
Base-station antenna height correction factor
where,
hB = Base-station antenna height. Unit: meter (m).
or
where,
hB = Base-station antenna height. Unit: foot (ft).
Base-station antenna gain correction factor
where,
GB = Base-station antenna gain. Unit: decibel with respect to half wave dipole antenna (dBd)
Mobile-station antenna height correction factor
where,
hM = Mobile-station antenna height. Unit: meter(m).
Mobile-antenna gain correction factor
where,
GM
|
https://en.wikipedia.org/wiki/International%20Society%20for%20Intelligence%20Research
|
The International Society for Intelligence Research (ISIR) is a scientific society for researchers in human intelligence. It was founded by Douglas K. Detterman of Case Western Reserve University in 2000.
The society advocates for ongoing support for scientific research on cognitive ability. A 2018 New Statesman article called two editors of Intelligence "eugenicists" and that ISIR conferences have included speakers who are part of "the infiltration of mainstream academia by eugenicists".
The society runs the journal Intelligence.
Presidents
The following persons are or have been president of the society:
2010 Douglas K. Detterman
2011 Earl B. Hunt
2012 Linda Gottfredson
2013 David Lubinski
2014 Aljoscha C. Neubauer
2015 Michael McDaniel
2016 Richard J. Haier
2017 Timothy Bates
2018 William Revelle
2019 Rex Jung
|
https://en.wikipedia.org/wiki/Potential%20vorticity
|
In fluid mechanics, potential vorticity (PV) is a quantity which is proportional to the dot product of vorticity and stratification. This quantity, following a parcel of air or water, can only be changed by diabatic or frictional processes. It is a useful concept for understanding the generation of vorticity in cyclogenesis (the birth and development of a cyclone), especially along the polar front, and in analyzing flow in the ocean.
Potential vorticity (PV) is seen as one of the important theoretical successes of modern meteorology. It is a simplified approach for understanding fluid motions in a rotating system such as the Earth's atmosphere and ocean. Its development traces back to the circulation theorem by Bjerknes in 1898, which is a specialized form of Kelvin's circulation theorem. Starting from Hoskins et al., 1985, PV has been more commonly used in operational weather diagnosis such as tracing dynamics of air parcels and inverting for the full flow field. Even after detailed numerical weather forecasts on finer scales were made possible by increases in computational power, the PV view is still used in academia and routine weather forecasts, shedding light on the synoptic scale features for forecasters and researchers.
Baroclinic instability requires the presence of a potential vorticity gradient along which waves amplify during cyclogenesis.
Bjerknes circulation theorem
Vilhelm Bjerknes generalized Helmholtz's vorticity equation (1858) and Kelvin's circulation theorem (1869) to inviscid, geostrophic, and baroclinic fluids, i.e., fluids of varying density in a rotational frame which has a constant angular speed. If we define circulation as the integral of the tangent component of velocity around a closed fluid loop and take the integral of a closed chain of fluid parcels, we obtain
(1)
where is the time derivative in the rotational frame (not inertial frame), is the relative circulation, is projection of the area surrounded by the fluid loop on the
|
https://en.wikipedia.org/wiki/William%20Revelle
|
William Roger Revelle (born c. 1944) is a psychology professor at Northwestern University working in personality psychology. Revelle studies the biological basis of personality and motivation, psychometric theory, the structure of daily mood, and models of attention and memory.
Early life and education
Revelle was raised in La Jolla, California. His father, Roger Revelle, was an early theorist in global warming.
Revelle graduated from Pomona College in 1965, abandoning a mathematics major in favor of psychology. He spent two years in Sarawak, Malaysia, as a volunteer in the Peace Corps before earning his PhD in psychology from the University of Michigan in 1973. He became a member of the Northwestern Faculty in 1973.
Career
Revelle has previously served as the President (2005-2009) of the International Society for the Study of Individual Differences (ISSID), the President (2008-2009) of the Association for Research in Personality (ARP), and the President (1984) of the Society of Multivariate Experimental Psychology (SMEP).
Currently, he is vice-chair of the Governing Board of the Bulletin of the Atomic Scientists, having previously served as Chair (2009-2012). He also serves as the President (2018–present) of the International Society for Intelligence Research (ISIR).
Additionally, he is a Fellow of the American Association for the Advancement of Science (AAAS; 1996–present), the Association for Psychological Science (APS; 1994–present), the American Psychological Association (APA Division 5; 2011–present), and the Society for Personality and Social Psychology (SPSP; 2015–present).
He resides in Evanston, Illinois.
Bibliography
Selected publications
Software
|
https://en.wikipedia.org/wiki/Diagnosis%20%28artificial%20intelligence%29
|
As a subfield in artificial intelligence, diagnosis is concerned with the development of algorithms and techniques that are able to determine whether the behaviour of a system is correct. If the system is not functioning correctly, the algorithm should be able to determine, as accurately as possible, which part of the system is failing, and which kind of fault it is facing. The computation is based on observations, which provide information on the current behaviour.
The expression diagnosis also refers to the answer of the question of whether the system is malfunctioning or not, and to the process of computing the answer. This word comes from the medical context where a diagnosis is the process of identifying a disease by its symptoms.
Example
An example of diagnosis is the process of a garage mechanic with an automobile. The mechanic will first try to detect any abnormal behavior based on the observations on the car and his knowledge of this type of vehicle. If he finds out that the behavior is abnormal, the mechanic will try to refine his diagnosis by using new observations and possibly testing the system, until he discovers the faulty component; the mechanic plays an important role in the vehicle diagnosis.
Expert diagnosis
The expert diagnosis (or diagnosis by expert system) is based on experience with the system. Using this experience, a mapping is built that efficiently associates the observations to the corresponding diagnoses.
The experience can be provided:
By a human operator. In this case, the human knowledge must be translated into a computer language.
By examples of the system behaviour. In this case, the examples must be classified as correct or faulty (and, in the latter case, by the type of fault). Machine learning methods are then used to generalize from the examples.
The main drawbacks of these methods are:
The difficulty acquiring the expertise. The expertise is typically only available after a long period of use of the system
|
https://en.wikipedia.org/wiki/Urine-indicator%20dye
|
Urine-indicator dye is a mythical substance that is supposed to be able to react with urine to form a colored cloud in a swimming pool or hot tub, thus indicating the location of people who are urinating while they are in the water. A 2015 report from the National Swimming Pool Foundation called this "the most common pool myth of all time", with nearly half of Americans surveyed by researchers believing that the dye existed.
Urine is difficult to detect, as many of the naturally occurring compounds within urine are unstable and react freely with common disinfectants, such as chlorine, creating a large number of disinfection by-product (DBP) compounds from the original organic chemicals in urine.
Rumours of the origin of urine indicator-dye go back at least as far as 1958, and the story is commonly told to children by parents who do not want them to urinate in the pool. A 1985 biography of Orson Welles describes him using such a dye as part of a prank in 1937.
|
https://en.wikipedia.org/wiki/Email%20bomb
|
On Internet usage, an email bomb is a form of net abuse that sends large volumes of email to an address to overflow the mailbox, overwhelm the server where the email address is hosted in a denial-of-service attack (DoS attack) or as a smoke screen to distract the attention from important email messages indicating a security breach.
Methods
There are three methods of perpetrating an email bomb: mass mailing, list linking and zip bombing.
Mass mailing
Mass mailing consists of sending numerous duplicate emails to the same email address. These types of mail bombs are simple to design but their extreme simplicity means they can be easily detected by spam filters. Email-bombing using mass mailing is also commonly performed as a DDoS attack by employing the use of "zombies" botnets; hierarchical networks of computers compromised by malware and under the attacker's control. Similar to their use in spamming, the attacker instructs the botnet to send out millions of emails, but unlike normal botnet spamming, the emails are all addressed to only one or a few addresses the attacker wishes to flood. This form of email bombing is similar to other DDoS flooding attacks. As the targets are frequently the dedicated hosts handling website and email accounts of a business, this type of attack can be devastating to both services of the host.
This type of attack is more difficult to defend against than a simple mass-mailing bomb because of the multiple source addresses and the possibility of each zombie computer sending a different message or employing stealth techniques to defeat spam filters.
List linking
List linking, also known as "email cluster bomb", means signing a particular email address up to several email list subscriptions. The victim then has to unsubscribe from these unwanted services manually. The attack can be carried out automatically with simple scripts: this is easy, almost impossible to trace back to the perpetrator, and potentially very destructive. A massive
|
https://en.wikipedia.org/wiki/Decorrelation
|
Decorrelation is a general term for any process that is used to reduce autocorrelation within a signal, or cross-correlation within a set of signals, while preserving other aspects of the signal. A frequently used method of decorrelation is the use of a matched linear filter to reduce the autocorrelation of a signal as far as possible. Since the minimum possible autocorrelation for a given signal energy is achieved by equalising the power spectrum of the signal to be similar to that of a white noise signal, this is often referred to as signal whitening.
Process
Although most decorrelation algorithms are linear, non-linear decorrelation algorithms also exist.
Many data compression algorithms incorporate a decorrelation stage. For example, many transform coders first apply a fixed linear transformation that would, on average, have the effect of decorrelating a typical signal of the class to be coded, prior to any later processing. This is typically a Karhunen–Loève transform, or a simplified approximation such as the discrete cosine transform.
By comparison, sub-band coders do not generally have an explicit decorrelation step, but instead exploit the already-existing reduced correlation within each of the sub-bands of the signal, due to the relative flatness of each sub-band of the power spectrum in many classes of signals.
Linear predictive coders can be modelled as an attempt to decorrelate signals by subtracting the best possible linear prediction from the input signal, leaving a whitened residual signal.
Decorrelation techniques can also be used for many other purposes, such as reducing crosstalk in a multi-channel signal, or in the design of echo cancellers.
In image processing decorrelation techniques can be used to enhance or stretch, colour differences found in each pixel of an image. This is generally termed as 'decorrelation stretching'.
The concept of decorrelation can be applied in many other fields.
In neuroscience, decorrelation is used in the an
|
https://en.wikipedia.org/wiki/United%20States%20Department%20of%20Veterans%20Affairs%20emblems%20for%20headstones%20and%20markers
|
The United States Department of Veterans Affairs (VA) maintains many cemeteries specifically devoted to veterans. Most have various rules regarding what must take place in order to be interred there.
Procedure
The VA only permits graphics on government-furnished headstones or markers that are approved emblems of belief, the Civil War Union Shield (including those who served in the U.S. military through the Spanish–American War), the Civil War Confederate Southern Cross of Honor, and the Medal of Honor insignia. Arlington National Cemetery has similar restrictions on headstones, though it is maintained by US Department of the Army.
The religious symbols are rendered as simple inscriptions without sculptural relief or coloring other than black. The emblem of belief is an optional feature.
Generally the VA adds a new symbol a few months after receiving a petition from a faith group. However, the Wiccan symbol was only added in 2007 to settle a lawsuit filed on behalf of several families by Americans United for the Separation of Church and State in November 2006. A separate parallel lawsuit was filed on behalf of two Wiccan churches and three families by the American Civil Liberties Union in September 2006, which was resolved by the same settlement.
The first interfaith headstone, which includes a Wiccan pentacle for Jan Deanna O'Rourke and a Presbyterian Cross for her husband, was installed at Arlington National Cemetery on May 1, 2007, and dedicated on July 4, 2007.
Headstone and marker symbols
The following emblems and emblem numbers are publicized as available for government headstones and markers as of May 2023. A process is in place to consider approving additional religious or belief system emblems requested by the families of individuals eligible for these headstones and markers.
Each emblem is given its official USVA name and designation, with added additional links for related symbolism (*) and for related movements (†). Explanatory footnotes are provide
|
https://en.wikipedia.org/wiki/Road%20Blaster
|
is an interactive movie video game developed by Data East featuring animation by Toei Animation, originally released exclusively in Japan as a laserdisc-based arcade game in 1985. The player assumes the role of a vigilante who must avenge the death of his wife by pursuing the biker gang responsible for her death in a modified sports car. The game would later be ported to a variety of home formats such as the MSX and Sharp X1 (VHD format), Sega CD (under the title of Road Blaster FX), LaserActive (in Mega-LD format), PlayStation and Sega Saturn (in a compilation with Thunder Storm). The Sega CD and Mega-LD versions were released outside of Japan under titles of Road Avenger and Road Prosecutor respectively.
Gameplay
As with other laserdisc-based arcade games from the same time, the gameplay consists of on-screen instructions overlaid over pre-recorded full motion video animated footage of high-speed chases and vehicular combat. The player controls the crosshair from a first-person perspective, to steer their car toward the correct directions according to the green arrows flashing and beeping beside it, while controlling the gas pedal, brake and booster whenever they light up.
The game has nine stages. Upon successfully completing a level, the player is graded on the reaction time. Different difficulty levels can be selected. In Normal Mode, pop-up icons and audio tones signal when to turn left or right, brake, hit turbo, or hit other cars. In Hard Mode, there are no on-screen icons to guide the player.
Plot
The story of Road Blaster is inspired by revenge thriller films such as Mad Max. In the late 1990s United States, the player assumes the role of a vigilante who drives a customized sports car in order to get revenge on a biker gang responsible for his wife's death on their honeymoon. After recovering from his own injuries, he upgrades his car and goes on a rampage through nine areas. His goal is to seek out the gang's female boss and complete his vengeance.
De
|
https://en.wikipedia.org/wiki/Nimrod%20%28computer%29
|
The Nimrod, built in the United Kingdom by Ferranti for the 1951 Festival of Britain, was an early computer custom-built to play Nim, inspired by the earlier Nimatron. The twelve-by-nine-by-five-foot (3.7-by-2.7-by-1.5-meter) computer, designed by John Makepeace Bennett and built by engineer Raymond Stuart-Williams, allowed exhibition attendees to play a game of Nim against an artificial intelligence. The player pressed buttons on a raised panel corresponding with lights on the machine to select their moves, and the Nimrod moved afterward, with its calculations represented by more lights. The speed of the Nimrod's calculations could be reduced to allow the presenter to demonstrate exactly what the computer was doing, with more lights showing the state of the calculations. The Nimrod was intended to demonstrate Ferranti's computer design and programming skills rather than to entertain, though Festival attendees were more interested in playing the game than the logic behind it. After its initial exhibition in May, the Nimrod was shown for three weeks in October 1951 at the Berlin Industrial Show before being dismantled.
The game of Nim running on the Nimrod is a candidate for one of the first video games, as it was one of the first computer games to have any sort of visual display of the game. It appeared only four years after the 1947 invention of the cathode-ray tube amusement device, the earliest known interactive electronic game to use an electronic display, and one year after Bertie the Brain, a computer similar to the Nimrod which played tic-tac-toe at the 1950 Canadian National Exhibition. The Nimrod's use of light bulbs rather than a screen with real-time visual graphics, however, much less moving graphics, does not meet some definitions of a video game.
Development
In the summer of 1951, the United Kingdom held the Festival of Britain, a national exhibition held throughout the UK to promote the British contribution to science, technology, industrial design,
|
https://en.wikipedia.org/wiki/Kat%C4%9Btov%E2%80%93Tong%20insertion%20theorem
|
The Katětov–Tong insertion theorem is a theorem of point-set topology proved independently by Miroslav Katětov and Hing Tong in the 1950s. The theorem states the following:
Let be a normal topological space and let be functions with g upper semicontinuous, h lower semicontinuous and . Then there exists a continuous function with
This theorem has a number of applications and is the first of many classical insertion theorems. In particular it implies the Tietze extension theorem and consequently Urysohn's lemma, and so the conclusion of the theorem is equivalent to normality.
|
https://en.wikipedia.org/wiki/Sequential%20coupling
|
In object-oriented programming, sequential coupling (also known as temporal coupling) is a form of coupling where a class requires its methods to be called in a particular sequence. This may be an anti-pattern, depending on context.
Methods whose name starts with Init, Begin, Start, etc. may indicate the existence of sequential coupling.
Using a car as an analogy, if the user steps on the gas without first starting the engine, the car does not crash, fail, or throw an exception - it simply fails to accelerate.
Sequential coupling can be refactored with the template method pattern to overcome the problems posed by the usage of this anti-pattern.
|
https://en.wikipedia.org/wiki/Conical%20function
|
In mathematics, conical functions or Mehler functions are functions which can be expressed in terms of Legendre functions of the first and second kind,
and
The functions were introduced by Gustav Ferdinand Mehler, in 1868, when expanding in series the distance of a point on the axis of a cone to a point located on the surface of the cone. Mehler used the notation to represent these functions. He obtained integral representation and series of functions representations for them. He also established an addition theorem
for the conical functions. Carl Neumann obtained an expansion of the functions in terms
of the Legendre polynomials in 1881. Leonhardt introduced for the conical functions the equivalent of the spherical harmonics in 1882.
External links
G. F. Mehler "Ueber die Vertheilung der statischen Elektricität in einem von zwei Kugelkalotten begrenzten Körper" Journal für die reine und angewandte Mathematik 68, 134 (1868).
G. F. Mehler "Ueber eine mit den Kugel- und Cylinderfunctionen verwandte Function und ihre Anwendung in der Theorie der Elektricitätsvertheilung" Mathematische Annalen 18 p. 161 (1881).
C. Neumann "Ueber die Mehler'schen Kegelfunctionen und deren Anwendung auf elektrostatische Probleme" Mathematische Annalen 18 p. 195 (1881).
G. Leonhardt " Integraleigenschaften der adjungirten Kegelfunctionen" Mathematische Annalen 19 p. 578 (1882).
Milton Abramowitz and Irene Stegun (Eds.) Handbook of Mathematical Functions (Dover, 1972) p. 337
A. Gil, J. Segura, N. M. Temme "Computing the conical function $P^{\mu}_{-1/2+i\tau}(x)$" SIAM J. Sci. Comput. 31(3), 1716–1741 (2009).
Tiwari, U. N.; Pandey, J. N. The Mehler-Fock transform of distributions. Rocky Mountain J. Math. 10 (1980), no. 2, 401–408.
Special functions
|
https://en.wikipedia.org/wiki/P2X%20purinoreceptor
|
The P2X receptors, also ATP-gated P2X receptor cation channel family, is a protein family that consists of cation-permeable ligand-gated ion channels that open in response to the binding of extracellular adenosine 5'-triphosphate (ATP). They belong to a larger family of receptors known as the ENaC/P2X superfamily. ENaC and P2X receptors have similar 3-D structures and are homologous. P2X receptors are present in a diverse array of organisms including humans, mouse, rat, rabbit, chicken, zebrafish, bullfrog, fluke, and amoeba.
Physiological roles
P2X receptors are involved in a variety of physiological processes, including:
Modulation of cardiac rhythm and contractility
Modulation of vascular tone
Mediation of nociception, especially chronic pain
Contraction of the vas deferens during ejaculation
Contraction of the urinary bladder during micturition
Platelet aggregation
Macrophage activation
Apoptosis
Neuronal-glial integration
Tissue distribution
P2X receptors are expressed in cells from a wide variety of animal tissues. On presynaptic and postsynaptic nerve terminals and glial cells throughout the central, peripheral and autonomic nervous systems, P2X receptors have been shown to modulate synaptic transmission. Furthermore, P2X receptors are able to initiate contraction in cells of the heart muscle, skeletal muscle, and various smooth muscle tissues, including that of the vasculature, vas deferens and urinary bladder. P2X receptors are also expressed on leukocytes, including lymphocytes and macrophages, and are present on blood platelets. There is some degree of subtype specificity as to which P2X receptor subtypes are expressed on specific cell types, with P2X1 receptors being particularly prominent in smooth muscle cells, and P2X2 being widespread throughout the autonomic nervous system. However, such trends are very general and there is considerable overlap in subunit distribution, with most cell types expressing more than one subunits. For example, P
|
https://en.wikipedia.org/wiki/International%20Conference%20on%20Distributed%20Computing%20Systems
|
The International Conference on Distributed Computing Systems (ICDCS) is the oldest conference in the field of distributed computing systems in the world. It was launched by the IEEE Computer Society Technical Committee on Distributed Processing (TCDP) in October 1979, and is sponsored by such committee. It was started as an 18-month conference until 1983 and became an annual conference since 1984. The ICDCS has a long history of significant achievements and worldwide visibility, and has recently celebrated its 37th year.
Location history
2019: Dallas, Texas, United States
2018: Vienna, Austria
2017: Atlanta, GA, United States
2016: Nara, Japan
2015: Columbus, Ohio, United States
2014: Madrid, Spain
2013: Philadelphia, Pennsylvania, United States
2012: Macau, China
2011: Minneapolis, Minnesota, United States
2010: Genoa, Italy
2009: Montreal, Quebec, Canada
2008: Beijing, China
2007: Toronto, Ontario, Canada
2006: Lisbon, Portugal
2005: Columbus, Ohio, United States
2004: Keio University, Japan
2003: Providence, RI, United States
2002: Vienna, Austria
2001: Phoenix, AZ, United States
2000: Taipei, Taiwan
1999: Austin, TX, United States
1998: Amsterdam, The Netherlands
1997: Baltimore, MD, United States
1996: Hong Kong
1995: Vancouver, Canada
1994: Poznań, Poland
1993: Pittsburgh, PA, United States
1992: Yokohama, Japan
1991: Arlington, TX, United States
1990: Paris, France
1989: Newport Beach, CA, United States
1988: San Jose, CA, United States
1987: Berlin, Germany
1986: Cambridge, MA, United States
1985: Denver, CO, United States
1984: San Francisco, CA, United States
1983: Hollywood, FL, United States
1981: Versailles, France
1979: Huntsville, AL, United States
See also
List of distributed computing conferences
External links
ICDCS 2018 - July 2–July 5, 2018, Vienna, Austria
ICDCS 2007 - June 25–June 29, 2007, Toronto, Canada.
ICDCS 2006 - July 4–July 7, 2006, Lisbon, Portugal.
ICDCS 2005 - July 6–July 10, 2005, Co
|
https://en.wikipedia.org/wiki/Aging%20in%20cats
|
Aging in cats is the process by which cats change over the course of their natural lifespans. The average lifespan of a domestic cat may range from 10 to 13 years. As cats senesce, they undergo predictable changes in health and behavior. Dental disease and loss of olfaction are common as cats age, affecting eating habits. Arthritis and sarcopenia are also common in older cats. How a cat's health is affected by aging may be managed through modifications in a cat's diet, accessibility adjustments, and cognitive stimulation.
Average lifespan among domestic cats
The average lifespan of domestic cats has increased in recent decades. It has risen from seven years in the 1980s, to nine years in 1995, to about 15 years in 2021. Reliable information on the lifespans of domestic cats is varied and limited. Nevertheless, a number of studies have investigated the matter and have come up with noteworthy estimates. Estimates of mean lifespan in these studies range between 13 and 20 years, with a single value in the neighborhood of 15 years. At least one study found a median lifespan value of 14 years and a corresponding interquartile range of 9 to 17 years. Maximum lifespan has been estimated at values ranging from 22 to 30 years although there have been claims of cats living longer than 30 years. According to the 2010 edition of the Guinness World Records, the oldest cat ever recorded was Creme Puff, who died in 2005, aged 38 years, 3 days. Female cats typically outlive male cats, and crossbred cats typically outlive purebred cats. It has also been found that the greater a cat's weight, the lower its life expectancy on average.
A common misconception in cat aging (and dog aging) is that a cat ages the equivalent of what a human would age in seven years each year. This is inaccurate due to the inconsistencies in aging as well as there being far more accurate equations to predict a cat's age in "cat years". A more accurate equation often used by veterinarians to predict cat yea
|
https://en.wikipedia.org/wiki/Bisection%20bandwidth
|
In computer networking, if the network is bisected into two equal-sized partitions, the bisection bandwidth of a network topology is the bandwidth available between the two partitions. Bisection should be done in such a way that the bandwidth between two partitions is minimum. Bisection bandwidth gives the true bandwidth available in the entire system. Bisection bandwidth accounts for the bottleneck bandwidth of the entire network. Therefore bisection bandwidth represents bandwidth characteristics of the network better than any other metric.
Bisection bandwidth calculations
For a linear array with n nodes bisection bandwidth is one link bandwidth. For linear array only one link needs to be broken to bisect the network into two partitions.
For ring topology with n nodes two links should be broken to bisect the network, so bisection bandwidth becomes bandwidth of two links.
For tree topology with n nodes can be bisected at the root by breaking one link, so bisection bandwidth is one link bandwidth.
For Mesh topology with n nodes, links should be broken to bisect the network, so bisection bandwidth is bandwidth of links.
For Hyper-cube topology with n nodes, n/2 links should be broken to bisect the network, so bisection bandwidth is bandwidth of n/2 links.
Significance of bisection bandwidth
Theoretical support for the importance of this measure of network performance was developed in the PhD research of Clark Thomborson (formerly Clark Thompson). Thomborson proved that important algorithms for sorting, Fast Fourier transformation, and matrix-matrix multiplication become communication-limited—as opposed to CPU-limited or memory-limited—on computers with insufficient bisection bandwidth. F. Thomson Leighton's PhD research tightened Thomborson's loose bound on the bisection bandwidth of a computationally-important variant of the De Bruijn graph known as the shuffle-exchange network. Based on Bill Dally's analysis of latency, average-case throughput, and h
|
https://en.wikipedia.org/wiki/Genotropism
|
Genotropism is defined as the reciprocal attraction between carriers of the same or related latent recessive genes. Developed by the Hungarian psychiatrist Léopold Szondi in the 1930s, the theory concludes that instinct is biological and genetic in origin. Szondi believed that these genes regulated the "possibilities of fate" and was the working principle of the familial unconscious.
Overview
Genotropism consists of the theory that genes influence human behavior. While identified as entities, genes exist in groups because evolution favors cooperation. Within each gene group, it is possible to detect specific needs that function as mechanisms of screening and natural selection.
Szondi arrived a sort of genetic determinism, a philosophical theory of predestination. "The latent hereditary factors in human beings, the recessive genes, do not remain dormant or inactive within the human organism, but exert a very important and even decisive influence upon its behavior. This latent or recessive gene theory claims that these non-dominant hereditary factors determine the Object selection, voluntary and involuntary, of the individual. The drives resulting from these latent genes, therefore, direct the individual's selection of love objects, friendships, occupations, diseases, and forms of death. Hence, from the very beginning of the human's existence there is a hidden plan of life guided by 'Instinctual drives'."
Instinctual drives
In Szondi's theory, each "need" (a link between genes and behavior) comprises a polarity of positive and negative tendencies. Needs also group together in polarities to form larger wholes called "instinctual drives." Together, behavior tendencies, needs, and drives combine to form patterned wholes.
Szondi created a drive theory that determines that every drive has at least four genes. "The four Szondian drives are (1) contact, (2) sexual, (3) paroxysmal, and (4) ego. They are implicated in their corresponding psychiatric disorders and equival
|
https://en.wikipedia.org/wiki/MKS%20Toolkit
|
MKS Toolkit is a software package produced and maintained by PTC that provides a Unix-like environment for scripting, connectivity and porting Unix and Linux software to Microsoft Windows. It was originally created for MS-DOS, and OS/2 versions were released up to version 4.4. Several editions of each version, such as MKS Toolkit for developers, power users, enterprise developers and interoperability are available, with the enterprise developer edition being the most complete.
Before PTC, MKS Toolkit was owned by MKS Inc. In 1999, MKS acquired a company based in Fairfax, Virginia, USA called Datafocus Inc. The Datafocus product NuTCRACKER had included the MKS Toolkit since 1994 as part of its Unix compatibility technology. The MKS Toolkit was also licensed by Microsoft for the first two versions of their Windows Services for Unix, but later dropped in favor of Interix after Microsoft purchased the latter company.
Version 10.0 was current .
Overview
The MKS Toolkit products offer functionality in the following areas:
Command shell environments of Bourne shell, KornShell, Bash, C shell, Tcl shell
Traditional Unix commands (400+), including grep, awk, sed, vi, ls, kill
Windows specific commands (70+), including registry, shortcut, desktop, wcopy, db, dde, userinfo
Tape and archive commands, including tar, cpio, pax, zip, bzip2, ar
Remote connectivity, including ssh, remote shell, telnet, xterm, kterm, rexec, rlogin
Porting APIs, including fork(), signals, alarms, threads
Graphical porting APIs, including X, ncurses, Motif, OpenGL
Supported operating systems
MKS Toolkit products support all IA-32 and x64 of the Microsoft Windows operating systems. There is some loss of functionality running IA-32 versions on Windows 9x. Earlier versions ran on MS-DOS and compatible operating systems.
See also
Cygwin
MinGW
Hamilton C shell
UnxUtils
UWIN
GnuWin32
|
https://en.wikipedia.org/wiki/Comparison%20triangle
|
Define as the 2-dimensional metric space of constant curvature . So, for example, is the Euclidean plane, is the surface of the unit sphere, and is the hyperbolic plane.
Let be a metric space. Let be a triangle in , with vertices , and . A comparison triangle in for is a triangle in with vertices , and such that , and .
Such a triangle is unique up to isometry.
The interior angle of at is called the comparison angle between and at . This is well-defined provided and are both distinct from .
|
https://en.wikipedia.org/wiki/All-pass%20filter
|
An all-pass filter is a signal processing filter that passes all frequencies equally in gain, but changes the phase relationship among various frequencies. Most types of filter reduce the amplitude (i.e. the magnitude) of the signal applied to it for some values of frequency, whereas the all-pass filter allows all frequencies through without changes in level.
Common applications
A common application in electronic music production is in the design of an effects unit known as a "phaser", where a number of all-pass filters are connected in sequence and the output mixed with the raw signal.
It does this by varying its phase shift as a function of frequency. Generally, the filter is described by the frequency at which the phase shift crosses 90° (i.e., when the input and output signals go into quadrature – when there is a quarter wavelength of delay between them).
They are generally used to compensate for other undesired phase shifts that arise in the system, or for mixing with an unshifted version of the original to implement a notch comb filter.
They may also be used to convert a mixed phase filter into a minimum phase filter with an equivalent magnitude response or an unstable filter into a stable filter with an equivalent magnitude response.
Active analog implementation
Implementation using low-pass filter
The operational amplifier circuit shown in adjacent figure implements a single-pole active all-pass filter that features a low-pass filter at the non-inverting input of the opamp. The filter's transfer function is given by:
which has one pole at -1/RC and one zero at 1/RC (i.e., they are reflections of each other across the imaginary axis of the complex plane). The magnitude and phase of H(iω) for some angular frequency ω are
The filter has unity-gain magnitude for all ω. The filter introduces a different delay at each frequency and reaches input-to-output quadrature at ω=1/RC (i.e., phase shift is 90°).
This implementation uses a low-pass filter at the
|
https://en.wikipedia.org/wiki/Stanford%20Web%20Credibility%20Project
|
The Stanford Web Credibility Project, which involves assessments of website credibility conducted by the Stanford University Persuasive Technology Lab, is an investigative examination of what leads people to believe in the veracity of content found on the Web. The goal of the project is to enhance website design and to promote further research on the credibility of Web resources.
Origins
The Web has become an important channel for exchanging information and services, resulting in a greater need for methods to ascertain the credibility of websites. In response, since 1998, the Stanford Persuasive Technology Lab (SPTL) has investigated what causes people to believe, or not, what they find online. SPTL provides insight into how computers can be designed to change what people think and do, an area called captology. Directed by experimental psychologist B.J. Fogg, the Stanford team includes social scientists, designers, and technologists who research and design interactive products that motivate and influence their users.
Objectives
The ongoing research of the Stanford Web Credibility Project includes:
Performing quantitative research on Web credibility
Collecting all public information on Web credibility
Acting as a clearinghouse for this information
Facilitating research and discussion about Web credibility
Collaborating with academic and industry research groups
How Do People Evaluate a Web Site's Credibility?
A study by the Stanford Web Credibility Project, How Do People Evaluate a Web Site's Credibility? Results from a Large Study, published in 2002, invited 2,684 "average people" to rate the credibility of websites in ten content areas. The study evaluated the credibility of two live websites randomly assigned from one of ten content categories: e-commerce, entertainment, finance, health, news, nonprofit, opinion or review, search engines, sports, and travel. A total of one hundred sites were assessed.
This study was launched jointly with a para
|
https://en.wikipedia.org/wiki/Outline%20of%20algebraic%20structures
|
In mathematics, there are many types of algebraic structures which are studied. Abstract algebra is primarily the study of specific algebraic structures and their properties. Algebraic structures may be viewed in different ways, however the common starting point of algebra texts is that an algebraic object incorporates one or more sets with one or more binary operations or unary operations satisfying a collection of axioms.
Another branch of mathematics known as universal algebra studies algebraic structures in general. From the universal algebra viewpoint, most structures can be divided into varieties and quasivarieties depending on the axioms used. Some axiomatic formal systems that are neither varieties nor quasivarieties, called nonvarieties, are sometimes included among the algebraic structures by tradition.
Concrete examples of each structure will be found in the articles listed.
Algebraic structures are so numerous today that this article will inevitably be incomplete. In addition to this, there are sometimes multiple names for the same structure, and sometimes one name will be defined by disagreeing axioms by different authors. Most structures appearing on this page will be common ones which most authors agree on. Other web lists of algebraic structures, organized more or less alphabetically, include Jipsen and PlanetMath. These lists mention many structures not included below, and may present more information about some structures than is presented here.
Study of algebraic structures
Algebraic structures appear in most branches of mathematics, and one can encounter them in many different ways.
Beginning study: In American universities, groups, vector spaces and fields are generally the first structures encountered in subjects such as linear algebra. They are usually introduced as sets with certain axioms.
Advanced study:
Abstract algebra studies properties of specific algebraic structures.
Universal algebra studies algebraic structures abstractly, r
|
https://en.wikipedia.org/wiki/Buffalo%20network-attached%20storage%20series
|
The Buffalo TeraStation network-attached storage series are network-attached storage devices.
The current lineup includes the LinkStation and TeraStation series. These devices have undergone various improvements since they were first produced, and have expanded to include a Windows Storage Server-based operating system.
History
Buffalo released the first TeraStation model, the HD-HTGL/R5, in December 2004. The second generation models, the TS-TGL/R5, was released the following year with uninterrupted operation and improved operational stability. This was followed up with the TeraStation Pro and the TeraStation Pro II in 2006, which offered iSCSI support, as well as 2U rackmount models. in 2008, the fourth generation TS-X models were released with hot swapping and replication, along with IU rackmount versions.
TeraStation
The TeraStation is a network-attached storage device using a PowerPC or ARM architecture processor. Many TeraStation models are shipped with enterprise-grade internal hard drives mounted in a RAID array. Since January 2012, the TeraStation uses LIO for its iSCSI target.
LinkStation
The LinkStation is a network-attached storage device using a PowerPC or ARM architecture processor designed for personal use, aiming to serve as a central media hub and backup storage for a household. Compared to the TeraStation series, LinkStation devices typically offer more streamlined UI and media server features.
Current Product Lineup
LinkStation
The LinkStation is notable among the Linux community both in Japan and in the US/Europe for being "hackable" into a generic Linux appliance and made to do tasks other than the file storage and sharing tasks for which it was designed. As the device runs on Linux, and included changes to the Linux source code, Buffalo was required to release their modified versions of source code as per the terms of the GNU General Public License. Due to the availability of source code and the relatively low cost of the device, there
|
https://en.wikipedia.org/wiki/Fragment%20molecular%20orbital
|
The fragment molecular orbital method (FMO) is a computational method that can be used to calculate very large molecular systems with thousands of atoms using ab initio quantum-chemical wave functions.
History of FMO and related methods
The fragment molecular orbital method (FMO) was developed by Kazuo Kitaura and coworkers in 1999. FMO is deeply interconnected with the energy decomposition analysis (EDA) by Kazuo Kitaura and Keiji Morokuma, developed in 1976. The main use of FMO is to compute very large molecular systems by dividing them into fragments and performing ab initio or density functional quantum-mechanical calculations of fragments and their dimers, whereby the Coulomb field from the whole system is included. The latter feature allows fragment calculations without using caps.
The mutually consistent field (MCF) method had introduced the idea of self-consistent fragment calculations in their embedding potential, which was later used with some modifications in various methods including FMO. There had been other methods related to FMO including the incremental correlation method by H. Stoll (1992).
Later, other methods closely related to FMO were proposed including the kernel energy method of L. Huang and the electrostatically embedded many-body expansion by E. Dahlke,
S. Hirata and later M. Kamiya suggested approaches also very closely related to FMO. Effective fragment molecular orbital (EFMO) method combines some features of the effective fragment potentials (EFP) and FMO. A detailed perspective on the fragment-based method development can be found in a review.
Introduction to FMO
In addition to the calculation of the total properties, such as the energy,
energy gradient, dipole moment etc., an interaction energy is obtained for
each pair of fragments. This pair interaction energy can be further
decomposed into electrostatic, exchange, charge transfer and dispersion
contributions. This analysis is known as the pair interaction energy
decompositi
|
https://en.wikipedia.org/wiki/Austin%20Hobart%20Clark
|
Austin Hobart Clark (December 17, 1880 – October 28, 1954) was an American zoologist. He was born in Wellesley, Massachusetts and died in Washington, D.C. His research covered a wide range of topics including oceanography, marine biology, ornithology, and entomology.
Biography
The son of Theodore Minot Clark and Jeannette French Clark, Clark obtained his Bachelor of Arts at Harvard University in 1903. He had five children with his first wife Mary Wendell Upham, whom he married on March 6, 1906. Mary died in December 1931 and Clark was remarried in 1933 to Leila Gay Forbes.
In 1901, Clark organized a scientific expedition to Isla Margarita in Venezuela. From 1903 to 1905, he conducted research in the Antilles. From 1906 to 1907, he led a scientific team aboard the 1882 USS Albatross. In 1908, he took a post at the National Museum of Natural History, which he held until his retirement in 1950.
Clark had important and various roles in a number of learned societies: he was president of the Entomological Society of Washington, vice president of the American Geophysical Union, and directed the press service of the American Association for the Advancement of Science.
Clark was author to more than 600 publications written in English, French, Italian, German, and Russian. Some of the most well-known include Animals of Land and Sea (1925), Nature Narratives (two volumes, 1929 and 1931), The New Evolution (1930), and Animals Alive (1948).
Several animal species and genera were first scientifically described by Clark, including the Lesser Antillean macaw (1905), the Martinique parrot (1905), the Dominican green-and-yellow macaw (1908), the mulga parrot (1910), the crustacean genus Laomenes (1919) or the starfish species Copidaster lymani (1948).
Zoogenesis
Clark is best known for his evolutionary theory called zoogenesis, which he introduced in his book The New Evolution: Zoogenesis (1930). His theory challenged the single tree view of evolution, according to Clark the
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.