source
stringlengths
31
227
text
stringlengths
9
2k
https://en.wikipedia.org/wiki/Vapor%20polishing
Vapor polishing is a method of polishing plastics to reduce the surface roughness or improve clarity. Typically, a component is exposed to a chemical vapor causing the surface to flow thereby improving the surface finish. This method of polishing is frequently used to return clear materials to an optical quality finish after machining. Vapor polishing works well in the internal features of components. Feature size changes of the plastic component generally do not occur. Post stress relieving is usually required as vapor polishing sets up surface stresses that can cause crazing. Plastics that respond well to vapor polishing are polycarbonate, acrylic, polysulfone, PEI, and ABS. The technique is also being used to improve the surface of objects created with 3D printing techniques. As the printer deposits layer upon layer of material to build the object, the surface is often not entirely smooth. The smoothness of the surface can be greatly increased by vapor polishing. Notes Materials science
https://en.wikipedia.org/wiki/Nvidia%20Quadro%20Plex
The Nvidia Quadro Plex is an external graphics processing unit (Visual Computing System) designed for large-scale 3D visualizations. The system consists of a box containing a pair of high-end Nvidia Quadro graphics cards featuring a variety of external video connectors. A special PCI Express card is installed in the host computer, and the two are connected by VHDCI cables. Specifications The Nvidia Quadro Plex system supports up to four GPUs per unit. It connects to the host PC via a small form factor PCI Express card connected to the host, and a Nvidia Quadro Plex Interconnect Cable. The system is housed in an external case that is approximately 9.49 inches in height, 5.94 inches in width, and 20.55 inches in depth and weighs about 19 pounds. The system relies heavily on Nvidia's SLI technology. Targeted audiences The Plex is aimed at large CGI animation companies, such as Pixar and DreamWorks Animation. This product is one of several professional graphics solutions on the market today, along with ATI's FireGL and Matrox's professional graphics cards. See also Sun Visualization System - uses Nvidia Quadro Plex for 3D rendering and graphics acceleration SGI Virtu VS product line - supports Quadro Plex External links Nvidia.com Quadro Plex VCS Quadro Plex Comparison Chart RTX 3090 - NVIDIA GeForce Graphics Third party information on the Quadro Plex: Gizmodo: Nvidia Announces Quadro Plex, Monster Graphics for Pros Graphics hardware Quadro Plex Visual effects Multi-monitor
https://en.wikipedia.org/wiki/Intelligence%3A%20Knowns%20and%20Unknowns
Intelligence: Knowns and Unknowns is a report issued in 1995 by a task force created by the Board of Scientific Affairs of the American Psychological Association (APA). It was subsequently published in the February 1996 issue of the peer-reviewed journal American Psychologist. Background The Board of Scientific Affairs (BSA) of the APA had concluded that after the publication of The Bell Curve (1994) and the following debate that there were "serious misunderstandings" and "that there was urgent need for an authoritative report on these issues—one that all sides could use as a basis for discussion". Furthermore, "Another unfortunate aspect of the debate was that many participants made little effort to distinguish scientific issues from political ones, Research findings were often assessed not so much on their merits or their scientific standing as on their supposed political implications." The report stated that "The charge to our Task Force was to prepare a dispassionate survey of the state of the art: to make clear what has been scientifically established, what is presently in dispute, and what is still unknown. In fulfilling that charge, the only recommendations we shall make are for further research and calmer debate." Published in August 1995, the report was authored by a task force of 11 experts. The APA Board on the Advancement of Psychology in the Public Interest (BAPPI) nominated one member of the Task Force. The Committee on Psychological Tests and Assessment nominated another. A third was nominated by the APA Council of Representatives. The other members were chosen by an extended consultative process with the aim of representing a broad range of expertise and opinion. Ulric Neisser was appointed chair. Three of the experts were also among the 52 signatories to "Mainstream Science on Intelligence", an editorial published in 1994. Members of BSA and BAPPI were asked to comment on a preliminary draft of the report. The entire Task Force gave unanimous su
https://en.wikipedia.org/wiki/Rare%20Breeds%20Survival%20Trust
The Rare Breeds Survival Trust is a conservation charity whose purpose is to secure the continued existence and viability of the native farm animal genetic resources (FAnGR) of the United Kingdom. It was founded in 1973 by Joe Henson to preserve native breeds; since then, no UK-native breed has become extinct. It maintains a watch list of rare native breeds of cattle, sheep, pigs, horses, goats and poultry, and an approved list of farm parks. Projects have included the collection of genetic material to ensure the future of rare breeds in a farm animal "gene bank". This project received publicity in the wake of the foot-and-mouth disease crisis in the UK and was supported by Charles III (while he was the Prince of Wales). The Trust also supports the Manx Ark Project, which provides sanctuary for rare breeds at several sites in the Isle of Man. See also Genetic diversity Similar organizations The Livestock Conservancy Ark of Taste Cobthorn Trust Gesellschaft zur Erhaltung alter und gefährdeter Haustierrassen
https://en.wikipedia.org/wiki/Battle%20Cross%20%281982%20video%20game%29
Battle Cross is a shoot 'em up released in arcades by Omori Electric (also known as Omori Electronic) in 1982. The player controls a fighter spacecraft by moving around the screen and shooting enemies. In 1984 Sony published a version for the Sony Hit-Bit 75 MSX computer. Gameplay
https://en.wikipedia.org/wiki/Maryam%20Mirzakhani
Maryam Mirzakhani (, ; 12 May 1977 – 14 July 2017) was an Iranian mathematician and a professor of mathematics at Stanford University. Her research topics included Teichmüller theory, hyperbolic geometry, ergodic theory, and symplectic geometry. On 13 August 2014, Mirzakhani was honored with the Fields Medal, the most prestigious award in mathematics, becoming the first woman to win the prize, as well as the first Iranian. The award committee cited her work in "the dynamics and geometry of Riemann surfaces and their moduli spaces". On 14 July 2017, Mirzakhani died of breast cancer at the age of 40. Early life and education Mirzakhani was born on 12 May 1977 in Tehran, Iran. As a child, she attended Tehran Farzanegan School, part of the National Organization for Development of Exceptional Talents (NODET). In her junior and senior years of high school, she won the gold medal for mathematics in the Iranian National Olympiad, thus allowing her to bypass the national college entrance exam. In 1994, Mirzakhani became the first Iranian woman to win a gold medal at the International Mathematical Olympiad in Hong Kong, scoring 41 out of 42 points. The following year, in Toronto, she became the first Iranian to achieve the full score and to win two gold medals in the International Mathematical Olympiad. Later in her life, she collaborated with friend, colleague, and Olympiad silver medalist, Roya Beheshti Zavareh (), on their book 'Elementary Number Theory, Challenging Problems', (in Persian) which was published in 1999. Mirzakhani and Zavareh together were the first women to compete in the Iranian National Mathematical Olympiad and won gold and silver medals in 1995, respectively. On 17 March 1998, after attending a conference consisting of gifted individuals and former Olympiad competitors, Mirzakhani and Zavareh, along with other attendees, boarded a bus in Ahvaz en route to Tehran. The bus was involved in an accident wherein it fell off a cliff, killing seven of the p
https://en.wikipedia.org/wiki/Logarithmic%20convolution
In mathematics, the scale convolution of two functions and , also known as their logarithmic convolution is defined as the function when this quantity exists. Results The logarithmic convolution can be related to the ordinary convolution by changing the variable from to : Define and and let , then Logarithms
https://en.wikipedia.org/wiki/ISO%2013485
ISO 13485 Medical devices -- Quality management systems -- Requirements for regulatory purposes is a voluntary standard, published by International Organization for Standardization (ISO) for the first time in 1996, and contains a comprehensive quality management system for the design and manufacture of medical devices. The latest version of this standard supersedes earlier documents such as EN 46001 (1993 and 1996) and EN 46002 (1996), the previously published ISO 13485 (1996 and 2003), and ISO 13488 (also 1996). The current ISO 13485 edition was published on 1 March 2016. Background Though it is tailored to the industry's quality system expectations and regulatory requirements, an organization does not need to be actively manufacturing medical devices or their components to seek certification to this standard, in contrast to the automotive sector's ISO/TS 16949, where only firms with an active request for quotation, or on the bid list, of an International Automotive Task Force supply chain manufacturer can seek registration. Reason for use While it remains a stand-alone document, ISO 13485 is generally harmonized with ISO 9001. A principal difference, however, is that ISO 9001 requires the organization to demonstrate continual improvement, whereas ISO 13485 requires only that the certified organization demonstrate the quality system is effectively implemented and maintained. Additionally, the ISO 9001 requirements regarding customer satisfaction are absent from the medical device standard. Other specific differences include: the promotion and awareness of regulatory requirements as a management responsibility. Examples of market-specific regulatory requirements include 21 CFR 820, the Quality System Regulation for medical devices sold in the United States, enforced by the U.S. Food and Drug Administration (FDA), or the Medical Devices Directive 93/42/EEC, required for doing business in the European Union controls in the work environment to ensure product
https://en.wikipedia.org/wiki/Unifacial%20cambium
The unifacial cambium (pl. cambia or cambiums) produces cells to the interior of its cylinder. These cells differentiate into xylem tissue. Unlike the more common bifacial cambium found in later woody plants, the unifacial cambium does not produce phloem to its exterior. Also in contrast to the bifacial cambium, the unifacial cambium is unable to expand its circumference with anticlinal cell division. Cell elongation provides a limited amount of expansion. Unifacial cambium plant morphology and life cycles The unifacial cambium allowed plants to grow as tall as 50 metres. Lacking secondary phloem, unifacial cambium plants developed alternative strategies to long range nutrient transport. For example, the stems of lycophyte trees were covered in photosynthesizing leaf bases. Due to the limited capacity for circumference growth, unifacial cambium plants had very little wood compared to modern woody plants. Xylem tissue in unifacial cambium plants was particularly structurally efficient. Additional structural support was provided in lycophytes by a special periderm tissue in the outer cortex. Lycophyte trees exhibit determinate growth. These trees appear to have lived for most of their life cycle as a 'stump', establishing root networks underground, before shooting up rapidly, releasing spores, and dying shortly thereafter. External links 'Key innovations, convergence, and success: macroevolutionary lessons from plant phylogeny', article by Michael J. Donahue from Paleobiology 31(2), 2005 (pdf) Devonian Times: More About Lycopsids Plant physiology Plant anatomy
https://en.wikipedia.org/wiki/Hydrogen%20sensor
A hydrogen sensor is a gas detector that detects the presence of hydrogen. They contain micro-fabricated point-contact hydrogen sensors and are used to locate hydrogen leaks. They are considered low-cost, compact, durable, and easy to maintain as compared to conventional gas detecting instruments. Key issues There are five key issues with hydrogen detectors: Reliability: Functionality should be easily verifiable. Performance: Detection 0.5% hydrogen in air or better Response time < 1 second. Lifetime: At least the time between scheduled maintenance. Cost: Goal is $5 per sensor and $30 per controller. Additional requirements Measurement range coverage of 0.1–10.0% concentration Operation in temperatures of −30 °C to 80 °C Accuracy within 5% of full scale Function in an ambient air gas environment within a 10–98% relative humidity range Resistance to hydrocarbon and other interference. Lifetime greater than 10 years Types of microsensors There are various types of hydrogen microsensors, which use different mechanisms to detect the gas. Palladium is used in many of these, because it selectively absorbs hydrogen gas and forms the compound palladium hydride. Palladium-based sensors have a strong temperature dependence which makes their response time too large at very low temperatures. Palladium sensors have to be protected against carbon monoxide, sulfur dioxide and hydrogen sulfide. Optical fibre hydrogen sensors Several types of optical fibre surface plasmon resonance (SPR) sensor are used for the point-contact detection of hydrogen: Fiber Bragg grating coated with a palladium layer – Detects the hydrogen by metal hindrance. Micromirror – With a palladium thin layer at the cleaved end, detecting changes in the backreflected light. Tapered fibre coated with palladium – Hydrogen changes the refractive index of the palladium, and consequently the amount of losses in the evanescent wave. Other types Electrochemical hydrogen sensor – low (ppm) levels of hy
https://en.wikipedia.org/wiki/5
5 (five) is a number, numeral and digit. It is the natural number, and cardinal number, following 4 and preceding 6, and is a prime number. It has garnered attention throughout history in part because distal extremities in humans typically contain five digits. Evolution of the Arabic digit The evolution of the modern Western digit for the numeral 5 cannot be traced back to the Indian system, as for the digits 1 to 4. The Kushana and Gupta empires in what is now India had among themselves several forms that bear no resemblance to the modern digit. The Nagari and Punjabi took these digits and all came up with forms that were similar to a lowercase "h" rotated 180°. The Ghubar Arabs transformed the digit in several ways, producing from that were more similar to the digits 4 or 3 than to 5. It was from those digits that Europeans finally came up with the modern 5. While the shape of the character for the digit 5 has an ascender in most modern typefaces, in typefaces with text figures the glyph usually has a descender, as, for example, in . On the seven-segment display of a calculator and digital clock, it is represented by five segments at four successive turns from top to bottom, rotating counterclockwise first, then clockwise, and vice-versa. It is one of three numbers, along with 4 and 6, where the number of segments matches the number. Mathematics Five is the third smallest prime number, and the second super-prime. It is the first safe prime, the first good prime, the first balanced prime, and the first of three known Wilson primes. Five is the second Fermat prime, the second Proth prime, and the third Mersenne prime exponent, as well as the third Catalan number and the third Sophie Germain prime. Notably, 5 is equal to the sum of the only consecutive primes 2 + 3 and it is the only number that is part of more than one pair of twin primes, (3, 5) and (5, 7). It also forms the first pair of sexy primes with 11, which is the fifth prime number and Heegner n
https://en.wikipedia.org/wiki/Dipole%20graph
In graph theory, a dipole graph, dipole, bond graph, or linkage, is a multigraph consisting of two vertices connected with a number of parallel edges. A dipole graph containing edges is called the dipole graph, and is denoted by . The dipole graph is dual to the cycle graph . The honeycomb as an abstract graph is the maximal abelian covering graph of the dipole graph , while the diamond crystal as an abstract graph is the maximal abelian covering graph of . Similarly to the Platonic graphs, the dipole graphs form the skeletons of the hosohedra. Their duals, the cycle graphs, form the skeletons of the dihedra.
https://en.wikipedia.org/wiki/Underwater%20explosion
An underwater explosion (also known as an UNDEX) is a chemical or nuclear explosion that occurs under the surface of a body of water. While useful in anti-ship and submarine warfare, underwater bombs are not as effective against coastal facilities. Properties of water Underwater explosions differ from in-air explosions due to the properties of water: Mass and incompressibility (all explosions) – water has a much higher density than air, which makes water harder to move (higher inertia). It is also relatively hard to compress (increase density) when under pressure in a low range (up to about 100 atmospheres). These two together make water an excellent conductor of shock waves from an explosion. Effect of neutron exposure on salt water (nuclear explosions only) – most underwater blast scenarios happen in seawater, not fresh or pure water. The water itself is not much affected by neutrons but salt is strongly affected. When exposed to neutron radiation during the microsecond of active detonation of a nuclear pit, water itself does not typically "activate", or become radioactive. The two elements in water, hydrogen and oxygen, can absorb an extra neutron, becoming deuterium and oxygen-17 respectively, both of which are stable isotopes. Even oxygen-18 is stable. Radioactive atoms can result if a hydrogen atom absorbs two neutrons, an oxygen atom absorbs three neutrons, or oxygen-16 undergoes a high energy neutron (n-p) reaction to produce a short-lived nitrogen-16. In any typical scenario, the probability of such multiple captures in significant numbers in the short time of active nuclear reactions around a bomb is very low. They are somewhat greater when the water is continuously irradiated, as in the closed-loop primary cooling system of a nuclear reactor. However, salt in seawater readily absorbs neutrons into both the sodium-23 and chlorine-35 atoms, which change to radioactive isotopes. Sodium-24 has a half-life of about 15 hours, while that of chlorine-36 (which
https://en.wikipedia.org/wiki/Post-orbital%20constriction
In physical anthropology, post-orbital constriction is the narrowing of the cranium (skull) just behind the eye sockets (the orbits, hence the name) found in most non-human primates and early hominins. This constriction is very noticeable in non-human primates, slightly less so in Australopithecines, even less in Homo erectus and completely disappears in modern Homo sapiens. Post-orbital constriction index in non-human primates and hominin range in category from increased constriction, intermediate, reduced constriction and disappearance. The post-orbital constriction index is defined by either a ratio of minimum frontal breadth (MFB), behind the supraorbital torus, divided by the maximum upper facial breadth (BFM), bifrontomalare temporale, or as the maximum width behind the orbit of the skull. Cranial Evolution Measurement of cranial capacity in hominis has been long used to examine the evolutionary development of increased brain size, allowing for comparing and contrasting among hominin skulls and between primates and hominins. Similarly, the post-orbital constriction index has become a form to compare and contrast craniums with the possibility of determining the relative age and evolutionary place of a new found hominin. Cranial capacity and post-orbital constriction index can demonstrate a correlation between increased brain size and reduced post-orbital constriction. For example, the average cranial capacity for Australopithecines is 440 cc, and the post-orbital constriction index is 0.66. However, with the evolutionary change in brain size in Australopithecines to the Homo genus, the average cranial capacity for Homo Habilis is 640 cc, and the post-orbital constriction index is 0.72. More specifically, in a departure from Homo erectus, Homo sapiens manifests a absolutely reduced post-orbital constriction (post-orbital constriction disappears) and a post-orbital constriction index of 0.92 due to increase in cranial capacity, about 1,350 cc. From the Austral
https://en.wikipedia.org/wiki/Chromosome%20abnormality
A chromosomal abnormality, chromosomal anomaly, chromosomal aberration, chromosomal mutation, or chromosomal disorder, is a missing, extra, or irregular portion of chromosomal DNA. These can occur in the form of numerical abnormalities, where there is an atypical number of chromosomes, or as structural abnormalities, where one or more individual chromosomes are altered. Chromosome mutation was formerly used in a strict sense to mean a change in a chromosomal segment, involving more than one gene. Chromosome anomalies usually occur when there is an error in cell division following meiosis or mitosis. Chromosome abnormalities may be detected or confirmed by comparing an individual's karyotype, or full set of chromosomes, to a typical karyotype for the species via genetic testing. Numerical abnormality An abnormal number of chromosomes is known as aneuploidy, and occurs when an individual is either missing a chromosome from a pair (resulting in monosomy) or has more than two chromosomes of a pair (trisomy, tetrasomy, etc.). Aneuploidy can be full, involving a whole chromosome missing or added, or partial, where only part of a chromosome is missing or added. Aneuploidy can occur with sex chromosomes or autosomes. An example of trisomy in humans is Down syndrome, which is a developmental disorder caused by an extra copy of chromosome 21; the disorder is therefore also called trisomy 21. An example of monosomy in humans is Turner syndrome, where the individual is born with only one sex chromosome, an X. Sperm aneuploidy Exposure of males to certain lifestyle, environmental and/or occupational hazards may increase the risk of aneuploid spermatozoa. In particular, risk of aneuploidy is increased by tobacco smoking, and occupational exposure to benzene, insecticides, and perfluorinated compounds. Increased aneuploidy is often associated with increased DNA damage in spermatozoa. Structural abnormalities When the chromosome's structure is altered, this can take several
https://en.wikipedia.org/wiki/Outline%20of%20transport
The following outline is provided as an overview of and topical guide to transport: Transport or transportation – movement of people and goods from one place to another. Essence of transport Driving involves controlling a vehicle, usually a motor vehicle such as a truck, bus, or automobile. For motorcycles, bicycles and animals, it is called riding. Shipping, transporting of goods and cargo, by land, sea, and air Travel, movement of people, by land, sea, and air Types of transport By availability Private transport Public transport (public transit) Modes and vehicles Intermodal passenger transport Aviation Aviation Fixed-wing aircraft Airship (dirigible) Autogyro Balloon Blimp Helicopter Human-powered aircraft Parachute (downward air transport only) Rocket Projectile (goods only, normally explosives) / Human cannonball Supersonic transport Zeppelin Animal-powered transport Animal-powered transport Animals domesticated for transport camel, Arabian, and Bactrian carabao deer dog sled dog Dogcart (dog-drawn) elephant equine donkey mule hinny horse pack horse draught horse riding horse coach horse llama moose ostrich ox reindeer sheep yak Turtles were used for riding as a sport in early 20th-century Australia Dolphins (to carry markers to attach to detected mines) Pigeon (for carrying messages) Animal-powered vehicles barge (sometimes pulled by humans) berlin (vehicle) Brougham (carriage) carriage cart chaise charabanc chariot (ancient form sometimes used in combat, later a racing machine, later a name for something entirely different in carriages) coach Conestoga wagon curricle dogcart dray ferry float gig governess cart Hansom cab horsecar horse-drawn boat horse-powered boat Experiment (horse-powered boat) howdah litter (vehicle) (sometimes carried by humans, mainly used with equines, though occasionally camels) mail coach Michigan logging wheels omnibus bullock cart pantechnicon van
https://en.wikipedia.org/wiki/Operation%20Cyber%20Condition%20Zebra
Operation Cyber Condition Zebra is a network operations campaign conducted by the United States Navy to deny network intrusion and establish an adequate computer network defense posture to provide defense-in-depth and warfighting capability. The operation specifies that perimeter security for legacy networks will deny intrusions and data infiltration, that firewalls will be maintained through risk assessment and formal adjudication of legacy application waiver requests, and that legacy networks will be shut down as quickly a possible after enterprise networks (such as the NMCI) are established. Its name is an analogue of the term "material condition Zebra," which is a standard configuration of equipment systems set on a warship to provide the greatest degree of subdivision and tightness to the ship. It is set immediately and automatically when general quarters is sounded.
https://en.wikipedia.org/wiki/Mimosine
Mimosine or leucenol is a toxic non-protein amino acid chemically similar to tyrosine. It occurs in some Mimosa spp. (including M. pudica) and all members of the closely related genus Leucaena. This compound, also known as leucenol, was first isolated from the seeds of Leucaena glauca Benth., and was later investigated by Adams and coworkers. Properties Mimosine melts with decomposition. The hydrochloride salt melts at 174.5–175.0 °C with decomposition; the hydrobromide decomposes at 179.5 °C, and the hydroiodide decomposes at 183–183.5 °C. Mimosine only forms monobasic acids, but the methyl ester forms a dihydrochloride, C7H9O2N2(COOMe)•2 HCl•½ H2O, mp. 175–6 °C. Biological effects Mimosine arrests dividing cells in the late G1 phase by inhibiting DNA replication initiation. In ruminants, mimosine is degraded to 3,4- and 2,3-dihydroxypyridone (3,4- and 2,3-DHP). Although toxicosis has occurred in Australia, Papua New Guinea, Africa and Florida, it has not been recorded in any other tropical and subtropical regions. Goats in Burma lost hair when fed a diet containing 50% of Leucaena. Goats and cattle in Hawaii are able to degrade the 3,4-DHP ruminally. Tolerance might be related to the presence or absence of microbes tolerant to mimosine and 3,4-DHP. It is known that at least Australian goats do not share the abilities of their Hawaiian counterparts. Bickel and Wibaut found in feeding experiments with rats and mice that leucenol is probably the toxic constituent of Leucaena glauca seeds, but they did not observe with these animals the loss of hair that seems to occur when these seeds are fed to cattle. Aung from Myanmar isolated the new subspecies of Klebsiella pneumoniae that can degrade mimosine. Besides, Moe Thida Htun, from Myanmar also found another new sub-species of Baccilus aureus that can degrade mimosine. Some rhizobia are known to produce rhizomimosinase, which consumes pyridoxal 5′-phosphate to degrade mimosine into 3,4 dihydroxypyridine, pyruvate
https://en.wikipedia.org/wiki/The%20One%20Minutes
The One Minutes is a global platform for one-minute videos. The One Minutes Foundation produces and distributes One Minutes, providing a platform for people to create and connect through short, accessible video art. History The One Minutes was initiated in 1998 by Katja van Stiphout and Michal Buttink, two students of the Sandberg Institute, Masters of Art and Design. The institute’s director Jos Houweling was asked to fill in an hour of airtime on local television, SALTO, once a month from midnight to 1 a.m. and offered this to two of his students. They invited fellow students and friends to fill the timeslot with one-minute films. A new format was born. Within the inexorable limitation of 60 seconds, the endless possibilities of video were revealed. The hour at midnight grew into a worldwide platform, where television channels, arts organisations and film festivals adopted segments of One Minutes, showcasing one-minutes at film festivals, art organisations and cultural institutes. In 1999, The One Minutes Foundation was founded, under direction of Jos Houweling and supported by Sandberg Institute. Since then, The One Minutes Awards have been held annually to acknowledge the best One Minutes of the year. The One Minutes held their first workshop in China in 2000 at Xiamen University. Ever since, The One Minutes has been a bridge of cultural exchange between international and Chinese artists, filmmakers and students. In 2008, Chinese artists took part in the Venice Biennale as part of The One Minutes. Since 2009, East China Normal University and Shanghai Dragon TV have been organising the yearly The One Minutes International Competition, which is broadcast by Dragon TV, Shanghai Media Group’s satellite broadcaster. One Minutes were also shown at EXPO Shanghai in 2010. Since 2011, The One Minutes has been a returning section in the annual Shanghai International TV Festival. In 2014, under the direction of Julia van Mourik, The One Minutes started a new curated pr
https://en.wikipedia.org/wiki/Nathan%20D%27Laryea
Nathan Amarkine D'Laryea (born 3 September 1985, in Manchester) is an English former footballer who played as a defender. He was a product of the Manchester City academy and played for clubs including Rochdale, Macclesfield Town and Hyde FC. Career Early career D'Laryea started out as a trainee at Manchester City. Before being released by City, he had a spell on loan at Macclesfield Town during the 2006–07 season. He joined Rochdale but made only eight appearances in two seasons and he quit professional football to study for a degree in English Language at the University of Manchester. Hyde In July 2009 D'Laryea signed for Conference North semi-professional club Hyde United. He made his debut for Hyde in the first game of the 2009–2010 season against Stafford Rangers. He scored his first Hyde goal in March 2010 in the Tameside derby against Stalybridge Celtic. He became a regular for Hyde at centre back until late in his first season when he acquired an injury that kept him out for the remainder of the season and the start of 2010–11 season. Personal life His twin brother, Jonathan D'Laryea, is also a professional footballer and currently plays for AFC Mansfield. The D'Laryea brothers are of Ghanaian descent. In 2012 D'Laryea was working as a secondary school teacher at Dixons Allerton Academy in Bradford and where he teaches maths. He recently moved to Manchester To work at Loreto High School, Chorlton. Referee In April 2011, it was announced that he had commenced a post-playing career as referee for Leisure Leagues 6-a-side football in Oldham.
https://en.wikipedia.org/wiki/Control%20track%20longitudinal%20timecode
Control track longitudinal timecode, or CTL timecode, developed by JVC in the early 1990s, is a unique technique for embedding, or striping, reference SMPTE timecode onto a videotape. Similar to the way VITC timecode is embedded in the vertical interval area of a video signal, CTL timecode embeds SMPTE timecode in the control track area of helical scan video recordings. The advantage of both VITC and CTL timecode is that an audio track does not have to be sacrificed for linear timecode. Though a very effective technology, and still probably in limited use today, CTL timecode never really caught on. JVC is apparently the only manufacturer that included CTL timecode capability in their video products, and this was limited to select professional S-VHS equipment. When it was introduced, there was much negativity about CTL timecode, because people misunderstood how it worked. Many incorrectly assumed that CTL timecode was nothing more than a control track pulse signal. Control Track Pulse: Most are familiar with the digital "counters" on VHS recorders and camcorders, viewed via the onscreen display (OSD) and/or a dedicated LED display. These numbers are sometimes in real-time format (hours:minutes:seconds), but are often only an ambiguous 4 digit sequential counter. These numbers advance up or down based on the machine counting a tape's control track pulses. This type of display is useful only as a simple and temporary reference, as it is very inaccurate, and the counter is reset to zero when a tape is inserted. A basic 4 digit counter is almost completely worthless, as their rate of advance was never standardized by manufacturers. Conversely, CTL timecode is an absolute timecode with specific digital references for every frame of video. Thus, a tape with CTL timecode can always display current timecode position accurately, even if the tape is moved from one machine to another. CTL timecode embedding can be transferred when making a copy, as long as the recording
https://en.wikipedia.org/wiki/Contemporary%20Physics%20Education%20Project
The Contemporary Physics Education Project (CPEP) is an "organization of teachers, educators, and physicists" formed in 1987. The group grew out of the Conference on the Teaching of Modern Physics held at Fermilab in 1986, organized by the American Association of Physics Teachers. The group's first effort aimed to supply a chart for particle physics teaching that would rival the Periodic Table of the elements. The first version of this chart was published in 1989. CPEP has created five charts emphasizing contemporary aspects of physics research: particles and interactions; fusion and plasma physics; nuclear science; and cosmology; and gravity.. Almost half a million of these charts and similar products have been distributed. The group has created website support for teaching for each of the charts. CPEP received the 2017 "Excellence in Physics Education Award" from the American Physical Society, "for leadership in providing educational materials on contemporary physics topics to students for over 25 years." Offshoots of CPEP include the book, "The Charm of Strange Quarks: Mysteries and Revolutions of Particle Physics" (2000), by R. Michael Barnett, Henry Muehry, and Helen R. Quinn, three of the founders of CPEP. See also the web site "The Particle Adventure: The Fundamentals of Matter and Force". R. Michael Barnett described the formation and early days of CPEP in a Nobel Symposium Lecture in 2002.
https://en.wikipedia.org/wiki/Petrophysics
Petrophysics (from the Greek πέτρα, petra, "rock" and φύσις, physis, "nature") is the study of physical and chemical rock properties and their interactions with fluids. A major application of petrophysics is in studying reservoirs for the hydrocarbon industry. Petrophysicists work together with reservoir engineers and geoscientists to understand the porous media properties of the reservoir. Particularly how the pores are interconnected in the subsurface, controlling the accumulation and migration of hydrocarbons. Some fundamental petrophysical properties determined are lithology, porosity, water saturation, permeability, and capillary pressure. The petrophysicists workflow measures and evaluates these petrophysical properties through well-log interpretation (i.e. in-situ reservoir conditions) and core analysis in the laboratory. During well perforation, different well-log tools are used to measure the petrophysical and mineralogical properties through radioactivity and seismic technologies in the borehole. In addition, core plugs are taken from the well as sidewall core or whole core samples. These studies are combined with geological, geophysical, and reservoir engineering studies to model the reservoir and determine its economic feasibility. While most petrophysicists work in the hydrocarbon industry, some also work in the mining, water resources, geothermal energy, and carbon capture and storage industries. Petrophysics is part of the geosciences, and its studies are used by petroleum engineering, geology, geochemistry, exploration geophysics and others. Fundamental petrophysical properties The following are the fundamental petrophysical properties used to characterize a reservoir: Lithology: A description of the rock's physical characteristics, such as grain size, composition and texture. By studying the lithology of local geological outcrops and core samples, geoscientists can use a combination of log measurements, such as natural gamma, neutron, densi
https://en.wikipedia.org/wiki/Sinusoidal%20plane-wave%20solutions%20of%20the%20electromagnetic%20wave%20equation
Sinusoidal plane-wave solutions are particular solutions to the electromagnetic wave equation. The general solution of the electromagnetic wave equation in homogeneous, linear, time-independent media can be written as a linear superposition of plane-waves of different frequencies and polarizations. The treatment in this article is classical but, because of the generality of Maxwell's equations for electrodynamics, the treatment can be converted into the quantum mechanical treatment with only a reinterpretation of classical quantities (aside from the quantum mechanical treatment needed for charge and current densities). The reinterpretation is based on the theories of Max Planck and the interpretations by Albert Einstein of those theories and of other experiments. The quantum generalization of the classical treatment can be found in the articles on photon polarization and photon dynamics in the double-slit experiment. Explanation Experimentally, every light signal can be decomposed into a spectrum of frequencies and wavelengths associated with sinusoidal solutions of the wave equation. Polarizing filters can be used to decompose light into its various polarization components. The polarization components can be linear, circular or elliptical. Plane waves The plane sinusoidal solution for an electromagnetic wave traveling in the z direction is for the electric field and for the magnetic field, where k is the wavenumber, is the angular frequency of the wave, and is the speed of light. The hats on the vectors indicate unit vectors in the x, y, and z directions. is the position vector (in meters). The plane wave is parameterized by the amplitudes and phases where and Polarization state vector Jones vector All the polarization information can be reduced to a single vector, called the Jones vector, in the x-y plane. This vector, while arising from a purely classical treatment of polarization, can be interpreted as a quantum state vector. The connection
https://en.wikipedia.org/wiki/Ovoid%20%28polar%20space%29
In mathematics, an ovoid O of a (finite) polar space of rank r is a set of points, such that every subspace of rank intersects O in exactly one point. Cases Symplectic polar space An ovoid of (a symplectic polar space of rank n) would contain points. However it only has an ovoid if and only and q is even. In that case, when the polar space is embedded into the classical way, it is also an ovoid in the projective geometry sense. Hermitian polar space Ovoids of and would contain points. Hyperbolic quadrics An ovoid of a hyperbolic quadricwould contain points. Parabolic quadrics An ovoid of a parabolic quadric would contain points. For , it is easy to see to obtain an ovoid by cutting the parabolic quadric with a hyperplane, such that the intersection is an elliptic quadric. The intersection is an ovoid. If q is even, is isomorphic (as polar space) with , and thus due to the above, it has no ovoid for . Elliptic quadrics An ovoid of an elliptic quadric would contain points. See also Ovoid (projective geometry)
https://en.wikipedia.org/wiki/Phage%20ecology
Bacteriophages (phages), potentially the most numerous "organisms" on Earth, are the viruses of bacteria (more generally, of prokaryotes). Phage ecology is the study of the interaction of bacteriophages with their environments. Introduction to phage ecology Vastness of phage ecology Phages are obligate intracellular parasites meaning that they are able to reproduce only while infecting bacteria. Phages therefore are found only within environments that contain bacteria. Most environments contain bacteria, including our own bodies (called normal flora). Often these bacteria are found in large numbers. As a consequence, phages are found almost everywhere. As a rule of thumb, many phage biologists expect that phage population densities will exceed bacterial densities by a ratio of 10-to-1 or more (VBR or virus-to-bacterium ratio; see for a summary of actual data). As there exist estimates of bacterial numbers on Earth of approximately 1030, there consequently is an expectation that 1031 or more individual virus (mostly phage) particles exist , making phages the most numerous category of "organisms" on our planet. Bacteria (along with archaea) appear to be highly diverse and there possibly are millions of species. Phage-ecological interactions therefore are quantitatively vast: huge numbers of interactions. Phage-ecological interactions are also qualitatively diverse: There are huge numbers of environment types, bacterial-host types, and also individual phage types Studying phage ecology The study of phage ecology reflects established scientific disciplines in ecological studies in scope, the most obvious being general ecology. Accordingly, phage ecology is treated under the following heads "organismal" ecology, population ecology, community ecology, and ecosystem ecology. Phage ecology also may be considered (though mostly less well formally explored) from perspectives of phage behavioral ecology, evolutionary ecology, functional ecology, landscape ecology, mat
https://en.wikipedia.org/wiki/Off-air%20pickup
An off-air pickup or off-the-air pickup is a method by which the direct signal from a radio or television station is received and then rebroadcast by another station, CATV system, or satellite television feed. It was often used to distribute network broadcast programming to smaller markets which were normally outside of the range of major centers. Use in United States television In the 1950s, a broadband network of AT&T L-carrier and microwave relays known as Long Lines was constructed. These circuits could be (and were) used for normal telephone traffic, but were also used to relay the video signals of the three US commercial television networks (ABC, CBS, and NBC) to their various affiliated stations around the country. This enabled those affiliates to all carry network programming "live" at the same time. However, a handful of small-market TV stations were not able to be connected via long lines due to their remote geographic location. In some cases, network shows were recorded (either as a kinescope or, starting in the late 1950s and early 1960s, on videotape) and sent to the station via mail or courier, to be aired on a delayed basis (usually a one-week delay). This was a less than satisfactory situation for these stations because of the inferior quality of kinescope recordings, the possibility of a recording being lost or damaged in transit, and the fact that they would be unable to broadcast episodes of a given series on the same date and at the same time as most other network affiliates. Additionally, live broadcasts such as sporting events and breaking news coverage could not be carried, as they would be dated by the time a recording could be received and broadcast. Some of these stations, therefore, entered into an agreement with another affiliate of the same network located in a nearby market to pick up their broadcast signal off the air and relay it via their own transmitter during network programming. In most cases, this was accomplished by simply
https://en.wikipedia.org/wiki/Hospital-acquired%20pneumonia
Hospital-acquired pneumonia (HAP) or nosocomial pneumonia refers to any pneumonia contracted by a patient in a hospital at least 48–72 hours after being admitted. It is thus distinguished from community-acquired pneumonia. It is usually caused by a bacterial infection, rather than a virus. Hospital acquired pneumonia is the second most common nosocomial infection (after urinary tract infections) and accounts for 15–20% of the total. It is the most common cause of death among nosocomial infections and is the primary cause of death in intensive care units. It is also one of the most common infections acquired at the hospital in children around the world. Hospital acquired pneumonia typically lengthens a hospital stay by 1–2 weeks. Signs and symptoms New or progressive infiltrate on the chest X-ray with one of the following: Fever > 37.8 °C (100 °F) Purulent sputum Leukocytosis > 10,000 cells/μl In an elderly person, the first sign of hospital-acquired pneumonia may be mental changes or confusion. Other symptoms may include: A cough with greenish or pus-like phlegm (sputum) Fever and chills General discomfort, uneasiness, or ill feeling (malaise) Loss of appetite Nausea and vomiting Sharp chest pain that gets worse with deep breathing or coughing Shortness of breath Decreased blood pressure and fast heart rate Types Bacterial pneumonia: The majority of cases related to various rod shaped gram-negative organisms (52%) and Staphylococcus aureus (19%), usually of the MRSA type. Others are Haemophilus spp. (5%). In the ICU results were S. aureus (17.4%), Pseudomonas aeruginosa (17.4%), Klebsiella pneumoniae and Enterobacter spp. (18.1%), and Haemophilus influenzae (4.9%). Viral pneumonia: influenza and respiratory syncytial virus and, in the immunocompromised host, cytomegalovirus – cause 10–20% of infections. Ventilator-associated pneumonia Ventilator-associated pneumonia (VAP) is a sub-type of hospital-acquired pneumonia (HAP) which occurs in people
https://en.wikipedia.org/wiki/Photonic%20integrated%20circuit
A photonic integrated circuit (PIC) or integrated optical circuit is a microchip containing two or more photonic components which form a functioning circuit. This technology detects, generates, transports, and processes light. Photonic integrated circuits utilize photons (or particles of light) as opposed to electrons that are utilized by electronic integrated circuits. The major difference between the two is that a photonic integrated circuit provides functions for information signals imposed on optical wavelengths typically in the visible spectrum or near infrared (850–1650 nm). The most commercially utilized material platform for photonic integrated circuits is indium phosphide (InP), which allows for the integration of various optically active and passive functions on the same chip. Initial examples of photonic integrated circuits were simple 2-section distributed Bragg reflector (DBR) lasers, consisting of two independently controlled device sections – a gain section and a DBR mirror section. Consequently, all modern monolithic tunable lasers, widely tunable lasers, externally modulated lasers and transmitters, integrated receivers, etc. are examples of photonic integrated circuits. As of 2012, devices integrate hundreds of functions onto a single chip. Pioneering work in this arena was performed at Bell Laboratories. The most notable academic centers of excellence of photonic integrated circuits in InP are the University of California at Santa Barbara, USA, the Eindhoven University of Technology and the University of Twente in the Netherlands. A 2005 development showed that silicon can, even though it is an indirect bandgap material, still be used to generate laser light via the Raman nonlinearity. Such lasers are not electrically driven but optically driven and therefore still necessitate a further optical pump laser source. History Photonics is the science behind the detection, generation, and manipulation of photons. According to quantum mechanics and t
https://en.wikipedia.org/wiki/DNA%20clamp
A DNA clamp, also known as a sliding clamp, is a protein complex that serves as a processivity-promoting factor in DNA replication. As a critical component of the DNA polymerase III holoenzyme, the clamp protein binds DNA polymerase and prevents this enzyme from dissociating from the template DNA strand. The clamp-polymerase protein–protein interactions are stronger and more specific than the direct interactions between the polymerase and the template DNA strand; because one of the rate-limiting steps in the DNA synthesis reaction is the association of the polymerase with the DNA template, the presence of the sliding clamp dramatically increases the number of nucleotides that the polymerase can add to the growing strand per association event. The presence of the DNA clamp can increase the rate of DNA synthesis up to 1,000-fold compared with a nonprocessive polymerase. Structure The DNA clamp is an α+β protein that assembles into a multimeric, six-domain ring structure that completely encircles the DNA double helix as the polymerase adds nucleotides to the growing strand. Each domain is in turn made of two β-α-β-β-β structural repeats. The DNA clamp assembles on the DNA at the replication fork and "slides" along the DNA with the advancing polymerase, aided by a layer of water molecules in the central pore of the clamp between the DNA and the protein surface. Because of the toroidal shape of the assembled multimer, the clamp cannot dissociate from the template strand without also dissociating into monomers. The DNA clamp fold is found in bacteria, archaea, eukaryotes and some viruses. In bacteria, the sliding clamp is a homodimer composed of two identical beta subunits of DNA polymerase III and hence is referred to as the beta clamp. In archaea and eukaryotes, it is a trimer composed of three molecules of PCNA. The T4 bacteriophage also uses a sliding clamp, called gp45 that is a trimer similar in structure to PCNA but lacks sequence homology to either PCNA or
https://en.wikipedia.org/wiki/Australian%20Privacy%20Foundation
The Australian Privacy Foundation is an NGO formed for the purpose of protecting the privacy rights of Australians. Its aim is to focus public attention on emerging issues which pose a threat to the freedom and privacy of Australians, and also takes a leading role on issues of defending rights of individuals to control access to personal information and to be free of excessive intrusions. History The organisation was initially formed on 31 August 1987 at the Sebel Town House, Kings Cross, Sydney for the sole purpose of coordinating public resistance against the Australia Card, first proposed as part of the 1985 Federal Budget. The media were attracted to the launch on account of support for the movement by a number of high-profile persons including Peter Garrett, Alan Jones, Ben Lexcen, and Janine Haines. After the group led the charge in successfully defeating the Australia Card, it was clear to the group that an ongoing national voice for privacy protection was needed in Australia. During the remaining years of the 1980s, the Foundation's focus was to ensure that the tax file number (TFN) scheme did not become an ID Card in disguise, and to lobby both state and federal governments for the enactment of national privacy legislation. The first summit organised by the Foundation brought together officials from the Australian Taxation Office, politicians and privacy advocates together to examine the TFN proposals, and exposed many weaknesses in the proposals which the Government was subsequently forced to rectify. The organisation has remained strongly focussed on policy matters. It has made many submissions to Parliamentary Committees, government agencies, industry associations and corporations. It has run a number of campaigns, in relation to such matters as credit reporting, government proposals to create identification schemes, and privacy and the media. Since 2008, it has become more proactive, developing a series of Policy Statements on matters of particu
https://en.wikipedia.org/wiki/Palmar%20carpal%20arch
The palmar carpal arch is a joining of an artery to an artery, a circulatory anastomosis, known as an arterio-arterial anastomosis. The two connected arteries are the palmar carpal branch of the radial artery and the palmar carpal branch of the ulnar artery. This anastomosis is joined by a branch from the anterior interosseous artery above, and by recurrent branches from the deep palmar arch below, thus forming a palmar carpal network which supplies the articulations of the wrist and carpus. See also Dorsal carpal arch Deep palmar arch Superficial palmar arch
https://en.wikipedia.org/wiki/Disease%20informatics
Disease Informatics is the application of Information science by defining the diseases with the least error, identifying most of the targets to combat a cluster of diseases (Disease Causal Chain), and designing a holistic solution (Health strategy) to the problem.
https://en.wikipedia.org/wiki/List%20of%20commercial%20products%20based%20on%20Red%20Hat%20Enterprise%20Linux
There are a number of commercial products based on Red Hat Enterprise Linux (RHEL). Information about these products and the version of RHEL they are based on is often difficult to come by, since this fact is not widely publicised. Sometimes it is possible to run the 'uname -r' command to get the kernel release and then cross-reference it with the RHEL version history. Examples Asianux Asianux 1.0 is based on Red Hat Enterprise Linux 3.0. Asianux 2.0 is based on Red Hat Enterprise Linux 4.0. Autodesk Media and Entertainment Autodesk's Smoke, Flame and Lustre software all run on HP z800 machines pre-configured with a custom RHEL 5 distribution. An additional software package called the 'Discreet Kernel Utility' or DKU is added for additional proprietary device drivers and resources. Avaya Avaya's Communication Manager VoIP-PBX software is based on Red Hat Enterprise Linux 5. Amazon.com Amazon Linux available as the default Linux distribution on Amazon Web Services. F5 Networks BIG-IP The BIG-IP product line runs an operating system derived from Red Hat Enterprise Linux 6.5. Upgraded in version TMOS v.12. Check Point SecurePlatform Check Point SecurePlatform NG is based on Red Hat Linux 7.2 Check Point SecurePlatform NGX is based on Red Hat Enterprise Linux 3.0 Check Point SecurePlatform 2.6 has kernel based on Red Hat Enterprise Linux 5 and user space based on RHEL 3 Check Point SecurePlatform R70 is based on Red Hat Enterprise Linux 5. Check Point Gaia is based on Red Hat Enterprise Linux 5.2. Check Point SecurePlatform VSX R67 is based on Red Hat Enterprise Linux 5.2. Cisco Cisco Global Site Selector Cisco Unified Communications Manager Cisco Secure ACS (RADIUS and TACACS+ server) CloudLinux OS CloudLinux OS 5 is based on Red Hat Enterprise Linux 5.0 CloudLinux OS 6 is based on Red Hat Enterprise Linux 6.0 CloudLinux OS 7 is based on Red Hat Enterprise Linux 7.0 CloudLinux OS 8 is based on Red Hat Enterprise Linux 8.0 Crossbeam Sy
https://en.wikipedia.org/wiki/Trzebiatowski%20Institute%20of%20Low%20Temperature%20and%20Structure%20Research%20of%20the%20Polish%20Academy%20of%20Sciences
The Włodzimierz Trzebiatowski Institute of Low Temperature and Structure Research is a scientific institute in Wrocław, Poland. It is named after Włodzimierz Trzebiatowski, the Polish chemist, physicist and mathematician. Localization Until 1993 the Institute was placed at Gajowicka street and the Bishop Palace, that after the II World War was requisitioned by the Polish People's Republic government. Currently the Institute is placed at Okólna street in Wrocław. External links Institute webpage Scientific organisations based in Poland Buildings and structures in Wrocław Physics research institutes Institutes of the Polish Academy of Sciences
https://en.wikipedia.org/wiki/Photon%20polarization
Photon polarization is the quantum mechanical description of the classical polarized sinusoidal plane electromagnetic wave. An individual photon can be described as having right or left circular polarization, or a superposition of the two. Equivalently, a photon can be described as having horizontal or vertical linear polarization, or a superposition of the two. The description of photon polarization contains many of the physical concepts and much of the mathematical machinery of more involved quantum descriptions, such as the quantum mechanics of an electron in a potential well. Polarization is an example of a qubit degree of freedom, which forms a fundamental basis for an understanding of more complicated quantum phenomena. Much of the mathematical machinery of quantum mechanics, such as state vectors, probability amplitudes, unitary operators, and Hermitian operators, emerge naturally from the classical Maxwell's equations in the description. The quantum polarization state vector for the photon, for instance, is identical with the Jones vector, usually used to describe the polarization of a classical wave. Unitary operators emerge from the classical requirement of the conservation of energy of a classical wave propagating through lossless media that alter the polarization state of the wave. Hermitian operators then follow for infinitesimal transformations of a classical polarization state. Many of the implications of the mathematical machinery are easily verified experimentally. In fact, many of the experiments can be performed with polaroid sunglass lenses. The connection with quantum mechanics is made through the identification of a minimum packet size, called a photon, for energy in the electromagnetic field. The identification is based on the theories of Planck and the interpretation of those theories by Einstein. The correspondence principle then allows the identification of momentum and angular momentum (called spin), as well as energy, with the photon.
https://en.wikipedia.org/wiki/Sequencing%20by%20hybridization
Sequencing by hybridization is a class of methods for determining the order in which nucleotides occur on a strand of DNA. Typically used for looking for small changes relative to a known DNA sequence. The binding of one strand of DNA to its complementary strand in the DNA double-helix (known as hybridization) is sensitive to even single-base mismatches when the hybrid region is short or if specialized mismatch detection proteins are present. This is exploited in a variety of ways, most notably via DNA chips or microarrays with thousands to billions of synthetic oligonucleotides found in a genome of interest plus many known variations or even all possible single-base variations. The type of sequencing by hybridization described above has largely been displaced by other methods, including sequencing by synthesis, and sequencing by ligation (as well as pore-based methods). However hybridization of oligonucleotides is still used in some sequencing schemes, including hybridization-assisted pore-based sequencing, and reversible hybridization. Examples of commercial systems Affymetrix (true sequencing-by-hybridization) NABsys (Hybridization-assisted pore-based sequencing) Complete Genomics Inc. (reversible-hybridization of probes that call-out a single base with each hybridization) See also Sequencing by ligation
https://en.wikipedia.org/wiki/Thermal%20spraying
Thermal spraying techniques are coating processes in which melted (or heated) materials are sprayed onto a surface. The "feedstock" (coating precursor) is heated by electrical (plasma or arc) or chemical means (combustion flame). Thermal spraying can provide thick coatings (approx. thickness range is 20 microns to several mm, depending on the process and feedstock), over a large area at high deposition rate as compared to other coating processes such as electroplating, physical and chemical vapor deposition. Coating materials available for thermal spraying include metals, alloys, ceramics, plastics and composites. They are fed in powder or wire form, heated to a molten or semimolten state and accelerated towards substrates in the form of micrometer-size particles. Combustion or electrical arc discharge is usually used as the source of energy for thermal spraying. Resulting coatings are made by the accumulation of numerous sprayed particles. The surface may not heat up significantly, allowing the coating of flammable substances. Coating quality is usually assessed by measuring its porosity, oxide content, macro and micro-hardness, bond strength and surface roughness. Generally, the coating quality increases with increasing particle velocities. Variations Several variations of thermal spraying are distinguished: Plasma spraying Detonation spraying Wire arc spraying Flame spraying High velocity oxy-fuel coating spraying (HVOF) High velocity air fuel (HVAF) Warm spraying Cold spraying Spray and Fuse In classical (developed between 1910 and 1920) but still widely used processes such as flame spraying and wire arc spraying, the particle velocities are generally low (< 150 m/s), and raw materials must be molten to be deposited. Plasma spraying, developed in the 1970s, uses a high-temperature plasma jet generated by arc discharge with typical temperatures >15,000 K, which makes it possible to spray refractory materials such as oxides, molybdenum, etc. System overview
https://en.wikipedia.org/wiki/Master%20of%20Physics
A Master of Physics honours (or MPhys (Hons)) degree is a specific master's degree for courses in the field of physics. United Kingdom In England and Wales, the MPhys is an undergraduate award available after pursuing a four-year course of study at a university. In Scotland the course has a five-year duration. In some universities, the degree has the variant abbreviation MSci. These are taught courses, with a research element in the final year — this can vary from a small component to an entire year working with a research group — and are not available as postgraduate qualifications in most cases, although depending on institution the final year can be considered as approximately equivalent to an MSc. Structure In terms of course structure, MPhys degrees usually follow the pattern familiar from bachelor's degrees with lectures, laboratory work, coursework and exams each year. Usually one, or more commonly two, substantial projects are to be completed in the fourth year which may well have research elements. At the end of the second or third years, there is usually a threshold of academic performance in examinations to be reached to allow progression into the final year. Final results are, in most cases, awarded on the standard British undergraduate degree classification scale, although some universities award something structurally similar to 'Distinction', 'Merit', 'Pass' or 'Fail', as this is often the way that taught postgraduate master's degrees are classified. Degree schemes It is usual for there to be some variation in the MPhys schemes, to allow for students to study the area of physics which most interests them. For example, Lancaster University's physics department offer the following schemes: MPhys Physics MPhys Physics, Astrophysics and Cosmology MPhys Physics with Particle Physics and Cosmology MPhys Physics with Space Science MPhys Physics with Biomedical Physics MPhys Theoretical Physics MPhys Theoretical Physics with Mathematics These schemes will
https://en.wikipedia.org/wiki/Jardines%20de%20la%20Reina
Jardines de la Reina () is an archipelago in the southern part of Cuba, in the provinces of Camagüey and Ciego de Ávila. It was named by Christopher Columbus to honour the Queen of Spain, Isabella I of Castile. Since 1996 a marine reserve was established covering a large swath of the archipelago. In 2010, Jardines de la Reina was established as a national park (). With an area of , it is one of Cuba's largest protected areas. Geography It is located in the Caribbean Sea, between the Gulf of Ana Maria (north-west), Gulf of Guacanayabo (south) and Caballones Channel (west). It extends on a general north-west to south-east direction, paralleling the Cuban coast for from Cayo Breton to Cayos Mordazo. Cuba's second largest archipelago (smaller only than Jardines del Rey), it is formed by more than 600 cays and islands. Other cays in the archipelago include Caguamas, Cayos Cinco Balas, Cayo Anclitas, Cayo Algodon Grande, Cayos Pingues and Cayo Granada. Part of the archipelago is also known as Laberinto de las Doce Leguas (The Labyrinth of the Twelve Leagues) Table of Islands 1) The Islands area and population data retrieved from the 2012 census. Fishing The archipelago is a popular destination for diving and sport fly-fishing. Only catch and release fly-fishing and a limited, well-regulated lobster fishery is allowed in the park, although many other fisheries occur surrounding the park and close to cays out of the park limits. It used to be one of Fidel Castro's favorite fishing spots. Species of fish found here include Cubera snapper, Bonefish, Yellowfin grouper, Black grouper, Atlantic goliath grouper as well as Strombus gigas (the large Caribbean conch) and Whale shark. Besides being an extraordinary site for fly fishing, one its main attractions for diving is the abundance of reef sharks. Diving Jardines de la Reina (The Gardens of the Queen) is one of the most popular scuba destinations of Cuba. The underwater landscapes include canyons, pinnacles and caves.
https://en.wikipedia.org/wiki/Floor%20effect
In statistics, a floor effect (also known as a basement effect) arises when a data-gathering instrument has a lower limit to the data values it can reliably specify. This lower limit is known as the "floor". The "floor effect" is one type of scale attenuation effect; the other scale attenuation effect is the "ceiling effect". Floor effects are occasionally encountered in psychological testing, when a test designed to estimate some psychological trait has a minimum standard score that may not distinguish some test-takers who differ in their responses on the test item content. Giving preschool children an IQ test designed for adults would likely show many of the test-takers with scores near the lowest standard score for adult test-takers (IQ 40 on most tests that were currently normed as of 2010). To indicate differences in current intellectual functioning among young children, IQ tests specifically for young children are developed, on which many test-takers can score well above the floor score. An IQ test designed to help assess intellectually disabled persons might intentionally be designed with easier item content and a lower floor score to better distinguish among individuals taking the test as part of an assessment process. See also Ceiling effect (statistics)
https://en.wikipedia.org/wiki/Virtual%20knot
In knot theory, a virtual knot is a generalization of knots in 3-dimensional Euclidean space, , to knots in thickened surfaces modulo an equivalence relation called stabilization/destabilization. Here is required to be closed and oriented. Virtual knots were first introduced by . Overview In the theory of classical knots, knots can be considered equivalence classes of knot diagrams under the Reidemeister moves. Likewise a virtual knot can be considered an equivalence of virtual knot diagrams that are equivalent under generalized Reidemeister moves. Virtual knots allow for the existence of, for example, knots whose Gauss codes which could not exist in 3-dimensional Euclidean space. A virtual knot diagram is a 4-valent planar graph, but each vertex is now allowed to be a classical crossing or a new type called virtual. The generalized moves show how to manipulate such diagrams to obtain an equivalent diagram; one move called the semi-virtual move involves both classical and virtual crossings, but all the other moves involve only one variety of crossing. Virtual knots are important, and there is a strong relation between Quantum Field Theory and virtual knots. Virtual knots themselves are fascinating objects, and having many connections to other areas of mathematics. Virtual knots have many exciting connections with other fields of knots theory. The unsolved problem shown is an important motivation to the study of virtual knots. See section 1.1 of this paper [KOS] for the background and the history of this problem. Kauffman submitted a solution in the case of the product manifold of closed oriented surface and the closed interval, by introducing virtual 1-knots . It is open in the other cases. Witten’s path integral for Jones polynomial is written for links in any compact 3-manifold formally, but the calculus is not done even in physics level in any case other than the 3-sphere (the 3-ball, the 3-space R3). This problem is also open in physics level. In the
https://en.wikipedia.org/wiki/Rigi%20%28software%29
Rigi is an interactive graph editor tool for software reverse engineering using the white box method, i.e. necessitating source code, thus it is mainly aimed at program comprehension. Rigi is distributed by its main author, Hausi A. Müller and the Rigi research group at the University of Victoria. Rigi provides interactive links from the graphs it produces to the source code, but not vice versa. Rigi renders trees and grid-layout graphs using its own internal engine, but relies on University of Passau's GraphEd for more advanced layouts. The public version of Rigi has built-in parsers ("fact extractors") for C and Cobol, and can leverage the C++ parser of IBM Visual Age. It can also accept external data in an RSF format (it introduced), so external parses can also feed it data, for example SHriMP tool's Java parser. Some efforts were made to integrate Rigi in Microsoft Visual Studio .NET. Early versions of Bauhaus were also built on top of Rigi; the author of this latter tool notes that the combination was rather slow for graphs having more than 500 nodes. Rigi was reportedly used to analyze some (undisclosed) embedded software at Nokia, in the range of hundreds of thousands of lines of code, and was met with positive feedback from the Nokia engineers. Active development of Rigi has ceased in 1999, with the last official version released in 2003. A 2008 paper noted that "Rigi is a mature tool that is still used in research and popular in teaching, but it is currently no longer actively evolved and is in bug-fix mode." See also Imagix 4D Rational Rose Sourcetrail
https://en.wikipedia.org/wiki/Benny%20Benson
John Ben Benson Jr. (September 12, 1912 – July 2, 1972) was an Alaska native best known for designing the flag of Alaska. Benson was age 14 years old when he won a contest in 1927 to design the flag for the Territory of Alaska, which became a U.S. state on January 3, 1959. Life Early life Benson was born in Chignik, Alaska, on September 12, 1912. He was a Qawalangin (Fox Islander) Unangan born to a Swedish-American father, John Ben "Benny" Benson Sr., and Aleut-Russian mother, Tatiana Ioannovna Dediukhina, from a village near Unalaska. When he was three years old, his mother died, forcing his father to send him and his brother Carl to an orphanage, as Benson's father could not take care of them. Benson grew up at the Jesse Lee Home for Children in Unalaska and later in Seward. Fox Farm After graduating from school in 1932, Benson left the Jesse Lee Home. He returned to the Aleutian Islands to work with his father fishing and, for a time, trapping blue foxes on Chirikof Island and on John's privately own Ugaiushak Island. The rate for furs began to decline, so Benson moved to Seattle in 1936. He used the $1,000 prize from the flag design competition to enroll in the Hemphill Diesel Engineering School for Diesel engine repair. In 1938, Benson married Betty Van Hise. The couple's first child, Anna May, was born in October 1938. Their second daughter, Charlotte Abbot, was born in June 1940. Benson divorced in 1950 and moved with his daughters to Kodiak where he became an airplane mechanic for Kodiak Airways. Later years and death Benson met his sister in the mid 1950s, 30 years after their separation. His sister died soon after. His brother Carl also died in 1965. Benson's right leg had to be amputated in 1969 due to circulatory problems. Shortly after that, in 1972, he met and married a former Jesse Lee Home resident, Anna Sophie Jenks. Benson had several stepchildren and grandchildren. He died of a heart attack in Kodiak, Alaska, on July 2 of that year, at the
https://en.wikipedia.org/wiki/Transition%20constraint
A transition constraint is a way of enforcing that the data does not enter an impossible state because of a previous state. For example, it should not be possible for a person to change from being "married" to being "single, never married". The only valid states after "married" might be "divorced", "widowed", or "deceased". This is the database-centric interpretation of the term. In formal models in computer security, a transition constraint is a property that governs every valid transition from a state of the model to a successor state. It can be viewed as complementary to the state criteria that pertain to states per se but have no bearing on transitions between successive states.
https://en.wikipedia.org/wiki/Plica%20semilunaris%20of%20conjunctiva
The plica semilunaris is a small fold of bulbar conjunctiva on the medial canthus of the eye. It functions during movement of the eye, to help maintain tear drainage via the lacrimal lake, and to permit greater rotation of the globe, for without the plica the conjunctiva would attach directly to the eyeball, restricting movement. It is the vestigial remnant of the nictitating membrane (the "third eyelid") which is drawn across the eye for protection, and is present in other animals such as birds, reptiles, and fish, but is rare in mammals, mainly found in monotremes and marsupials. Its associated muscles are also vestigial. It is loose, thus eye movements are not restricted by it. Only one species of primate, the Calabar angwantibo, is known to have a functioning nictitating membrane. With ocular allergies, the lacrimal caruncle and plica semilunaris may be inflamed and pruritic (itchy) due to histamine release in the tissue and tear film.
https://en.wikipedia.org/wiki/Mixed%20complementarity%20problem
Mixed Complementarity Problem (MCP) is a problem formulation in mathematical programming. Many well-known problem types are special cases of, or may be reduced to MCP. It is a generalization of nonlinear complementarity problem (NCP). Definition The mixed complementarity problem is defined by a mapping , lower values and upper values . The solution of the MCP is a vector such that for each index one of the following alternatives holds: ; ; . Another definition for MCP is: it is a variational inequality on the parallelepiped . See also Complementarity theory
https://en.wikipedia.org/wiki/Earnings%20response%20coefficient
In financial economics, finance, and accounting, the earnings response coefficient, or ERC, is the estimated relationship between equity returns and the unexpected portion of (i.e., new information in) companies' earnings announcements. Development Arbitrage pricing theory describes the theoretical relationship between information that is known to market participants about a particular equity (e.g., a common stock share of a particular company) and the price of that equity. Under the strong form of the efficient market hypothesis, equity prices are expected in the aggregate to reflect all relevant information at a given time. Market participants with superior information are expected to exploit that information until share prices have effectively impounded the information. Therefore, in the aggregate, a portion of changes in a company's share price is expected to result from changes in the relevant information available to the market. The ERC is an estimate of the change in a company's stock price due to the information provided in a company's earnings announcement. The ERC is expressed mathematically as follows: UR = the unexpected return a = benchmark rate b = earning response coefficient (ern-u) = (actual earnings less expected earnings) = unexpected earnings e = random movement Earnings response coefficient research attempts to identify and explain the differential market response to earnings information of different firms. An Earnings response coefficient measures the extent of security’s abnormal market return in response to the unexpected component of reported earnings of the firm issuing that security. and The relationship between stock returns to profit to determine the extent of the response that occurs to as the Earnings Response Coefficient (ERC). Some studies reveal there are four factors that affect Earnings Response Coefficient (ERC), namely : beta, capital structure, persistence and growth. Reasons Reasons for differential market respons
https://en.wikipedia.org/wiki/Protection%20Profile
A Protection Profile (PP) is a document used as part of the certification process according to ISO/IEC 15408 and the Common Criteria (CC). As the generic form of a Security Target (ST), it is typically created by a user or user community and provides an implementation independent specification of information assurance security requirements. A PP is a combination of threats, security objectives, assumptions, security functional requirements (SFRs), security assurance requirements (SARs) and rationales. A PP specifies generic security evaluation criteria to substantiate vendors' claims of a given family of information system products. Among others, it typically specifies the Evaluation Assurance Level (EAL), a number 1 through 7, indicating the depth and rigor of the security evaluation, usually in the form of supporting documentation and testing, that a product meets the security requirements specified in the PP. The National Institute of Standards and Technology (NIST) and the National Security Agency (NSA) have agreed to cooperate on the development of validated U.S. government PPs. Purpose A PP states a security problem rigorously for a given collection of system or products, known as the Target of Evaluation (TOE) and to specify security requirements to address that problem without dictating how these requirements will be implemented. A PP may inherit requirements from one or more other PPs. In order to get a product evaluated and certified according to the CC, the product vendor has to define a Security Target (ST) which may comply with one or more PPs. In this way a PP may serve as a template for the product's ST. Problem areas Although the EAL is easiest for laymen to compare, its simplicity is deceptive because this number is rather meaningless without an understanding the security implications of the PP(s) and ST used for the evaluation. Technically, comparing evaluated products requires assessing both the EAL and the functional requirements. Unfo
https://en.wikipedia.org/wiki/The%20Journal%20of%20Object%20Technology
The Journal of Object Technology is an online scientific journal welcoming manuscripts describing theoretical, empirical, conceptual, and experimental results in the area of software and language engineering, including programming paradigms software language engineering model-based and model-driven engineering requirement engineering software architecture software validation & verification software maintenance and evolution software analytics software development process and methodology Each issue contains columns by regular columnists and peer-reviewed papers. Columnists include Dave Thomas, Won Kim, Bertrand Meyer and John McGregor. Its first issue appeared in May 2002, in response to the need for an international journal covering the object-oriented and component-based development field. From 1986 to 2001, the Journal of Object-Oriented Programming played much of that role, but after it was sold to a new owner in 2000, it was announced in the following year that it would cease publication at the end of the year. The launching of The Journal of Object Technology was largely a response to this event, as leaders from the community realized a flagship publication was needed. Since then the journal evolved by broadening its audience. Platinum open access The Journal of Object Technology is open-access and completely free to both readers and authors ("platinum" model). This model has been adopted since the journal's visionary creation in 2002 and applies to all contributions (available under the Creative Commons license). If you are interested in the journal's view on open access and in understanding what are the pros and cons of the different open access models you can have a look at the following editorial Open Access: all you wanted to know and never dared to ask. Continuous publication scheme The Journal of Object Technology uses a continuous publication scheme whereby regular papers, upon acceptance, are immediately added into a dynamic annual issue
https://en.wikipedia.org/wiki/Dorsal%20carpal%20branch%20of%20the%20ulnar%20artery
The dorsal carpal branch of the ulnar artery arises from the ulnar artery immediately above the pisiform bone, and winds backward beneath the tendon of the flexor carpi ulnaris; it passes across the dorsal surface of the carpus beneath the extensor tendons, to anastomose with a corresponding branch of the radial artery. Immediately after its origin, it gives off a small branch, which runs along the ulnar side of the fifth metacarpal bone, and supplies the ulnar side of the dorsal surface of the little finger.
https://en.wikipedia.org/wiki/Palmar%20carpal%20branch%20of%20radial%20artery
The palmar carpal branch of the radial artery is a small branch of the radial artery which arises near the lower border of the pronator quadratus, and, running across the front of the carpus, anastomoses with the palmar carpal branch of the ulnar artery.
https://en.wikipedia.org/wiki/Tubby%20protein
The tubby protein is encoded by the TUB gene. It is an upstream cell signaling protein common to multicellular eukaryotes. The first tubby gene was identified in mice, and proteins that are homologous to tubby are known as "tubby-like proteins" (TULPs). They share a common and characteristic tertiary structure that consists of a beta barrel packed around an alpha helix in the central pore. The gene derives its name from its role in metabolism; mice with a mutated tubby gene develop delayed-onset obesity, sensorineural hearing loss, and retinal degeneration. Structure Tubby proteins are classified as α+β proteins and have a 12-beta stranded barrel surrounding a central alpha helix. Tubby proteins can bind the small cell signaling molecule phosphatidylinositol, which is typically localized to the cell membrane. A similar structural fold to the Tubby like proteins has been identified in the Scramblase family of proteins. Function Tubby proteins have been implicated as transcription factors and as potential signaling factors coupled to G-protein activity. They are associated with neuronal differentiation and development, and in mammals are implicated in three disease processes when mutated: obesity, retinal degeneration, and hearing loss. In mice, mutations in tubby proteins are known to affect life span and fat storage as well as carbohydrate metabolism. Tubby domains associate with cytoplasmic side of cell membranes through binding of different phosphoinositides Human proteins containing this domain TUB; TULP1; TULP2; TULP3; TULP4; External links SCOP tubby fold
https://en.wikipedia.org/wiki/Thorpe%E2%80%93Ingold%20effect
The Thorpe–Ingold effect, gem-dimethyl effect, or angle compression is an effect observed in chemistry where increasing steric hindrance favours ring closure and intramolecular reactions. The effect was first reported by Beesley, Thorpe, and Ingold in 1915 as part of a study of cyclization reactions. It has since been generalized to many areas of chemistry. The comparative rates of lactone formation (lactonization) of various 2-hydroxybenzenepropionic acids illustrate the effect. The placement of an increasing number of methyl groups accelerates the cyclization process. One application of this effect is addition of a quaternary carbon (e.g., a gem-dimethyl group) in an alkyl chain to increase the reaction rate and/or equilibrium constant of cyclization reactions. An example of this is an olefin metathesis reaction: In the field of peptide foldamers, amino acid residues containing quaternary carbons such as 2-aminoisobutyric acid are used to promote formation of certain types of helices. One proposed explanation for this effect is that the increased size of the substituents increases the angle between them. As a result, the angle between the other two substituents decreases. By moving them closer together, reactions between them are accelerated. It is thus a kinetic effect. The effect also has some thermodynamic contribution as the in silico strain energy decreases on going from cyclobutane to 1-methylcyclobutane and 1,1-dimethylcyclobutane by a value between 8 kcal/mole and 1.5 kcal/mole. A noteworthy example of the Thorpe-Ingold effect in supramolecular catalysis is given by diphenylmethane derivatives provided with guanidinium groups. These compounds are active in the cleavage of the RNA model compound HPNP. Substitution of the methylene group of the parent diphenylmethane spacer with cyclohexylidene and adamantylidene moieties enhances catalytic efficiency, with gem dialkyl effect accelerations of 4.5 and 9.1, respectively. See also Chelate effect Flipp
https://en.wikipedia.org/wiki/Comparison%20of%20integrated%20development%20environments
The following tables list notable software packages that are nominal IDEs; standalone tools such as source-code editors and GUI builders are not included. These IDEs are listed in alphabetic order of the supported language. ActionScript Ada Assembly BASIC C/C++ C# COBOL Common Lisp Component Pascal D Eiffel Erlang Go to this page: Source code editors for Erlang Fortran F# Groovy Haskell Haxe Go to this page: Comparison of IDE choices for Haxe programmers Java Java has strong IDE support, due not only to its historical and economic importance, but also due to a combination of reflection and static-typing making it well-suited for IDE support. Some of the leading Java IDEs (such as IntelliJ and Eclipse) are also the basis for leading IDEs in other programming languages (e.g. for Python, IntelliJ is rebranded as PyCharm, and Eclipse has the PyDev plugin.) Open Closed JavaScript Julia Lua Pascal, Object Pascal Perl PHP Python R Racket Ruby Scala Smalltalk Tcl Unclassified IBM Rational Business Developer Mule (software) Visual Basic .NET See also Comparison of assemblers Graphical user interface builder List of compilers Source-code editor Game integrated development environment
https://en.wikipedia.org/wiki/Oat%20milk
Oat milk is a plant milk derived from whole oat (Avena spp.) grains by extracting the plant material with water. Oat milk has a creamy texture and mild oatmeal-like flavor, and is manufactured in various flavors, such as sweetened, unsweetened, vanilla, and chocolate. Unlike other plant milks having origins as early as the 13th century, oat milk was developed in the 1990s by the Swedish scientist Rickard Öste. Over 2017–2019, oat milk sales in the United States increased 10 fold, and one major manufacturer, Oatly, reported a three-fold increase in worldwide sales. As of late 2020, the oat milk market became second-largest among plant milks in the United States, following the leader, almond milk, but exceeding the sales of soy milk. By 2020, oat milk products included coffee creamer, yogurt alternatives, ice cream, and chocolate. Oat milk may be consumed to replace dairy in vegan diets, or in cases of medical conditions where dairy is incompatible, such as lactose intolerance or an allergy to cow milk. Compared to milk and other plant-based beverages, oat milk has relatively low environmental impact due to its comparatively low land and water needs for production. History Invention Soy milk and almond milk predate all other alternative milks, including oat milk, both as cultural and commercial products. Since the early 20th century, soy milk made its way from Asia to European and American grocery stores, initially as a dairy substitute due to lactose intolerance. The increase in consumption of soy milk since its global distribution created a large market for plant-based beverages like oat milk. The first recorded instance of an oat-based plant beverage was in the early 1990s, when Rickard Öste developed oat milk. Öste was working as a food scientist at Lund University in Lund, Sweden, researching lactose intolerance and sustainable food systems, when he invented the drink. Soon after, Öste founded Oatly, the first commercial manufacturer of oat milk. History of
https://en.wikipedia.org/wiki/Simple%20Sensor%20Interface%20protocol
The Simple Sensor Interface (SSI) protocol is a simple communications protocol designed for data transfer between computers or user terminals and smart sensors. The SSI protocol is an Application layer protocol as in the OSI model. The SSI protocol has been developed jointly by Nokia, Vaisala, Suunto, Ionific, Mermit and University of Oulu. Currently SSI is being developed within the Mimosa Project, part of the European Union Framework Programmes for Research and Technological Development. The SSI protocol is used in point-to-point communications over UART and networking nanoIP applications. SSI also provides polling sensors and streaming sensor data. For RFID sensor tags SSI specifies memory map for sensor data. The criteria for SSI protocol development are: general purpose simple – minimal overhead small footprint on the server (sensor) side Sample implementation of the SSI protocol for MSP430 microcontrollers will be published as open source during August 2006 by Nokia. SSI message structure An SSI message consists of a 2-byte header and an n-byte payload. The header consists of a one byte address (wildcard is '?', 0x3F in ASCII) and a one byte message/command type. The different possible values for the message/command type are presented in SSI v1.2 command base. SSI v1.2 command base The group of commands: Q – query A – query reply C – sensor discovery N – discovery reply Z – reset G – get sensor configuration S – set sensor configuration are used to find and configure sensor units utilizing the SSI-protocol. The group of commands: R – request sensor data V – data response D – data response with status field are used to read sensor data infrequently. For data streaming purposes defined commands are: O – create sensor observer Y – observer created K – delete observer U – observer finished L – request sensor listener J – sensor listener created. V – data response M – data response with many data points Point-to-point SSI Point-t
https://en.wikipedia.org/wiki/Wolfson%20Centre%20for%20Magnetics
Wolfson Centre for Magnetics (WCM) is a research and knowledge centre operating within School of Engineering at Cardiff University. Research WCM is a centre for research, teaching and technology transfer over a wide spectrum of magnetics, including magnetic engineering, magnetic materials, magnetic devices, and the physics of magnetism. Research within WCM focuses on several areas related to production, characterisation and applications of magnetic materials. The scope of the centre's research activities has recently been broadened by the addition of several new academic staff, allowing the Wolfson Centre for Magnetics to capitalise on the anticipated growth of research opportunities in collaborative, interdisciplinary projects in magnetism and magnetic materials. Computer aided design in magnetics, electromagnetic machines, magnetic imaging, high permeability materials, magnetostriction, magnetic sensors and actuators, nanomagnetic materials, magnetic thin films and multilayers, magnetic material for data storage, theory and modelling of magnetic materials are all areas of research currently under investigation in the centre. Postgraduate research and industrial contract projects are carried out with the support of equipment and research facilities recently upgraded through major infrastructure investments. WCM has collaborative links with leading research groups in magnetics throughout Europe, Asia, North and South America, Japan, China, India and Korea. Its members participate regularly in conferences, networking and collaborative research projects both nationally and internationally. Facilities The Wolfson Centre for Magnetics has a wide range of state of the facilities within its laboratories, which support the research and industrial consultancy activities. Ongoing investment, including an award from the EPSRC/EST Joint Infrastracture Fund (JIF) in 2003 helped in continuous improvement of existing infrastructure and a Strategic Research Investment Fund
https://en.wikipedia.org/wiki/Meijer%20G-function
In mathematics, the G-function was introduced by as a very general function intended to include most of the known special functions as particular cases. This was not the only attempt of its kind: the generalized hypergeometric function and the MacRobert E-function had the same aim, but Meijer's G-function was able to include those as particular cases as well. The first definition was made by Meijer using a series; nowadays the accepted and more general definition is via a line integral in the complex plane, introduced in its full generality by Arthur Erdélyi in 1953. With the modern definition, the majority of the established special functions can be represented in terms of the Meijer G-function. A notable property is the closure of the set of all G-functions not only under differentiation but also under indefinite integration. In combination with a functional equation that allows to liberate from a G-function G(z) any factor zρ that is a constant power of its argument z, the closure implies that whenever a function is expressible as a G-function of a constant multiple of some constant power of the function argument, f(x) = G(cxγ), the derivative and the antiderivative of this function are expressible so too. The wide coverage of special functions also lends power to uses of Meijer's G-function other than the representation and manipulation of derivatives and antiderivatives. For example, the definite integral over the positive real axis of any function g(x) that can be written as a product G1(cxγ)·G2(dxδ) of two G-functions with rational γ/δ equals just another G-function, and generalizations of integral transforms like the Hankel transform and the Laplace transform and their inverses result when suitable G-function pairs are employed as transform kernels. A still more general function, which introduces additional parameters into Meijer's G-function, is Fox's H-function and is used for Matrix transform by Ram Kishore Saxena One application of the Meijer G-func
https://en.wikipedia.org/wiki/Biological%20oceanography
Biological oceanography is the study of how organisms affect and are affected by the physics, chemistry, and geology of the oceanographic system. Biological oceanography may also be referred to as ocean ecology, in which the root word of ecology is Oikos (oικoσ), meaning ‘house’ or ‘habitat’ in Greek. With that in mind, it is of no surprise then that the main focus of biological oceanography is on the microorganisms within the ocean; looking at how they are affected by their environment and how that affects larger marine creatures and their ecosystem. Biological oceanography is similar to marine biology, but is different because of the perspective used to study the ocean. Biological oceanography takes a bottom-up approach (in terms of the food web), while marine biology studies the ocean from a top-down perspective. Biological oceanography mainly focuses on the ecosystem of the ocean with an emphasis on plankton: their diversity (morphology, nutritional sources, motility, and metabolism); their productivity and how that plays a role in the global carbon cycle; and their distribution (predation and life cycle). History In 325 BC, Pytheas of Massalia, a Greek geographer, explored much of the coast of England and Norway and developed the means of determining latitude from the declination of the North Star. His account of tides is also one of the earliest accounts that suggest a relationship between them and the moon. This relationship was later developed by English monk Bede in De Temporum Ratione (The Reckoning of Time) around 700 AD. Understanding the ocean began with the general exploration and voyaging for trade. Some notable events closer to our time, include Prince Henry the Navigator’s ocean exploration in the 1400s. In 1513, Ponce de Leon described the Florida Current. In 1674, Robert Boyle investigated the relationship between salinity, temperature, and pressure in the depths of the ocean. Captain James Cook’s voyages were responsible for the extensive da
https://en.wikipedia.org/wiki/Common%20Hybrid%20Interface%20Protocol%20System
Common Hybrid Interface Protocol System (CHIPS) is the definition of a computer network that consists of a mixture of common serial data protocols such as RS-232 and RS-485, or can be even PC keyboard interface communication. CHIPS may also consist of Bluetooth and Wi-Fi for wireless communication can be installed on all major hardware platforms. There are several CHIPS projects and products available today where such systems are i.e. MISOLIMA DOLLx8 and Olivetti's "Mael Gateasy". As new bus systems are gaining market shares, there will always be needs for CHIPS to enable serial network protocols to be integrated into one single connection point. By using CHIPS, it will be possible to control I/O data from different sources and systems without having the need to install several serial interface cards and drivers. CHIPS users will, in most cases, be able to work with several serial data transceiver sources at the same time. Such serial data might originate from PC Keyboards, CANbus, RS and wireless communication where all data connects into one or several CHIPS units that communicate over the mixed serial data protocols. Due to some mixed baud rates between the connected systems, the compatibility with CHIPS means that some devices will have reduced transfer rates, but CHIPS are primary designed for Lab-, office-, home-, factory- and building automation also used in Internet of Things.
https://en.wikipedia.org/wiki/Biologically%20inspired%20cognitive%20architectures
Biologically Inspired Cognitive Architectures (BICA) was a DARPA project administered by the Information Processing Technology Office (IPTO). BICA began in 2005 and is designed to create the next generation of cognitive architecture models of human artificial intelligence. Its first phase (Design) ran from September 2005 to around October 2006, and was intended to generate new ideas for biological architectures that could be used to create embodied computational architectures of human intelligence. The second phase (Implementation) of BICA was set to begin in the spring of 2007, and would have involved the actual construction of new intelligent agents that live and behave in a virtual environment. However, this phase was canceled by DARPA, reportedly because it was seen as being too ambitious. Now BICA is a transdisciplinary study that aims to design, characterise and implement human-level cognitive architectures. There is also BICA Society, a scientific nonprofit organization formed to promote and facilitate this study. On their website, they have an extensive comparison table of various cognitive architectures.
https://en.wikipedia.org/wiki/Double%20dabble
In computer science, the double dabble algorithm is used to convert binary numbers into binary-coded decimal (BCD) notation. It is also known as the shift-and-add-3 algorithm, and can be implemented using a small number of gates in computer hardware, but at the expense of high latency. Algorithm The algorithm operates as follows: Suppose the original number to be converted is stored in a register that is n bits wide. Reserve a scratch space wide enough to hold both the original number and its BCD representation; bits will be enough. It takes a maximum of 4 bits in binary to store each decimal digit. Then partition the scratch space into BCD digits (on the left) and the original register (on the right). For example, if the original number to be converted is eight bits wide, the scratch space would be partitioned as follows: Hundreds Tens Ones Original 0010 0100 0011 11110011 The diagram above shows the binary representation of 24310 in the original register, and the BCD representation of 243 on the left. The scratch space is initialized to all zeros, and then the value to be converted is copied into the "original register" space on the right. 0000 0000 0000 11110011 The algorithm then iterates n times. On each iteration, any BCD digit which is at least 5 (0101 in binary) is incremented by 3 (0011); then the entire scratch space is left-shifted one bit. The increment ensures that a value of 5, incremented and left-shifted, becomes 16 (10000), thus correctly "carrying" into the next BCD digit. Essentially, the algorithm operates by doubling the BCD value on the left each iteration and adding either one or zero according to the original bit pattern. Shifting left accomplishes both tasks simultaneously. If any digit is five or above, three is added to ensure the value "carries" in base 10. The double-dabble algorithm, performed on the value 24310, looks like this: 0000 0000 0000 11110011 Initialization 0000 0000 0001 11100110 Shift 000
https://en.wikipedia.org/wiki/Cooperative%20breeding
Cooperative breeding is a social system characterized by alloparental care: offspring receive care not only from their parents, but also from additional group members, often called helpers. Cooperative breeding encompasses a wide variety of group structures, from a breeding pair with helpers that are offspring from a previous season, to groups with multiple breeding males and females (polygynandry) and helpers that are the adult offspring of some but not all of the breeders in the group, to groups in which helpers sometimes achieve co-breeding status by producing their own offspring as part of the group's brood. Cooperative breeding occurs across taxonomic groups including birds, mammals, fish, and insects. Costs for helpers include a fitness reduction, increased territory defense, offspring guarding and an increased cost of growth. Benefits for helpers include a reduced chance of predation, increased foraging time, territory inheritance, increased environmental conditions and an inclusive fitness. Inclusive fitness is the sum of all direct and indirect fitness, where direct fitness is defined as the amount of fitness gained through producing offspring. Indirect fitness is defined as the amount of fitness gained through aiding the offspring of related individuals, that is, relatives are able to indirectly pass on their genes through increasing the fitness of related offspring. This is also called kin selection. For the breeding pair, costs include increased mate guarding and suppression of subordinate mating. Breeders receive benefits as reductions in offspring care and territory maintenance. Their primary benefit is an increased reproductive rate and survival. Cooperative breeding causes the reproductive success of all sexually mature adults to be skewed towards one mating pair. This means the reproductive fitness of the group is held within a select few breeding members and helpers have little to no reproductive fitness. With this system, breeders gain an incre
https://en.wikipedia.org/wiki/Potentiometric%20sensor
A potentiometric sensor is a type of chemical sensor that may be used to determine the analytical concentration of some components of the analyte gas or solution. These sensors measure the electrical potential of an electrode when no current is present. Principle The signal is measured as the potential difference (voltage) between the working electrode and the reference electrode. The working electrode's potential must depend on the concentration of the analyte in the gas or solution phase. The reference electrode is needed to provide a defined reference potential. ................. Classification of sensors Potentiometric solid state gas sensors have been generally classified into three broad groups. Type I sensors have an electrolyte containing mobile ions of the chemical species in the gas phase that it is monitoring. The commercial product, YSZ oxygen sensor, is an example of type I. Type II sensors do not have mobile ions of the chemical species to be sensed, but an ion related to the target gas can diffuse in the solid electrolyte to allow equilibration with the atmosphere. Therefore, type I and type II sensors have the same design with gas electrodes combined with metal and an electrolyte where oxidized or reduced ions can be electrochemically equilibrated through the electrochemical cell. In the third type of electrochemical sex , auxiliary phases are added to the electrodes to enhance the selectivity and stability. Type III sensors make the electrode concept even more confusing. With respect to the design of a solid state sensor, the auxiliary phase looks as part of the electrode. But it cannot be an electrode because auxiliary phase materials are not generally good electrical conductor. In spite of this confusion, type III design offers more feasibility in terms of designing various sensors with different auxiliary materials and electrolytes. See also Potentiometer
https://en.wikipedia.org/wiki/Taenia%20of%20fourth%20ventricle
In the brain, the taenia of the fourth ventricle (lingula, tenia of fourth ventricle) are two narrow bands of white matter, one on either side, which complete the lower part of the roof of the fourth ventricle. Each consists of a vertical and a horizontal part. The vertical part is continuous below the obex with the gracile nucleus, to which it is adherent by its lateral border. The horizontal portion extends transversely across the inferior peduncle, below the striae medullares, and roofs in the lower and posterior part of the lateral recess; it is attached by its lower margin to the inferior peduncle, and partly encloses the choroid plexus, which, however, projects beyond it like a cluster of grapes; and hence this part of the tænia has been termed the cornucopia. Additional images See also Fourth ventricle
https://en.wikipedia.org/wiki/Brake%20force
Brake force, also known as Brake Power, is a measure of braking power of a vehicle. Railways In the case of railways, it is important that staff are aware of the brake force of a train so sufficient brake power will be available to bring the train to a halt within the required distance from a given speed. In simple terms the brake force of a train should be relative to the sum of the brake force that can be exerted by all the vehicles in the train relative to the weight of the train, excluding problems that may occur such as wheels locking and sliding under braking. Modern freight wagons typically have brakes that can be operated from the locomotive, these are sometimes referred to as fitted freights. Older wagons typically were not fitted with brakes that could be operated from the locomotive, sometimes these are referred to as unfitted freights. These unfitted freights would typically have brake vans attached to provide additional braking force and operated at a reduced field. While very early passenger trains might have had brakes that would have been applied by a brakesman riding in say every second carriage modern passenger vehicles have brakes that will be applied to all vehicles. There are a certain of cases in modern practice where operating at higher at higher speeds can could the brake force above a certain speed would be insufficient to stop the train within the required distance. Some cases arise because the brake force to weight ratio of a locomotive of typically 80 to 120 tonnes in weight is often less than that of passenger vehicle in the 40 tonne range; and it may be that locomotives running by themselves require several coaches to be attached if the train is to run at its maximum permitted speed with a suitable brake force to weight ratio. Another issue that can arise is when a locomotive is hauling a set of coaches, typically a multiple unit, where it is unable to work the brakes of the unit. In this case additional braking vehicles may
https://en.wikipedia.org/wiki/Bing%20Mobile
Bing for mobile (formerly Live Search Mobile) is a search tool for handheld mobile devices from Microsoft as part of their Bing search engine. It is designed for mobile device displays. Bing Mobile is built into Windows Mobile and Windows Phone as proprietary software, accessed via the Search key on Windows Phone 7 and Windows Phone 8 devices. It is also available on Windows Phone 8.1 (and integrated with Microsoft Cortana where available), and can be downloaded for other platforms, including and Android. Windows Phone With the launch of Windows Phone, the new operating system included the Bing Hub, a centralised hub for contextual web searches and Bing Maps for navigation built-in which gained some additional features such as showing users real-time traffic updates, street-view photos, 3D graphics, and directions. To make Windows Phone hardware more uniform Microsoft requires all Windows Phones feature a dedicated "Search button" that opens up Bing Mobile. With Windows Phone 7.5, the first major update to Windows Phone Microsoft included several new features to the Bing Hub including the new Bing Vision application that allows users to scan QR Codes, books, price tags, and various other items but unlike Google Goggles, can not scan any object due to its limited functionality, other than scanning objects Bing Vision comes with a built-in version of the Bing Translator that can scan texts and translate it into the phone's language, Engadget did a test they scanned the back cover a book and it only picked up about 90 percent of the words in the summary after the text was found, Bing Vision's translation feature supports 26 languages. Bing Maps also received new features namely the Bing Local Scout which takes a look at businesses around the user's location and gives them a list of local restaurants, bars, and shopping centres. The Bing Hub can also open up Bing Local Scout and got a new feature added in Windows Phone 7.5 called Bing Audio, a service similar to Sha
https://en.wikipedia.org/wiki/Peptide%20library
A peptide library is a tool for studying proteins. Peptide libraries typically contain a large number of peptides that have a systematic combination of amino acids. Usually, the peptide library is synthesized on a solid phase, mostly on resin, which can be made as a flat surface or beads. The peptide library is a popular tool for drug design, protein–protein interactions, and other biochemical and pharmaceutical applications. Synthetic peptide libraries are synthesized without utilizing phage or other biological systems. There are at least five subtypes of synthetic peptide libraries that differ from each other by the method which was used for the synthesis of the library. The subtypes are: Overlapping peptide libraries Truncation peptide libraries Random libraries Alanine scanning libraries Positional or scrambled peptide libraries. This form of peptide synthesis is limited to a peptide chain length of approximately 70 amino acids, and is generally unsuitable for the study of larger proteins. For 20 amino acids, this results in an upper limit of 2070 possible combinations, not considering the plethora of available amino acids with pre-installed post-translational modifications. From this number of total combinations, the scope of the peptide library can be narrowed down for a more specific purpose by selecting which amino acids are desired at each point in the chain. For example, a peptide chain of 10 residues in length is used in native chemical ligation with a larger recombinantly expressed protein. Residue 1: alanine Residue 2: one of glutamine, glycine, arginine, glutamic acid, serine, or methionine Residue 3: any one of the 20 amino acids Residue 4: acetyllysine Residue 5: alanine Residue 6: isoleucine Residue 7: aspartic acid Residue 8: phenylalanine Residue 9: acetyllysine Residue 10: arginine with the carboxy terminal thioester With 7 possibilities at Residue 2 and 20 possibilities at Residue 3, the total would be or 140 different pol
https://en.wikipedia.org/wiki/Immanant
In mathematics, the immanant of a matrix was defined by Dudley E. Littlewood and Archibald Read Richardson as a generalisation of the concepts of determinant and permanent. Let be a partition of an integer and let be the corresponding irreducible representation-theoretic character of the symmetric group . The immanant of an matrix associated with the character is defined as the expression Examples The determinant is a special case of the immanant, where is the alternating character , of Sn, defined by the parity of a permutation. The permanent is the case where is the trivial character, which is identically equal to 1. For example, for matrices, there are three irreducible representations of , as shown in the character table: As stated above, produces the permanent and produces the determinant, but produces the operation that maps as follows: Properties The immanant shares several properties with determinant and permanent. In particular, the immanant is multilinear in the rows and columns of the matrix; and the immanant is invariant under simultaneous permutations of the rows or columns by the same element of the symmetric group. Littlewood and Richardson studied the relation of the immanant to Schur functions in the representation theory of the symmetric group. The necessary and sufficient conditions for the immanant of a Gram matrix to be are given by Gamas's Theorem.
https://en.wikipedia.org/wiki/Breeding%20pair
Breeding pair is a pair of animals which cooperate over time to produce offspring with some form of a bond between the individuals. For example, many birds mate for a breeding season or sometimes for life. They may share some or all of the tasks involved: for example, a breeding pair of birds may split building a nest, incubating the eggs and feeding and protecting the young. The term is not generally used when a male has a harem of females, such as with mountain gorillas. True breeding pairs are usually found only in vertebrates, but there are notable exceptions, such as the Lord Howe Island stick insect. True breeding pairs are rare in amphibians or reptiles, although the Australian Shingleback is one exception with long-term pair-bonds. Some fish form short term pairs and the French angelfish is thought to pair-bond over a long term. True breeding pairs are quite common in birds. Breeding pair arrangements are rare in mammals, where the prevailing patterns are either that the male and female only meet for copulation (e.g. brown bear) or that dominant males have a harem of females (e.g. walrus). See also Pair bond Monogamous pairing in animals
https://en.wikipedia.org/wiki/Digital%20mixing%20console
In professional audio, a digital mixing console (DMC) is a type of mixing console used to combine, route, and change the dynamics, equalization and other properties of multiple audio input signals, using digital signal processing rather than analog circuitry. The digital audio samples, which is the internal representation of the analog inputs, are summed to what is known as a master channel to produce a combined output. A professional digital mixing console is a dedicated desk or control surface produced exclusively for the task and is typically more robust in terms of user control, processing power and quality of audio effects. However, a computer can also perform the same function since it can mimic its interface, input and output. Uses Digital mixing consoles are typically used in recording studios, public address systems, sound reinforcement systems, broadcasting, television, and film post-production. Common sound system problems and solutions Most DMCs are expensive and sophisticated tools. The most common issue related to the DMC is the complex structure, which can be difficult to navigate without previous experience or knowledge of the system. The user usually requires a basic understanding of signal flow, audio terminology, and hardware implementation. Part of the solution to alleviate operator issues is to automate whenever possible. The advent of modern digital computer technology has now made it possible to install sound system components that will, to some extent, operate themselves. A digital mixing console can offset the lack of operator experience because it can store the settings programmed by an expert mixing engineer. After everything is properly adjusted, that set-up is assigned a name and stored in the memory. Afterwards, a less knowledgeable operator can simply select that setting on their console or computer. One can easily program many different preset configurations or "snapshots," into the mixing console. Default configurations tha
https://en.wikipedia.org/wiki/Charged%20current
Charged current interactions are one of the ways in which subatomic particles can interact by means of the weak force. These interactions are mediated by the and bosons. In simple terms Charged current interactions are the most easily detected class of weak interactions. The weak force is best known for mediating nuclear decay. It has very short range, but is the only force (apart from gravity) to interact with neutrinos. The weak force is communicated via the W and Z exchange particles. Of these, the W-boson has either a positive or negative electric charge, and mediates neutrino absorption and emission by or with an electrically charged particle. During these processes, the W-boson induces electron or positron emission or absorption, or changing the flavour of a quark as well as its electrical charge, such as in beta decay or K-capture. By contrast, the Z particle is electrically neutral, and exchange of a Z-boson leaves the interacting particles’ quantum numbers unaffected, except for a transfer of momentum, spin, and energy. Because exchange of W bosons involves a transfer of electric charge (as well as a transfer of weak isospin, while weak hypercharge is not transferred), it is known as “charged current”. By contrast, exchanges of Z bosons involve no transfer of electrical charge, so it is referred to as a “neutral current”. In the latter case, the word “current” has nothing to do with electricity – it simply refers to the Z bosons’ movement between other particles. Definition The name 'charged current' arises due to currents of fermions coupled to the W bosons having electric charge. For example, the charged current contribution to the → elastic scattering amplitude is: where the charged currents describing the flow of one fermion into the other are given by: The W-Boson can couple to any particle with weak isospin (i.e. any left-handed Standard Model fermions). See also Neutral current W and Z bosons Notes
https://en.wikipedia.org/wiki/De%20Gua%27s%20theorem
In mathematics, De Gua's theorem is a three-dimensional analog of the Pythagorean theorem named after Jean Paul de Gua de Malves. It states that if a tetrahedron has a right-angle corner (like the corner of a cube), then the square of the area of the face opposite the right-angle corner is the sum of the squares of the areas of the other three faces: De Gua's theorem can be applied for proving a special case of Heron's formula. Generalizations The Pythagorean theorem and de Gua's theorem are special cases () of a general theorem about n-simplices with a right-angle corner, proved by P. S. Donchian and H. S. M. Coxeter in 1935. This, in turn, is a special case of a yet more general theorem by Donald R. Conant and William A. Beyer (1974), which can be stated as follows. Let U be a measurable subset of a k-dimensional affine subspace of (so ). For any subset with exactly k elements, let be the orthogonal projection of U onto the linear span of , where and is the standard basis for . Then where is the k-dimensional volume of U and the sum is over all subsets with exactly k elements. De Gua's theorem and its generalisation (above) to n-simplices with right-angle corners correspond to the special case where k = n−1 and U is an (n−1)-simplex in with vertices on the co-ordinate axes. For example, suppose , and U is the triangle in with vertices A, B and C lying on the -, - and -axes, respectively. The subsets of with exactly 2 elements are , and . By definition, is the orthogonal projection of onto the -plane, so is the triangle with vertices O, B and C, where O is the origin of . Similarly, and , so the Conant–Beyer theorem says which is de Gua's theorem. The generalisation of de Gua's theorem to n-simplices with right-angle corners can also be obtained as a special case from the Cayley–Menger determinant formula. De Gua's theorem can also be generalized to arbitrary tetrahedra and to pyramids. History Jean Paul de Gua de Malves (1713–85)
https://en.wikipedia.org/wiki/Targeted%20drug%20delivery
Targeted drug delivery, sometimes called smart drug delivery, is a method of delivering medication to a patient in a manner that increases the concentration of the medication in some parts of the body relative to others. This means of delivery is largely founded on nanomedicine, which plans to employ nanoparticle-mediated drug delivery in order to combat the downfalls of conventional drug delivery. These nanoparticles would be loaded with drugs and targeted to specific parts of the body where there is solely diseased tissue, thereby avoiding interaction with healthy tissue. The goal of a targeted drug delivery system is to prolong, localize, target and have a protected drug interaction with the diseased tissue. The conventional drug delivery system is the absorption of the drug across a biological membrane, whereas the targeted release system releases the drug in a dosage form. The advantages to the targeted release system is the reduction in the frequency of the dosages taken by the patient, having a more uniform effect of the drug, reduction of drug side-effects, and reduced fluctuation in circulating drug levels. The disadvantage of the system is high cost, which makes productivity more difficult, and the reduced ability to adjust the dosages. Targeted drug delivery systems have been developed to optimize regenerative techniques. The system is based on a method that delivers a certain amount of a therapeutic agent for a prolonged period of time to a targeted diseased area within the body. This helps maintain the required plasma and tissue drug levels in the body, thereby preventing any damage to the healthy tissue via the drug. The drug delivery system is highly integrated and requires various disciplines, such as chemists, biologists, and engineers, to join forces to optimize this system. Background In traditional drug delivery systems such as oral ingestion or intravascular injection, the medication is distributed throughout the body through the systemic
https://en.wikipedia.org/wiki/Identity%20driven%20networking
Identity driven networking (IDN) is the process of applying network controls to a network device access based on the identity of an individual or a group of individuals responsible to or operating the device. Individuals are identified, and the network is tuned to respond to their presence by context. The OSI model provides a method to deliver network traffic, not only to the system but to the application that requested or is listening for data. These applications can operate either as a system based user-daemon process, or as a user application such as a web browser. Internet security is built around the idea that the ability to request or respond to requests should be subjected to some degree of authentication, validation, authorization, and policy enforcement. Identity driven networking endeavors to resolve user and system based policy into a single management paradigm. Since the internet comprises a vast range of devices and applications there are also many boundaries and therefore ideas on how to resolve connectivity to users within those boundaries. An endeavor to overlay the system with an identity framework must first decide what an Identity is, determine it, and only then use existing controls to decide what is intended with this new information. The Identity A digital identity represents the connectedness between the real and some projection of an identity; and it may incorporate references to devices as well as resources and policies. In some systems, policies provide the entitlements that an identity can claim at any particular point in time and space. For example, a person may be entitled to some privileges during work from their workplace that may be denied from home out of hours. How it might work Before a user gets to the network there is usually some form of machine authentication, this probably verifies and configures the system for some basic level of access. Short of mapping a user to a MAC address prior or during this process (802.1x) it i
https://en.wikipedia.org/wiki/Control%20variable
A control variable (or scientific constant) in scientific experimentation is an experimental element which is constant (controlled) and unchanged throughout the course of the investigation. Control variables could strongly influence experimental results were they not held constant during the experiment in order to test the relative relationship of the dependent variable (DV) and independent variable (IV). The control variables themselves are not of primary interest to the experimenter. Usage A variable in an experiment which is held constant in order to assess the relationship between multiple variables, is a control variable. A control variable is an element that is not changed throughout an experiment because its unchanging state allows better understanding of the relationship between the other variables being tested. In any system existing in a natural state, many variables may be interdependent, with each affecting the other. Scientific experiments test the relationship of an IV (or independent variable: that element that is manipulated by the experimenter) to the DV (or dependent variable: that element affected by the manipulation of the IV). Any additional independent variable can be a control variable. A control variable is an experimental condition or element that is kept the same throughout the experiment, and it is not of primary concern in the experiment, nor will it influence the outcome of the experiment. Any unexpected (e.g.: uncontrolled) change in a control variable during an experiment would invalidate the correlation of dependent variables (DV) to the independent variable (IV), thus skewing the results, and invalidating the working hypothesis. This indicates the presence of a spurious relationship existing within experimental parameters. Unexpected results may result from the presence of a confounding variable, thus requiring a re-working of the initial experimental hypothesis. Confounding variables are a threat to the internal validity of
https://en.wikipedia.org/wiki/Outline%20of%20arithmetic
Arithmetic is an elementary branch of mathematics that is widely used for tasks ranging from simple day-to-day counting to advanced science and business calculations. Essence of arithmetic Elementary arithmetic Decimal arithmetic Decimal point Numeral Place value Face value History of arithmetic Arithmetic operations and related concepts Order of Operations Addition Summation – Answer after adding a sequence of numbers Additive Inverse Subtraction – Taking away numbers Multiplication – Repeated addition Multiple – Product of Multiplication Least Common Multiple Multiplicative Inverse Division – Repeated Subtraction Modulo – The remainder of division Quotient – Result of Division Quotition and Partition – How many parts are there, and what is the size of each part Fraction – A number that isn't whole, often shown as a divsion equation Decimal Fraction – Representation of a Fraction in the form of a number Proper Fraction – Fraction with a Numerator that is less than the Denominator Improper Fraction – Fractions with a Numerator that is any number Ratio – Showing how much one number can go into another Least Common Denominator – Least Common Multiple of 2 or more fractions' denominators Factoring – Breaking a number down into its products Fundamental theorem of arithmetic Prime number – Number divisable by only 1 or itself Prime number theorem Distribution of primes Composite number – Number made of 2 smaller integers Factor – A number that can be divided from it's original number to get a whole number Greatest Common Factor – Greatest Factor that is common between 2 numbers Euclid's algorithm for finding greatest common divisors Exponentiation (power) – Repreated Multiplication Square root – Reversal of a power of 2 (exponent of 1/2) Cube root – Reversal of a power of 3 (exponent of 1/3) Properties of Operations Associative property Distributive property Commutative property Factorial – Multiplication of numbers from the current number to 0 Types of numbers Re
https://en.wikipedia.org/wiki/Outline%20of%20algebra
The following outline is provided as an overview of and topical guide to algebra: Algebra is one of the main branches of mathematics, covering the study of structure, relation and quantity. Algebra studies the effects of adding and multiplying numbers, variables, and polynomials, along with their factorization and determining their roots. In addition to working directly with numbers, algebra also covers symbols, variables, and set elements. Addition and multiplication are general operations, but their precise definitions lead to structures such as groups, rings, and fields. Branches Pre-algebra Elementary algebra Boolean algebra Abstract algebra Linear algebra Universal algebra Algebraic equations An algebraic equation is an equation involving only algebraic expressions in the unknowns. These are further classified by degree. Linear equation – algebraic equation of degree one. Polynomial equation – equation in which a polynomial is set equal to another polynomial. Transcendental equation – equation involving a transcendental function of one of its variables. Functional equation – equation in which the unknowns are functions rather than simple quantities. Differential equation – equation involving derivatives. Integral equation – equation involving integrals. Diophantine equation – equation where the only solutions of interest of the unknowns are the integer ones. History History of algebra General algebra concepts Fundamental theorem of algebra – states that every non-constant single-variable polynomial with complex coefficients has at least one complex root. This includes polynomials with real coefficients, since every real number is a complex number with an imaginary part equal to zero. Equations – equality of two mathematical expressions Linear equation – an algebraic equation with a degree of one Quadratic equation – an algebraic equation with a degree of two Cubic equation – an algebraic equation with a degree of three Quartic equati
https://en.wikipedia.org/wiki/Outline%20of%20geometry
Geometry is a branch of mathematics concerned with questions of shape, size, relative position of figures, and the properties of space. Geometry is one of the oldest mathematical sciences. Classical branches Geometry Analytic geometry Differential geometry Euclidean geometry Non-Euclidean geometry Projective geometry Riemannian geometry Contemporary branches Absolute geometry Affine geometry Archimedes' use of infinitesimals Birational geometry Complex geometry Combinatorial geometry Computational geometry Conformal geometry Constructive solid geometry Contact geometry Convex geometry Descriptive geometry Digital geometry Discrete geometry Distance geometry Elliptic geometry Enumerative geometry Epipolar geometry Finite geometry Geometry of numbers Hyperbolic geometry Incidence geometry Information geometry Integral geometry Inversive geometry Klein geometry Lie sphere geometry Numerical geometry Ordered geometry Parabolic geometry Plane geometry Quantum geometry Ruppeiner geometry Spherical geometry Symplectic geometry Synthetic geometry Systolic geometry Taxicab geometry Toric geometry Transformation geometry Tropical geometry History of geometry History of geometry Timeline of geometry Babylonian geometry Egyptian geometry Ancient Greek geometry Euclidean geometry Pythagorean theorem Euclid's Elements Measurement of a Circle Indian mathematics Bakhshali manuscript Modern geometry History of analytic geometry History of the Cartesian coordinate system History of non-Euclidean geometry History of topology History of algebraic geometry General geometry concepts General concepts Geometric progression — Geometric shape — Geometry — Pi — angular velocity — linear velocity — De Moivre's theorem — parallelogram rule — Pythagorean theorem — similar triangles — trigonometric identity — unit circle — Trapezoid — Triangle — Theorem — point — ray — plane — line — line segment Measurements Bearing A
https://en.wikipedia.org/wiki/Blastoderm
A blastoderm (germinal disc, blastodisc) is a single layer of embryonic epithelial tissue that makes up the blastula. It encloses the fluid filled blastocoel. Gastrulation follows blastoderm formation, where the tips of the blastoderm begins the formation of the ectoderm, mesoderm, and endoderm. Formation The blastoderm is formed when the oocyte plasma membrane begins cleaving by invagination, creating multiple cells that arrange themselves into an outer sleeve to the blastocoel. In oviparous In chicken eggs, the blastoderm represents a flat disc after embryonic fertilization. At the edge of the blastoderm is the site of active migration by most cells. See also Blastodisc Embryology Cleavage Gastrulation
https://en.wikipedia.org/wiki/RF64
RF64 is a BWF-compatible multichannel audio file format enabling file sizes to exceed 4 GB. It has been specified by the European Broadcasting Union. It has been accepted as the ITU recommendation ITU-R BS.2088. The file format is designed to meet the requirements for multichannel sound in broadcasting and audio archiving. It is based on the Microsoft RIFF/WAVE format and Wave Format Extensible for multichannel parameters. Additions are made to the basic specification to allow for more than 4 GB file sizes when needed (the new maximum filesize is now approximately 16 exabytes). The format is transparent to the BWF and all its supplements and chunks. RF64 WAV files typically use the .wav file extension. Capability A maximum of 18 surround channels, stereo down mix channel and bit stream signals with non-PCM coded data can also be stored in the file format. RF64 can be used in the entire programme chain from capture to editing and play out and for short or long term archiving of multichannel files. Due to the inconsistent usage of CUE data definition, the additional requirement that CUE chunk names be stored in an additional LABL chunk, along with the inherent 32-bit limitation of the CUE chunk pointer index, the 2009 RF64 format also defines a new 'r64m' marker chunk. The RF64 file format should fulfill the longer-term need for multichannel sound in broadcasting and archiving. The required effort for software implementers is very small. The changes that will be needed to update existing systems will be reasonable in cost. Format In its basic form, the 32-bit chunk size field at offset 4 in the file is set to -1 (0xFFFFFFFF), and immediately following that a new 'ds64' chunk is inserted (before the FMT chunk). This new ds64 chunk will contain the 64-bit sizes of the DATA chunk(s), using a simple sequential table mechanism to point to additional DATA chunks. The first 4 bytes of the file are then changed from 'RIFF' to 'BW64'. RF64 files define the following e
https://en.wikipedia.org/wiki/Insertional%20mutagenesis
In molecular biology, insertional mutagenesis is the creation of mutations in DNA by the addition of one or more base pairs. Such insertional mutations can occur naturally, mediated by viruses or transposons, or can be artificially created for research purposes in the lab. Signature tagged mutagenesis This is a technique used to study the function of genes. A transposon such as the P element of Drosophila melanogaster is allowed to integrate at random locations in the genome of the organism being studied. Mutants generated by this method are then screened for any unusual phenotypes. If such a phenotype is found then it can be assumed that the insertion has caused the gene relating to the usual phenotype to be inactivated. Because the sequence of the transposon is known, the gene can be identified, either by sequencing the whole genome and searching for the sequence, or by using the polymerase chain reaction to amplify specifically that gene. Virus insertional mutagenesis Because many viruses integrate their own genomes into the genomes of their host cells in order to replicate, mutagenesis caused by viral infections is a fairly common occurrence. Not all integrating viruses cause insertional mutagenesis, however. Some DNA insertions will lead to no noticeable mutation. Historically, lentiviral vectors included strong viral promoters which had a side effect of insertional mutagenesis, nuclear DNA mutations that effect the function of a gene. These strong viral promotors were shown to be the main cause of cancer formation. As a result, viral promotors have been replaced by cellular promotors and regulatory sequences. For those viruses such as gammaretroviruses that tend to integrate their DNA in genetically unfavorable locations, the severity of any ensuing mutation depends entirely on the location within the host's genome wherein the viral DNA is inserted. If the DNA is inserted into the middle of an essential gene, the effects on the cell will be drastic. Ad
https://en.wikipedia.org/wiki/Test%20data
Test data plays a crucial role in software development by providing inputs that are used to verify the correctness, performance, and reliability of software systems. Test data encompasses various types, such as positive and negative scenarios, edge cases, and realistic user scenarios, and it aims to exercise different aspects of the software to uncover bugs and validate its behavior. By designing and executing test cases with appropriate test data, developers can identify and rectify defects, improve the quality of the software, and ensure it meets the specified requirements. Moreover, test data can also be used for regression testing to validate that new code changes or enhancements do not introduce any unintended side effects or break existing functionalities. Overall, the effective utilization of test data in software development significantly contributes to the production of reliable and robust software systems. Background Some data may be used in a confirmatory way, typically to verify that a given set of inputs to a given function produces some expected result. Other data may be used in order to challenge the ability of the program to respond to unusual, extreme, exceptional, or unexpected input. Test data may be produced in a focused or systematic way (as is typically the case in domain testing), or by using other, less-focused approaches (as is typically the case in high-volume randomized automated tests). Test data may be produced by the tester, or by a program or function that aids the tester. Test data may be recorded for reuse or used only once. Test data can be created manually, by using data generation tools (often based on randomness), or be retrieved from an existing production environment. The data set can consist of synthetic (fake) data, but preferably it consists of representative (real) data. Limitations Due to privacy rules and regulations like GDPR, PCI and HIPAA it is not allowed to use privacy sensitive personal data for testing. But an
https://en.wikipedia.org/wiki/Deductive%20language
A deductive language is a computer programming language in which the program is a collection of predicates ('facts') and rules that connect them. Such a language is used to create knowledge based systems or expert systems which can deduce answers to problem sets by applying the rules to the facts they have been given. An example of a deductive language is Prolog, or its database-query cousin, Datalog. History As the name implies, deductive languages are rooted in the principles of deductive reasoning; making inferences based upon current knowledge. The first recommendation to use a clausal form of logic for representing computer programs was made by Cordell Green (1969) at Stanford Research Institute (now SRI International). This idea can also be linked back to the battle between procedural and declarative information representation in early artificial intelligence systems. Deductive languages and their use in logic programming can also be dated to the same year when Foster and Elcock introduced Absys, the first deductive/logical programming language. Shortly after, the first Prolog system was introduced in 1972 by Colmerauer through collaboration with Robert Kowalski. Components The components of a deductive language are a system of formal logic and a knowledge base upon which the logic is applied. Formal Logic Formal logic is the study of inference in regards to formal content. The distinguishing feature between formal and informal logic is that in the former case, the logical rule applied to the content is not specific to a situation. The laws hold regardless of a change in context. Although first-order logic is described in the example below to demonstrate the uses of a deductive language, no formal system is mandated and the use of a specific system is defined within the language rules or grammar. As input, a predicate takes any object(s) in the domain of interest and outputs either one of two Boolean values: true or false. For example, consider the se
https://en.wikipedia.org/wiki/Moore%20School%20Lectures
Theory and Techniques for Design of Electronic Digital Computers (popularly called the "Moore School Lectures") was a course in the construction of electronic digital computers held at the University of Pennsylvania's Moore School of Electrical Engineering between July 8, 1946, and August 30, 1946, and was the first time any computer topics had ever been taught to an assemblage of people. The course disseminated the ideas developed for the EDVAC (then being built at the Moore School as the successor computer to the ENIAC) and initiated an explosion of computer construction activity in the United States and internationally, especially in the United Kingdom. Background The Moore School in Philadelphia, Pennsylvania was at the center of developments in high-speed electronic computing in 1946. On February 14 of that year it had publicly unveiled the ENIAC, the first general-purpose electronic digital computer, developed in secret beginning in 1943 for the Army's Ballistics Research Laboratory. Prior even to the ENIAC's completion, work had begun on a second-generation electronic digital computer, the EDVAC, which incorporated the stored program model. Work at the Moore School attracted researchers including John von Neumann, who served as a consultant to the EDVAC project, and Stan Frankel and Nicholas Metropolis of the Manhattan Project, who arrived to run one of the first major programs written for the ENIAC, a mathematical simulation for the hydrogen bomb project. World War II had spawned major national efforts in many forms of scientific research—continued in peacetime—that required computationally intensive analysis; the thirst for information about the new Moore School computing machines had not been slaked, but instead intensified, by the distribution of von Neumann's notes on the EDVAC's logical design. Rather than allow themselves to be inundated with requests for demonstrations or slow progress in computer research by withholding the benefits of the Moo
https://en.wikipedia.org/wiki/Helper%20dependent%20virus
A helper dependent virus, also termed a gutless virus, is a synthetic viral vector dependent on the assistance of a helper virus in order to replicate, and can be used for purposes such as gene therapy. Naturally-occurring satellite viruses are also helper virus dependent, and can sometimes be modified to become viral vectors. Viral vector Since the genome of the gutless virus does not include genes encoding the enzymes and/or structural proteins required to replicate, it is deemed safe for use in gene therapy since an infection cannot occur except in the presence of a suitable helper virus. Well established protocols allow scientists to propagate helper dependent viruses in the lab. However, using an actual helper virus poses problems when it comes to purification of a desired transgenic virus. Therefore, lab methods often utilize minimal fragments of the helper DNA that can serve this purpose without creating unwanted virus. This process usually involves the introduction of three separate DNA plasmids into a eukaryotic cell line through a process called transfection. These plasmids contain either transgenic DNA or replication and capsid encoding DNA, plus helper DNA. Every cell that is successfully transfected with all three DNA fragments will produce the necessary proteins to produce infective viruses. These viruses will only have transgenic DNA encapsidated and therefore once they've infected a patient's cell, they will not be capable of reproducing. Satellite virus Helper dependent viruses can also occur in nature without being "gutted". The term satellite virus has been given to a large group of viruses that all require the presence of another virus to replicate. Many of these are plant viruses, but animal viruses can be seen in the case of dependoviruses. Within the family parvoviridae, the dependovirus genus was given a distinct classification due to their dependence on another virus. The most widely known dependovirus is adeno-associated virus (AAV)
https://en.wikipedia.org/wiki/Helper%20virus
A helper virus is a virus that allows an otherwise-deficient coinfecting virus to replicate. These can be naturally occurring as with Hepatitis D virus, which requires Hepatitis B virus to coinfect cells in order to replicate. Helper viruses are also commonly used to replicate and spread viral vectors for gene expression and gene therapy. See also Helper dependent virus Virophage
https://en.wikipedia.org/wiki/Arthur%20Burks
Arthur Walter Burks (October 13, 1915 – May 14, 2008) was an American mathematician who worked in the 1940s as a senior engineer on the project that contributed to the design of the ENIAC, the first general-purpose electronic digital computer. Decades later, Burks and his wife Alice Burks outlined their case for the subject matter of the ENIAC having been derived from John Vincent Atanasoff. Burks was also for several decades a faculty member at the University of Michigan in Ann Arbor. Early life and education Burks was born in Duluth, Minnesota. He earned his B.A. in mathematics and physics from DePauw University in Greencastle, Indiana, in 1936 and his M.A. and Ph.D. in philosophy from the University of Michigan in Ann Arbor in 1937 and 1941, respectively. The Moore School The summer after obtaining his Ph.D., Burks moved to Philadelphia, Pennsylvania, and enrolled in the national defense electronics course offered by the University of Pennsylvania's Moore School of Electrical Engineering; his laboratory teaching assistant was J. Presper Eckert, a graduate student at the Moore School; a fellow student was John Mauchly, the chairman of the physics department at Ursinus College in nearby Collegeville, Pennsylvania. Both Burks and Mauchly sought and obtained teaching positions at the Moore School the following fall, and roomed together throughout the academic year. ENIAC When Mauchly and Eckert's proposed concept for an electronic digital computer was funded by the U.S. Army's Ballistics Research Laboratory in June 1943, Burks was added to the design team. Among his principal contributions to the project was the design of the high-speed multiplier unit. (Also during this time, Burks met and married Alice Rowe, a human computer employed at the Moore School.) In April 1945, with John Grist Brainerd, Burks was charged with writing the technical reports on the ENIAC for publication. Also during 1945 Burks assisted with the preliminary logical design of the ED
https://en.wikipedia.org/wiki/Quantum%20%28video%20game%29
Quantum is a color vector arcade video game developed at General Computer Corporation for Atari, Inc. and released in December 1982. It was designed by Betty Ryan () who was the first female developer at GCC. The premise of the game is related loosely to quantum physics; the player directs a probe with a trackball to encircle atomic "particles" for points, without touching various other particles. Once the particles are surrounded by the probe's tail they are destroyed. Gameplay High score table To enter initials for a high score, the player uses the trackball to circle letters in the same fashion used during gameplay. If the player achieves the highest score on the table, the initials screen is preceded by another on which adept players can use the trackball to draw their initials. Legacy A screenshot of a clone called Tachyon was previewed in Atari 8-bit family magazine ANALOG Computing, but the game was never completed. A remake, Quantum Recharged, was released in August 2023 for the Nintendo Switch, PlayStation 4, PlayStation 5, Windows, Xbox One, and Xbox Series X/S. It is part of Atari SA's Recharged remake series. See also Disco No. 1 Libble Rabble
https://en.wikipedia.org/wiki/Abu%20Abed
Abu Abed is a fictional character that forms the centerpiece of many jokes in Lebanon, though he is known throughout the Arab world. The Washington Post describes him as an "Archie Bunker-like figure who is a fumbling caricature of all the failings of the Lebanese." His full name is sometimes given as Abu Abed El Beyrouty and he is also called Abul Abed or Abu El-Abed. In illustrations, Abu Abed's most notable features are a large mustache and the red fez he wears on his head. Abu Abed's best friend is Abu Steif, with whom he spends much of the day in the Kahwat El Ejeez قهوة القزاز, an actual and well-known coffee shop in central Beirut. He is sometimes claimed to be a Sunni Beiruti. One example of the literally hundreds of jokes with Abu Abed is: During the 2006 Israel-Lebanon conflict, the most widely told joke among Lebanese was again about Abu Abed. The jokes goes: Abu Abed is sitting in the cafe when he calls Ehud Olmert, Prime Minister of Israel, to tell him not to come north of the border or he and four of his friends will give Israel trouble. Olmert laughs and tells Abu Abed that one Israeli battalion can easily overrun his neighborhood. This verbal contest escalates until Abu Abed says that he has collected thousands of fighters armed with shoulder-fired rockets and Olmert states that Israel has two million soldiers. "'Two million?' asks Abul Abed. 'In that case I am going to have to surrender. We simply do not have enough room to keep 2 million hostages.'" Notes and references External links Clean Abu El Abed Jokes, collected by Abdallah Hayar Official Abou El Abed web site Humor and wit characters Lebanese culture Fictional Lebanese people National personifications
https://en.wikipedia.org/wiki/OFFSystem
The Owner-Free File System (OFF System, or OFFS for short) is a peer-to-peer distributed file system in which all shared files are represented by randomized multi-used data blocks. Instead of anonymizing the network, the data blocks are anonymized and therefore, only data garbage is ever exchanged and stored and no forwarding via intermediate nodes is required. OFFS claims to have been created with the expressed intention "to cut off some gangrene-infested bits of the copyright industry." History OFFS development started within the hacktivism group The Big Hack in 2003 by the hackers Cheater512, CaptainMorgan, Aqlo and WhiteRaven. In 2004, a rudimentary version was finished, written in PHP, which was distributed as two demo CDs. Following these, SpectralMorning re-implemented the functionality in 2004 in C++, which led to the current "mainline" OFFS client. On August 14, 2006, CaptainMorgan posted a letter of "closing" addressed to the "Copyright Industry Associations of America", such as the RIAA and MPAA, stating that they have created OFFS with the purpose of ending "all of your problems with consumer copyright infringement." In 2008, the network consisted of around 50 nodes. On April 11, 2008, a beta test was held with a network size of over 100 nodes. Since SpectralMorning stopped work on OFFS in late 2008, only minor bug fix releases were made to mainline OFF. Starting from 2007, an alternative, but compatible client was developed, called BlocksNet. Written in Ruby and well-maintained, it saw major improvements over recent time. It has been under development until 2011. The client OFFLoad is a fork from mainline OFFS, which seemingly adds no features. Reasons for the fork are unclear. Another distantly related program is Monolith, which uses a similar principle to OFFS. It was created after OFFS and features no multi-use of blocks and no networking. Functional Principle The OFF System is a kind of anonymous, fully decentralized P2P file sharing program a
https://en.wikipedia.org/wiki/SoundEdit
SoundEdit was the first popular GUI-based audio editor for digitized audio. It was one of the first significant audio applications for personal computers in general. SoundEdit was known for its ease of use. It made audio manipulation accessible to those who were not musicians or audio professionals. History SoundEdit was one of three audio applications created during a sabbatical by Steve Capps during 1986. The Macintosh had no built-in sound input, so the MacRecorder audio digitizer was invented for this purpose in 1985 by Michael Lamoureux, a mathematics student at the University of California, Berkeley. The MacRecorder hardware and software was publicly released through the Berkeley Macintosh Users Group as a kit in late 1985. SoundEdit first shipped in January 1988, as part of a hardware product called MacRecorder Sound System, by a company called Farallon Computing (which eventually became Netopia). The default sampling rate in SoundEdit was 22,256 Hz. One of the major drivers for SoundEdit was Apple's HyperCard. With MacRecorder Sound System, stack makers could finally create alternatives to HyperCard's two built-in sounds. The other multimedia programs of the time, (Director and Authorware) also adopted it right away. In 1991, SoundEdit was bought by Macromind-Paracomp, which became Macromedia (now Adobe Systems). SoundEdit 16 Macromedia rebranded the program as 'SoundEdit 16' and expanded its capabilities to support CD-quality stereo audio (16-bit) and added QuickTime soundtrack editing and an audio plug-in architecture. The new version also supported handling of sound files bigger than available RAM. A free plug-in was included to convert sound files into '.SWA' format, an early form of mp3 file. These SWA files could be streamed to web viewers using by the free web audio player (downloadable from Macromedia) or by creating a custom interactive Shockwave experience embedded into a web page. Shockwave content is created by using the Director multim
https://en.wikipedia.org/wiki/Wabi%20%28software%29
Wabi is a discontinued commercial software application from Sun Microsystems that implements the Windows Win16 API specification on Solaris and AIX; a version for Linux was also released by Caldera Systems. Wabi runs applications developed for Windows 3.1, Windows 3.11, and Windows for Workgroups. History The technology was originally developed by Praxsys Technologies as the result of discussions in 1990 with Interactive Systems Corporation. The assets of Praxsys were acquired by Sun in the fall of 1992. The name "Wabi" was chosen for two reasons: its meaning in Japanese of balance or harmony, which conjured the notion of a more peaceful coexistence between Windows and Unix software; and, the more obvious implication of it standing for Windows Application Binary Interface, although before its release Sun declared that the name was not an acronym. Wabi 2.2B was licensed by Caldera for Linux, together with the also licensed Merge. Wabi development was discontinued in December 1997. Features Wabi requires a licensed Windows 3.x installation, unlike similar software that endeavors to implement the entire Windows API, such as Wine or select versions of OS/2. Wabi implements the lowest layers of the Windows environment in the form of the user.dll, kernel.dll, and gdi.dll libraries. All other Windows dlls depend on these three modules, so cloning this functionality allows Windows software to execute correctly on a foreign host system. This approach, as opposed to a full replacement, was thought by the engineering team to be the only rational methodology for success given both the size of Microsoft's ever-expanding efforts and the difficulties of the emulation being precise enough to run commercial software. Wabi was released for x86 and SPARC systems, and on PowerPC for AIX. To run an x86 Windows environment on SPARC systems, a code translation layer dynamically converts x86 instructions upon first use into SPARC instructions. Standardization attempt In conjunction w
https://en.wikipedia.org/wiki/Sicilicus
A sicilicus was an old Latin diacritical mark, , like a reversed C (Ɔ) placed above a letter and evidently deriving its name from its shape like a little sickle (which is sicilis in Latin). The ancient sources say that during the time of the Republic it was placed above a geminate consonant to indicate that the consonant counted twice, although there is hardly any epigraphic or paleographic evidence available from such an early time. When such geminate consonants began to be represented during classical times by writing the letter twice, the sicilicus naturally fell into disuse in this function, but continued to be used to indicate the doubling of vowels as an indication of length, in the developed form of the apex. Fontaine suggests that Plautus alludes to the sicilicus in the prologue to Menaechmi. See also Open O, although this is a full letter, and not a diacritic placed above a letter Antisigma, although this is a full letter, and not a diacritic placed above a letter Apex (diacritic), used for long vowels instead of long consonants Apostrophe, whose shape is derived from it Comma (punctuation), whose shape is similar Latin spelling and pronunciation
https://en.wikipedia.org/wiki/Isoaspartate
Isoaspartic acid (isoaspartate, isoaspartyl, β-aspartate) is an aspartic acid residue isomeric to the typical α peptide linkage. It is a β-amino acid, with the side chain carboxyl moved to the backbone. Such a change is caused by a chemical reaction in which the nitrogen atom on the N+1 following peptide bond (in black at top right of Figure 1) nucleophilically attacks the γ-carbon of the side chain of an asparagine or aspartic acid residue, forming a succinimide intermediate (in red). Hydrolysis of the intermediate results in two products, either aspartic acid (in black at left) or isoaspartic acid, which is a β-amino acid (in green at bottom right). The reaction also results in the deamidation of the asparagine residue. Racemization may occur leading to the formation of D-aminoacids. Kinetics of isoaspartyl formation Isoaspartyl formation reactions have been conjectured to be one of the factors that limit the useful lifetime of proteins. Isoaspartyl formation proceeds much more quickly if the asparagine is followed by a small, flexible residue (such as Gly) that leaves the peptide group open for attack. These reactions also proceed much more quickly at elevated pH (>10) and temperatures. Repair L-isoaspartyl methyltransferase repairs isoaspartate and D-aspartate residues by sticking a methyl group onto the side chain carboxyl group in the residue, creating an ester. The ester rapidly and spontaneously turns into the succinimide (red), and randomly turns back into normal aspartic acid (black) or isoaspartate again (green) for another attempt.
https://en.wikipedia.org/wiki/Telelogic
Telelogic AB was a software business headquartered in Malmö, Sweden. Telelogic was founded in 1983 as a research and development arm of Televerket, the Swedish department of telecom (now part of TeliaSonera). It was later acquired by IBM Rational, and exists under the IBM software group. Telelogic had operations in 22 countries and had been publicly traded since 1999. CEO and President in 2001 was Anders Lidbeck. On June 11, 2007, IBM announced that it had made a cash offer to acquire Telelogic. On August 29, 2007, the European Union opened an investigation into the acquisition. On March 5, 2008, European regulators approved the acquisition of Telelogic by the Swedish IBM subsidiary Watchtower AB. On April 28, 2008, IBM completed its purchase of Telelogic. Former Products Focal Point — System for management of product and project portfolios. DOORS — Requirements tracking tool. System Architect — Enterprise Architecture and Business Architecture modeling tool. Tau — SDL and UML modeling tool. Synergy — Task-based version control and configuration management system. Rhapsody — Systems engineering and executable UML modeling tool. DocExpress — Technical documentation tool, discontinued after the acquisition and superseded by Publishing Engine. Publishing Engine — Technical documentation tool All of these products have been continued under IBM's Rational Software division in the systems engineering and Product lifecycle management (PLM) "solutions" software line. IBM sold System Architect, Focal Point and several other software products to UNICOM Global in 2016. Acquisitions Telelogic acquired the following companies between 1999 and 2007:
https://en.wikipedia.org/wiki/Bullet%20Cluster
The Bullet Cluster (1E 0657-56) consists of two colliding clusters of galaxies. Strictly speaking, the name Bullet Cluster refers to the smaller subcluster, moving away from the larger one. It is at a comoving radial distance of . The object is of a particular note for astrophysicists, because gravitational lensing studies of the Bullet Cluster are claimed to provide the best evidence to date for the existence of dark matter. Observations of other galaxy cluster collisions, such as MACS J0025.4-1222, similarly support the existence of dark matter. Overview The major components of the cluster pair—stars, gas and the putative dark matter—behave differently during collision, allowing them to be studied separately. The stars of the galaxies, observable in visible light, were not greatly affected by the collision, and most passed right through, gravitationally slowed but not otherwise altered. The hot gas of the two colliding components, seen in X-rays, represents most of the baryonic, or "ordinary", matter in the cluster pair. The gases of the Intracluster medium interact electromagnetically, causing the gases of both clusters to slow much more than the stars. The third component, the dark matter, was detected indirectly by the gravitational lensing of background objects. In theories without dark matter, such as Modified Newtonian dynamics (MOND), the lensing would be expected to follow the baryonic matter; i.e. the X-ray gas. However, the lensing is strongest in two separated regions near (possibly coincident with) the visible galaxies. This provides support for the idea that most of the gravitation in the cluster pair is in the form of two regions of dark matter, which bypassed the gas regions during the collision. This accords with predictions of dark matter as only gravitationally interacting, other than weakly interacting. The Bullet Cluster is one of the hottest-known clusters of galaxies. It provides an observable constraint for cosmological models, which
https://en.wikipedia.org/wiki/EF%20hand
The EF hand is a helix–loop–helix structural domain or motif found in a large family of calcium-binding proteins. The EF-hand motif contains a helix–loop–helix topology, much like the spread thumb and forefinger of the human hand, in which the Ca2+ ions are coordinated by ligands within the loop. The motif takes its name from traditional nomenclature used in describing the protein parvalbumin, which contains three such motifs and is probably involved in muscle relaxation via its calcium-binding activity. The EF-hand consists of two alpha helices linked by a short loop region (usually about 12 amino acids) that usually binds calcium ions. EF-hands also appear in each structural domain of the signaling protein calmodulin and in the muscle protein troponin-C. Calcium ion binding site The calcium ion is coordinated in a pentagonal bipyramidal configuration. The six residues involved in the binding are in positions 1, 3, 5, 7, 9 and 12; these residues are denoted by X, Y, Z, -Y, -X and -Z. The invariant Glu or Asp at position 12 provides two oxygens for liganding calcium (bidentate ligand). The calcium ion is bound by both protein backbone atoms and by amino acid side chains, specifically those of the anionic amino acid residues aspartate and glutamate. These residues are negatively charged and will make a charge-interaction with the positively charged calcium ion. The EF hand motif was among the first structural motifs whose sequence requirements were analyzed in detail. Five of the loop residues bind calcium and thus have a strong preference for oxygen-containing side chains, especially aspartate and glutamate. The sixth residue in the loop is necessarily glycine due to the conformational requirements of the backbone. The remaining residues are typically hydrophobic and form a hydrophobic core that binds and stabilizes the two helices. Upon binding to Ca2+, this motif may undergo conformational changes that enable Ca2+-regulated functions as seen in Ca2+ effecto
https://en.wikipedia.org/wiki/Badlands%20%281984%20video%20game%29
is a video game released and developed by Konami on the LaserDisc system in 1984 and published by both Konami and Centuri. It first debuted at the Amusement and Music Operators Association (AMOA) Show in October 1983 and was later released to the public in early 1984. In addition to its LaserDisc version, two versions of a Badlands video game cabinet exist, one produced by Konami, and one by Centuri. Badlands follows a cowboy named Buck seeking vengeance on a gang of outlaws and its leader, Landolf, for the murder of his wife and children. Gameplay Badlands is a first-person shooter action-adventure video game set in a wild west fantasy world. The game's arcade cabinet consists of one large "shoot" button. Badlands' gameplay consists of animated cutscenes, requiring players to shoot and react to environmental hazards and enemies. The game uses a life system, granting the player three lives upon starting. Losing all lives ends the game. The aim of the game is to eliminate outlaws and claim their bounties. Reception In Japan, Game Machine listed Badlands on their September 15, 1984, issue as the second most-successful upright arcade unit of the month.