source
stringlengths 31
227
| text
stringlengths 9
2k
|
---|---|
https://en.wikipedia.org/wiki/Atomic%20and%20molecular%20astrophysics
|
Atomic astrophysics is concerned with performing atomic physics calculations that will be useful to astronomers and using atomic data to interpret astronomical observations. Atomic physics plays a key role in astrophysics as astronomers' only information about a particular object comes through the light that it emits, and this light arises through atomic transitions.
Molecular astrophysics, developed into a rigorous field of investigation by theoretical astrochemist Alexander Dalgarno beginning in 1967, concerns the study of emission from molecules in space. There are 110 currently known interstellar molecules. These molecules have large numbers of observable transitions. Lines may also be observed in absorption—for example the highly redshifted lines seen against the gravitationally lensed quasar PKS1830-211. High energy radiation, such as ultraviolet light, can break the molecular bonds which hold atoms in molecules. In general then, molecules are found in cool astrophysical environments. The most massive objects in our galaxy are giant clouds of molecules and dust known as giant molecular clouds. In these clouds, and smaller versions of them, stars and planets are formed. One of the primary fields of study of molecular astrophysics is star and planet formation. Molecules may be found in many environments, however, from stellar atmospheres to those of planetary satellites. Most of these locations are relatively cool, and molecular emission is most easily studied via photons emitted when the molecules make transitions between low rotational energy states. One molecule, composed of the abundant carbon and oxygen atoms, and very stable against dissociation into atoms, is carbon monoxide (CO). The wavelength of the photon emitted when the CO molecule falls from its lowest excited state to its zero energy, or ground, state is 2.6mm, or 115 gigahertz. This frequency is a thousand times higher than typical FM radio frequencies. At these high frequencies, molecules in th
|
https://en.wikipedia.org/wiki/Patient%20safety%20organization
|
A Patient Safety Organization (PSO) is a group, institution, or association that improves medical care by reducing medical errors. Common functions of patient safety organizations are data collection, analysis, reporting, education, funding, and advocacy. A PSO differs from a Federally designed Patient Safety Organization (PSO), which provides health care providers in the U.S. privilege and confidentiality protections for efforts to improve patient safety and the quality of patient care delivery (see 42 U.S.C. 299b-21 et seq. and www.PSO.AHRQ.gov.)
In the 1990s, reports in several countries revealed a staggering number of patient injuries and deaths each year due to avoidable errors and deficiencies in health care, among them adverse events and complications arising from poor infection control. In the United States, a 1999 report from the Institute of Medicine called for a broad national effort to prevent these events, including the establishment of patient safety centers, expanded reporting of adverse events, and development of safety programs in healthcare organizations. Although many PSOs are funded and run by governments, others have sprung from private entities such as industry, professional and consumer groups.
Functions
To achieve their goals, patient safety organizations may
Collect data on the prevalence and individual details of errors.
Analyze sources of error by root cause analysis.
Propose and disseminate methods for error prevention.
Design and conduct pilot projects to study safety initiatives, including monitoring of results.
Raise awareness and inform the public, health professionals, providers, purchasers, and employers.
Conduct fundraising and provide funding for research and safety projects.
Advocate for regulatory and legislative changes.
Governmental organizations
World Health Organization
World Alliance for Patient Safety
In response to a 2002 World Health Assembly Resolution, the World Health Organization (WHO) launched the Worl
|
https://en.wikipedia.org/wiki/Secondary%20palate%20development
|
The development of the secondary palate commences in the sixth week of human embryonic development. It is characterised by the formation of two palatal shelves on the maxillary prominences, the elevation of these shelves to a horizontal position, and then a process of palatal fusion between the horizontal shelves. The shelves will also fuse anteriorly upon the primary palate, with the incisive foramen being the landmark between the primary palate and secondary palate. This forms what is known as the roof of the mouth, or the hard palate.
The formation and development of the secondary palate occurs through signalling molecules SHH, BMP-2, FGF-8, among others.
Failure of the secondary palate to develop correctly may result in a cleft palate disorder.
Formation of palatal shelves
The formation of the vertical palatal shelves occurs during week 7 of embryological development, on the maxillary processes of the head of the embryo, lateral to the developing tongue.
Palatal shelf elevation
The elevation of the palatal shelves from a vertical position to a horizontal one occurs during week 8 of embryological development. The direct cause of this movement is unknown, but a number of possibilities have been identified as follows:
Muscular contraction;
Hydrostatic forces exerted by glycosaminoglycans and hyaluronan;
Mesenchymal reorganisation;
Mesenchyme cell contraction;
Epithelial reorganisation;
Movement of the developing tongue.
Suggested mechanisms for palatal fusion
Fusion between the two palatal shelves occurs during week 9 of embryonic development. In this time, the elevated palatal shelves join together to form one continuous structure, with the medial edge epithelium (the shelf surfaces which are closest to each other) disappearing.
The specific mechanism by which the medial edge epithelium disappears has been differed over by academics. The three most distinguished theories related to the explanation of palatal fusion are as follows:
Epithelial apoptosi
|
https://en.wikipedia.org/wiki/PikeOS
|
PikeOS is a commercial, hard real-time operating system (RTOS) that offers a separation kernel based hypervisor with multiple logical partition types for many other operating systems (OS), each called a GuestOS, and applications. It enables users to build certifiable smart devices for the Internet of things (IoT) according to the high quality, safety and security standards of different industries. For safety and security, critical real-time applications on controller-based systems without memory management unit (MMU) but with memory protection unit (MPU) PikeOS for MPU is available.
Overview
PikeOS was introduced in 2005 and combines a real-time operating system (RTOS) with a virtualization platform and Eclipse-based integrated development environment (IDE) for embedded systems. It is a commercial clone of L4 microkernel family. PikeOS has been developed for safety and security-critical applications with certification needs in the fields of aerospace, defense, automotive, transport, industrial automation, medical, network infrastructures, and consumer electronics. The PikeOS separation kernel (v5.1.3) is certified against Common Criteria at EAL5+.
A key feature of PikeOS is an ability to safely execute applications with different safety and security levels concurrently on the same computing platform. This is done by strict spatial and temporal segregation of these applications via software partitions. A software partition can be seen as a container with pre-allocated privileges that can have access to memory, central processing unit (CPU) time, input/output (I/O), and a predefined list of OS services. With PikeOS, the term application refers to an executable linked against the PikeOS application programming interface (API) library and running as a process inside a partition. The nature of the PikeOS application programming interface (API) allows applications to range from simple control loops up to full paravirtualized guest operating systems like Linux or hardwar
|
https://en.wikipedia.org/wiki/Office%20of%20the%20Chief%20Actuary
|
The Office of the Chief Actuary is a government agency that has responsibility for actuarial estimates regarding social welfare programs. In Canada, the Office of the Chief Actuary works with the Canada Pension Plan and the Old Age Security Program. In the United States, the Social Security Administration has an Office of the Chief Actuary that deals with Social Security, and the Centers for Medicare and Medicaid Services have an Office of the Actuary that deals with Medicare and Medicaid. A similar agency in the United Kingdom is called the Government Actuary's Department (GAD).
United States
Social Security Administration
In the U.S., the Office of the Chief Actuary at the Social Security Administration plans and directs a program of actuarial estimates and analyses relating to SSA-administered retirement, survivors and disability insurance programs and to proposed changes in those programs. It evaluates operations of the Federal Old-Age and Survivors Insurance Trust Fund and the Federal Disability Insurance Trust Fund, conducts studies of program financing, performs actuarial and demographic research on social insurance and related program issues, and projects future workloads.
In addition, the office is charged with conducting cost analyses relating to the Supplemental Security Income (SSI) program, a general-revenue financed, means-tested program for low-income aged, blind and disabled people. The Office provides technical and consultative services to the Commissioner, to the board of trustees of the Social Security Trust Funds, and its staff appears before Congressional Committees to provide expert testimony on the actuarial aspects of Social Security issues.
, the Chief Actuary of the Social Security Administration is Stephen Goss.
|
https://en.wikipedia.org/wiki/Hemerocallis%20fulva
|
Hemerocallis fulva, the orange day-lily, tawny daylily, corn lily, tiger daylily, fulvous daylily, ditch lily or Fourth of July lily (also railroad daylily, roadside daylily, outhouse lily, and wash-house lily), is a species of daylily native to Asia. It is very widely grown as an ornamental plant in temperate climates for its showy flowers and ease of cultivation. It is not a true lily in the genus Lilium, but gets its common name from the superficial similarity of its flowers to Lilium and from the fact that each flower lasts only one day.
Description
It is an herbaceous perennial plant growing from tuberous roots, with stems tall. The leaves are linear, long and broad. The flowers are across, orange-red, with a pale central line on each tepal; they are produced from early summer through late autumn on scapes of ten through twenty flowers, with the individual flowers opening successively, each one lasting only one day. Its fruit is a three-valved capsule long and broad which splits open at maturity and releases seeds.
Both diploid and triploid forms occur in the wild, but most cultivated plants are triploids which rarely produce seeds and primarily reproduce vegetatively by stolons. At least four botanical varieties are recognized, including the typical triploid var. fulva, the diploid, long-flowered var. angustifolia (syn.: var. longituba), the triploid var. Flore Pleno, which has petaloid stamens, and the evergreen var. aurantiaca.
Distribution
Orange daylily is native to Asia from the Caucasus east through the Himalaya through China, Japan, and Korea. Orange daylily persists where planted, making them a very good garden plant.
Hemerocallis fulva var. fulva has escaped from cultivation across much of the United States and parts of Canada and has become a weedy or invasive species. It persists also where dumped and spreads more or less rapidly by vegetative increase into woods and fields and along roadsides and ditches, hence its common name ditch lily.
|
https://en.wikipedia.org/wiki/SuperGrid%20%28hydrogen%29
|
In lossless power transmission, a supergrid with hydrogen is an idea for combining very long distance electric power transmission with liquid hydrogen distribution, to achieve superconductivity in the cables. The hydrogen is both a distributed fuel and a cryogenic coolant for the power lines, rendering them superconducting. The concept's advocates describe it as being in a "visionary" stage, for which no new scientific breakthrough is required but which requires major technological innovations before it could progress to a practical system. A system for the United States is projected to require "several decades" before it could be fully implemented.
One proposed design for a superconducting cable includes a superconducting bipolar DC line operating at ±50 kV, and 50 kA, transmitting about 2.5 GW for several hundred kilometers at zero resistance and nearly no line loss. High-voltage direct current (HVDC) lines have the capability of transmitting similar wattages, for example a 5 gigawatt HVDC system is being constructed along the southern provinces of China without the use of superconducting cables.
In the United States, a Continental SuperGrid 4,000 kilometers long might carry 40,000 to 80,000 MW in a tunnel shared with long distance high speed maglev trains, which at low pressure could allow cross continental journeys of one hour. The liquid hydrogen pipeline would both store and deliver hydrogen.
1.5% of the energy transmitted on the British AC Supergrid is lost (transformer, heating and capacitive losses). Of this, a little under two-thirds (or 1% on the British supergrid), represents "DC" (resistive) heating type losses. With superconductive power lines, the capacitive and transformer losses (in the unlikely event the transmission lines were still overhead AC lines) would remain the same. In addition, overhead lines do not lend themselves at all well physically to the incorporation of cryogenic hydrogen piping, due to the likely weight of the transmission med
|
https://en.wikipedia.org/wiki/CDC%20Kronos
|
Kronos is an operating system with time-sharing capabilities, written by Control Data Corporation in 1971. Kronos ran on the 60-bit CDC 6000 series mainframe computers and their successors. CDC replaced Kronos with the NOS operating system in the late 1970s, which were succeeded by the NOS/VE operating system in the mid-1980s.
The MACE operating system and APEX were forerunners to KRONOS. It was written by Control Data systems programmer Greg Mansfield, Dave Cahlander, Bob Tate and three others.
See also
CDC SCOPE
|
https://en.wikipedia.org/wiki/Bessel%E2%80%93Clifford%20function
|
In mathematical analysis, the Bessel–Clifford function, named after Friedrich Bessel and William Kingdon Clifford, is an entire function of two complex variables that can be used to provide an alternative development of the theory of Bessel functions. If
is the entire function defined by means of the reciprocal gamma function, then the Bessel–Clifford function is defined by the series
The ratio of successive terms is z/k(n + k), which for all values of z and n tends to zero with increasing k. By the ratio test, this series converges absolutely for all z and n, and uniformly for all regions with bounded |z|, and hence the Bessel–Clifford function is an entire function of the two complex variables n and z.
Differential equation of the Bessel–Clifford function
It follows from the above series on differentiating with respect to x that satisfies the linear second-order homogeneous differential equation
This equation is of generalized hypergeometric type, and in fact the Bessel–Clifford function is up to a scaling factor a Pochhammer–Barnes hypergeometric function; we have
Unless n is a negative integer, in which case the right-hand side is undefined, the two definitions are essentially equivalent; the hypergeometric function being normalized so that its value at z = 0 is one.
Relation to Bessel functions
The Bessel function of the first kind can be defined in terms of the Bessel–Clifford function as
when n is not an integer we can see from this that the Bessel function is not entire. Similarly, the modified Bessel function of the first kind can be defined as
The procedure can of course be reversed, so that we may define the Bessel–Clifford function as
but from this starting point we would then need to show was entire.
Recurrence relation
From the defining series, it follows immediately that
Using this, we may rewrite the differential equation for as
which defines the recurrence relationship for the Bessel–Clifford function. This is equivalent to a s
|
https://en.wikipedia.org/wiki/Degree%20of%20a%20polynomial
|
In mathematics, the degree of a polynomial is the highest of the degrees of the polynomial's monomials (individual terms) with non-zero coefficients. The degree of a term is the sum of the exponents of the variables that appear in it, and thus is a non-negative integer. For a univariate polynomial, the degree of the polynomial is simply the highest exponent occurring in the polynomial. The term order has been used as a synonym of degree but, nowadays, may refer to several other concepts (see Order of a polynomial (disambiguation)).
For example, the polynomial which can also be written as has three terms. The first term has a degree of 5 (the sum of the powers 2 and 3), the second term has a degree of 1, and the last term has a degree of 0. Therefore, the polynomial has a degree of 5, which is the highest degree of any term.
To determine the degree of a polynomial that is not in standard form, such as , one can put it in standard form by expanding the products (by distributivity) and combining the like terms; for example, is of degree 1, even though each summand has degree 2. However, this is not needed when the polynomial is written as a product of polynomials in standard form, because the degree of a product is the sum of the degrees of the factors.
Names of polynomials by degree
The following names are assigned to polynomials according to their degree:
Special case – zero (see , below)
Degree 0 – non-zero constant
Degree 1 – linear
Degree 2 – quadratic
Degree 3 – cubic
Degree 4 – quartic (or, if all terms have even degree, biquadratic)
Degree 5 – quintic
Degree 6 – sextic (or, less commonly, hexic)
Degree 7 – septic (or, less commonly, heptic)
Degree 8 – octic
Degree 9 – nonic
Degree 10 – decic
Names for degree above three are based on Latin ordinal numbers, and end in -ic. This should be distinguished from the names used for the number of variables, the arity, which are based on Latin distributive numbers, and end in -ary. For example, a degre
|
https://en.wikipedia.org/wiki/Ribosome%20recycling%20factor
|
Ribosome recycling factor or ribosome release factor (RRF) is a protein found in bacterial cells as well as eukaryotic organelles, specifically mitochondria and chloroplasts. It functions to recycle ribosomes after completion of protein synthesis (bacterial translation). In humans, the mitochrondrial version is coded by the MRRF gene.
Discovery
The ribosome recycling factor was discovered in the early 1970s by the work of Akira Kaji and Akikazu Hiroshima at the University of Pennsylvania. Their work described the requirement for two protein factors to release ribosomes from mRNA. These two factors were identified as RRF, an unknown protein until then, and Elongation Factor G (EF-G), a protein already identified and known to function in protein synthesis. RRF was originally called Ribosome Releasing Factor but is now called Ribosome Recycling Factor.
Function
RRF accomplishes the recycling of ribosomes by splitting ribosomes into subunits, thereby releasing the bound mRNA. This also requires the participation of EF-G (GFM2 in humans). Depending on the tRNA, IF1–IF3 may also perform recycling.
Loss of RRF function
In Bacteria (specifically Escherichia coli), loss of the gene encoding RRF is deleterious. This makes RRF a possible target for new antibacterial drugs.
Yeast mitochondrial RRF (mtRRF) is encoded by a gene in the cell nucleus. Loss of function of this gene leads to mitochondrial genome instability and respiratory incompetence.
Structure and binding to ribosomes
The crystal structure of RRF was first determined by X-ray diffraction in 1999. The most striking revelation was that RRF is a near-perfect structural mimic of tRNA, in both size and dimensions. One view of RRF can be seen here.
Despite the tRNA-mimicry, RRF binds to ribosomes quite differently from the way tRNA does. It has been suggested that ribosomes bind proteins (or protein domain) of similar shape and size to tRNA, and this, rather than function, explains the observed structural mimicr
|
https://en.wikipedia.org/wiki/Category%20of%20relations
|
In mathematics, the category Rel has the class of sets as objects and binary relations as morphisms.
A morphism (or arrow) R : A → B in this category is a relation between the sets A and B, so .
The composition of two relations R: A → B and S: B → C is given by
(a, c) ∈ S o R ⇔ for some b ∈ B, (a, b) ∈ R and (b, c) ∈ S.
Rel has also been called the "category of correspondences of sets".
Properties
The category Rel has the category of sets Set as a (wide) subcategory, where the arrow in Set corresponds to the relation defined by .
A morphism in Rel is a relation, and the corresponding morphism in the opposite category to Rel has arrows reversed, so it is the converse relation. Thus Rel contains its opposite and is self-dual.
The involution represented by taking the converse relation provides the dagger to make Rel a dagger category.
The category has two functors into itself given by the hom functor: A binary relation R ⊆ A × B and its transpose RT ⊆ B × A may be composed either as R RT or as RT R. The first composition results in a homogeneous relation on A and the second is on B. Since the images of these hom functors are in Rel itself, in this case hom is an internal hom functor. With its internal hom functor, Rel is a closed category, and furthermore a dagger compact category.
The category Rel can be obtained from the category Set as the Kleisli category for the monad whose functor corresponds to power set, interpreted as a covariant functor.
Perhaps a bit surprising at first sight is the fact that product in Rel is given by the disjoint union (rather than the cartesian product as it is in Set), and so is the coproduct.
Rel is monoidal closed, if one defines both the monoidal product A ⊗ B and the internal hom A ⇒ B by the cartesian product of sets. It is also a monoidal category if one defines the monoidal product by the disjoint union of sets.
The category Rel was the prototype for the algebraic structure called an allegory by Peter J. Freyd and An
|
https://en.wikipedia.org/wiki/277%20%28number%29
|
277 (two hundred [and] seventy-seven) is the natural number following 276 and preceding 278.
Mathematical properties
277 is the 59th prime number, and is a regular prime.
It is the smallest prime p such that the sum of the inverses of the primes up to p is greater than two.
Since 59 is itself prime, 277 is a super-prime. 59 is also a super-prime (it is the 17th prime), as is 17 (the 7th prime). However, 7 is the fourth prime number, and 4 is not prime. Thus, 277 is a super-super-super-prime but not a super-super-super-super-prime. It is the largest prime factor of the Euclid number 510511 = 2 × 3 × 5 × 7 × 11 × 13 × 17 + 1.
As a member of the lazy caterer's sequence, 277 counts the maximum number of pieces obtained by slicing a pancake with 23 straight cuts.
277 is also a Perrin number, and as such counts the number of maximal independent sets in an icosagon. There are 277 ways to tile a 3 × 8 rectangle with integer-sided squares, and 277 degree-7 monic polynomials with integer coefficients and all roots in the unit disk.
On an infinite chessboard, there are 277 squares that a knight can reach from a given starting position in exactly six moves.
277 appears as the numerator of the fifth term of the Taylor series for the secant function:
Since no number added to the sum of its digits generates 277, it is a self number. The next prime self number is not reached until 367.
|
https://en.wikipedia.org/wiki/List%20of%20wireless%20community%20networks%20by%20region
|
Megasurf wireless internet os covering the area.
Africa
DRC
Mesh Bukavu
Pamoja Net
Ghana
Akwapim Community Wireless Network
Somalia
Abaarso
South Africa
Rural Telehealth
Orange Farm:
Zenzeleni:
Home of Compassion
Tanzania
Sengerema Wireless Community Network
Kenya
TunapandaNET Community Network
Tunisia
Mesh SAYADA
Asia
Nepal Wireless Networking Project
Middle East
Oceania
Australia
Air Stream Wireless
Melbourne Wireless
TasWireless, Tasmania
Europe
Austria
FunkFeuer
Belgium
Bulgaria
Croatia
Cyprus
Czechia
Denmark
Estonia
Finland
France
Clermont Sans Fil
Germany
Freifunk
Greece
Athens Wireless Metropolitan Network
Wireless Thessaloniki
Sarantaporo.gr Wireless Community Network
Hungary
Ireland
CRCWN Cavan Rural Community Wireless Network. Free Wi-Fi and Fixed Wireless Internet.
(www.crcwn.online)
Italy
ninux
Progetto Neco
Latvia
Lithuania
Luxembourg
Malta
Netherlands
Wireless Leiden
Poland
Hyperboria PL (hyperboria.net.pl)
Portugal
Wirelesspt
Romania
Slovakia
Slovenia
wlan slovenija
Spain
guifi.net
Sweden
Pjodd
United Kingdom
Southampton Open Wireless Network
Americas
North America
Canada
Nova Scotia
Chebucto Community Network, Halifax Regional Municipality
Ontario
Toronto Mesh
Wireless Toronto
Wireless Nomad, Ontario and Toronto
Québec
Île Sans Fil, Montreal
ZAP Sherbrooke, Sherbrooke
Cuba
SNET (abbreviation of "Street Network"), nationwide underground community network
United States
Arizona
Tucson Mesh, Tucson
California
People's Open Network, Oakland
Illinois
Champaign-Urbana Community Wireless Network
Minnesota
Minneapolis wireless internet network
New York
NYC Mesh
Red Hook Wi-Fi
Oregon
Personal Telco, Portland, OR
Vermont
Newport Wireless Mesh, Newport City, Vermont
Washington
Seattle Wireless
West Virginia
West Virginia Broadband
South America
Argentina
Altermundi
Brazil
Coolab
See also
Computer network
Metropolitan area network
Wireless community ne
|
https://en.wikipedia.org/wiki/Not%20Just%20Another%20Bogus%20List
|
Not Just Another Bogus List (NJABL) was a DNS blacklist.
NJABL maintained a list of known and potential spam sources (open mail relays, open proxies, open form to mail HTTP gateways, dynamic IP pools, and direct spammers) for the purpose of being able to tag or refuse e-mail and thereby block spam from certain sources. NJABL automatically retests only listed open relays every 90 days.
The Open Proxy IPs portion (only) of NJABL data was used in Spamhaus XBL list
NJABL's dynamic IP list originally came from Dynablock but was maintained independently since Dynablock stopped updating December 2003. The SORBS dynamic IP list is also a development from Dynablock, but is more aggressively inclusive than NJABL's version.
As of March 1, 2013, NJABL is in the process of being shut down. The DNSBL zones have been emptied. After "the Internet" has had some time to remove NJABL from server configs, the NS's will be pointed off into unallocated space (192.0.2.0/24 TEST-NET-1) to hopefully make the shutdown obvious to those who were slower to notice.
As of April 29, the above-mentioned pointing of the DNSBL NS's into 192.0.2.0/24 has been done.
As of Jan 02/2019, the domain name njabl.org was set to expire and dns servers were switched to tucows autorenew servers which would cause any lookups by servers still not having removed the configuration to have rejections.
|
https://en.wikipedia.org/wiki/Composite%20Blocking%20List
|
In computer networking, the Composite Blocking List (CBL) is a DNS-based Blackhole List of suspected E-mail spam sending computer infections.
Overview
The CBL takes its source data from very large spamtraps/mail infrastructures, and only lists IPs exhibiting characteristics such as:
Open proxies of various sorts (HTTP, socks, AnalogX, wingate etc.)
Worms/viruses/botnets that do their own direct mail transmission, or are otherwise participating in a botnet.
Trojan horse or "stealth" spamware.
The CBL attempts to avoid listing real mail servers, but certain misconfigurations of mail servers can make the system appear infected (for example, servers that send HELO with 'localhost' or a similar incorrect domain.)
Entries automatically expire after a period of time.
The CBL does not provide public access to gathered evidence.
CBL data are used in Spamhaus XBL list.
See also
Comparison of DNS blacklists
CBL Index — estimate of outgoing spam reputation
External links
The CBL
CBL lookup and removal page
Computer security procedures
Spamming
|
https://en.wikipedia.org/wiki/International%20Software%20Testing%20Qualifications%20Board
|
The International Software Testing Qualifications Board (ISTQB) is a software testing certification board that operates internationally. Founded in Edinburgh in November 2002, the ISTQB is a non-profit association legally registered in Belgium.
ISTQB Certified Tester is a standardized qualification for software testers and the certification is offered by the ISTQB. The qualifications are based on a syllabus, and there is a hierarchy of qualifications and guidelines for accreditation and examination. More than 1 million ISTQB exams have been delivered and over 721,000 certifications issued; the ISTQB consists of 67 member boards worldwide representing more than 100 countries as of April 2021.
Product portfolio
Current ISTQB product portfolio follows a matrix approach characterized by
Levels, that identify progressively increasing learning objectives
Foundation
Advanced
Expert
Streams, that identify clusters of certification modules:
Core
Agile
Specialist
ISTQB streams focus on:
Core – these modules correspond to the “historical” ISTQB certifications and so they:
Cover software testing topic in a breadth-first, broad, horizontal way,
Are valid for any technology/ methodology/ application domain
Allow for a common understanding
Agile – these modules address testing practices specifically for the Agile SDLC
Specialist – these modules are new in the ISTQB product portfolio and address specific topics in a vertical way:
They can address specific quality characteristics (e.g.: Usability; Security; Performance; etc.)
They can address technologies that involve specific test approaches (e.g.: model based testing; mobile testing; etc.)
They can also be related to specific test activities (e.g.: test automation; test metrics management; etc.)
Pre-conditions
Pre-conditions relate to certification exams and provide a natural progression through the ISTQB Scheme which helps people pick the right certificate and informs them about what they need to know.
The
|
https://en.wikipedia.org/wiki/Total%20economic%20value
|
Total economic value (TEV) is a concept in cost–benefit analysis that refers to the value derived by people from a natural resource, a man-made heritage resource or an infrastructure system, compared to not having it. It appears in environmental economics as an aggregation of the (main function based) values provided by a given ecosystem.
Value classification
The value of an ecosystem can be distinguished as:
Use value — Can be split into Direct and Indirect use values:
Direct use value: Obtained through a removable product in nature (i.e., timber, fish, water).
Indirect use value: Obtained through a non-removable product in nature (i.e., sunset, waterfall).
Non-use value — Values for existence of the natural resource. For example, knowing that tigers are in the wild, even though you may never see them.
Option value: Placed on the potential future ability to use a resource even though it is not currently used and the likelihood of future use is very low. This reflects the willingness to preserve an option for potential future use.
Bequest value or existence value: Placed on a resource that will never be used by current individuals, derived from the value of satisfaction from preserving a natural environment or a historic environment (i.e., natural heritage or cultural heritage) for future generations.
Total economic value is the price of the customer's best alternative (the reference value) plus the economic value of whatever differentiates the offering from the alternative (the differentiation value).
See also
Green accounting
|
https://en.wikipedia.org/wiki/Virus%20hoax
|
A computer virus hoax is a message warning the recipients of a non-existent computer virus threat. The message is usually a chain e-mail that tells the recipients to forward it to everyone they know, but it can also be in the form of a pop-up window.
Identification
Most hoaxes are sensational in nature and easily identified by the fact that they indicate that the virus will do nearly impossible things, like blow up the recipient's computer and set it on fire, or less sensationally, delete everything on the user's computer. They often include fake announcements claimed to originate from reputable computer organizations together with mainstream news media. These bogus sources are quoted in order to give the hoax more credibility. Typically, the warnings use emotive language, stress the urgent nature of the threat and encourage readers to forward the message to other people as soon as possible.
Virus hoaxes are usually harmless and accomplish nothing more than annoying people who identify it as a hoax and wasting the time of people who forward the message. Nevertheless, a number of hoaxes have warned users that vital system files are viruses and encourage the user to delete the file, possibly damaging the system. Examples of this type include the jdbgmgr.exe virus hoax and the SULFNBK.EXE hoax.
Some consider virus hoaxes and other chain e-mails to be a computer worm in and of themselves. They replicate by social engineering—exploiting users' concern, ignorance, and disinclination to investigate before acting.
Hoaxes are distinct from computer pranks, which are harmless programs that perform unwanted and annoying actions on a computer, such as randomly moving the mouse, turning the screen display upside down, etc.
Action
Anti-virus specialists agree that recipients should delete virus hoaxes when they receive them, instead of forwarding them.
McAfee says:
F-Secure recommends:
Comparison
Telephone scam
A telephone scam, commonly operated from call centres based
|
https://en.wikipedia.org/wiki/High-mobility%20group
|
High-Mobility Group or HMG is a group of chromosomal proteins that are involved in the regulation of DNA-dependent processes such as
transcription, replication, recombination, and DNA repair.
Families
The HMG proteins are subdivided into 3 superfamilies each containing a characteristic functional domain:
HMGA – contains an AT-hook domain
HMGA1
HMGA2
HMGB – contains a HMG-box domain
HMGB1
HMGB2
HMGB3
HMGB4
HMGN – contains a nucleosomal binding domain
HMGN1
HMGN2
HMGN3
HMGN4
HMGN5
Proteins containing any of these embedded in their sequence are known as HMG motif proteins.
HMG-box proteins are found in a variety of eukaryotic organisms.
They were originally isolated from mammalian cells, and named according to their electrophoretic mobility in polyacrylamide gels.
Other families with HMG-box domain
SOX gene family
Sex-Determining Region Y Protein
SOX1, SOX2, etc.
TCF/LEF family (T cell factor/lymphoid enhancer factor family)
LEF1 (Lymphoid enhancer-binding factor 1)
TCF7 (TCF-1)
TCF7L1 (TCF-3)
TCF7L2 (TCF-4)
Function
HMG proteins are thought to play a significant role in various human disorders. Disruptions and rearrangements in the genes coding for some of the HMG proteins are associated with some common benign tumors. Antibodies to HMG proteins are found in patients with autoimmune diseases. The SRY gene on the Y Chromosome, responsible for male sexual differentiation, contains an HMG-Box domain. A member of the HMG family of proteins, HMGB1, has also been shown to have an extracellular activity as a chemokine, attracting neutrophils and mononuclear inflammatory cells to the infected liver. The high-mobility group protein such as HMO1 alters DNA architecture by binding, bending and looping. Furthermore, these HMG-box DNA-binding proteins increase the flexibility of the DNA upon binding.
In mammalian cells, the HMG non-histone proteins can modulate the activity of major DNA repair pathways including base excision repair, mismatch repair,
|
https://en.wikipedia.org/wiki/Pendulum%20%28mechanics%29
|
A pendulum is a body suspended from a fixed support so that it swings freely back and forth under the influence of gravity. When a pendulum is displaced sideways from its resting, equilibrium position, it is subject to a restoring force due to gravity that will accelerate it back toward the equilibrium position. When released, the restoring force acting on the pendulum's mass causes it to oscillate about the equilibrium position, swinging it back and forth. The mathematics of pendulums are in general quite complicated. Simplifying assumptions can be made, which in the case of a simple pendulum allow the equations of motion to be solved analytically for small-angle oscillations.
Simple gravity pendulum
A simple gravity pendulum is an idealized mathematical model of a real pendulum. This is a weight (or bob) on the end of a massless cord suspended from a pivot, without friction. Since in this model there is no frictional energy loss, when given an initial displacement it will swing back and forth at a constant amplitude. The model is based on these assumptions:
The rod or cord on which the bob swings is massless, inextensible and always remains taut.
The bob is a point mass.
Motion occurs only in two dimensions, i.e. the bob does not trace an ellipse but an arc.
The motion does not lose energy to friction or air resistance.
The gravitational field is uniform.
The support does not move.
The differential equation which represents the motion of a simple pendulum is
where is the magnitude of the gravitational field, is the length of the rod or cord, and is the angle from the vertical to the pendulum.
Small-angle approximation
The differential equation given above is not easily solved, and there is no solution that can be written in terms of elementary functions. However, adding a restriction to the size of the oscillation's amplitude gives a form whose solution can be easily obtained. If it is assumed that the angle is much less than 1 radian (often cite
|
https://en.wikipedia.org/wiki/Mechanome
|
The mechanome consists of the body, or ome, of data including cell and molecular processes relating to force and mechanical systems at molecular, cellular and tissue length scales - the fundamental "machine code" structures of the cell. The mechanome encompasses biological motors, like kinesin, myosin, RNAP, and Ribosome mechanical structures, like actin or the cytoskeleton and also proteomic and genomic components that are mechanosensitive and are involved in the response of cells to externally applied force.
A definition of the "Mechanome" extending to cell/organ/body given by Prof. Roger Kamm, at the 5th World Congress of Biomechanics Munich, includes understanding:
The complete state of stress existing from tissues to cells to molecules. The biological state that results from the distribution of forces. Requires knowledge of the distribution of force throughout the cell/organ/body, the functional interactions between these stresses and the fundamental biological processes.
The mechanome seeks to understand the fundamental physical-mechanical processes and events that affect biological function. An example at the molecular level includes the common structural designs used by kinesin and myosin motor proteins (such as dimer formation and mechanochemical cycles) that control their function and lead to properties such as processivity. The mechanome assembles the common features of these motors regardless of the "track" (microtubules, actin filaments, nucleotide based structures, membranes) they move on. A cytoskeletal example includes structures such as actin filament networks and bundles that can form from a variety of actin binding proteins that cross-link or bundle actin filaments leading to common mechanical changes of these structures. A cell machinery example includes common structures such as contractile ring formation formed by both actin and tubulin type structures leading to the same mechanical result of cell division.
In order to respond to loading ce
|
https://en.wikipedia.org/wiki/Complexity%20theory%20and%20organizations
|
Complexity theory and organizations, also called complexity strategy or complex adaptive organizations, is the use of the study of complexity systems in the field of strategic management and organizational studies. It draws from research in the natural sciences that examines uncertainty and non-linearity. Complexity theory emphasizes interactions and the accompanying feedback loops that constantly change systems. While it proposes that systems are unpredictable, they are also constrained by order-generating rules.
Complexity theory has been used in the fields of strategic management and organizational studies. Application areas include understanding how organizations or firms adapt to their environments and how they cope with conditions of uncertainty. Organizations have complex structures in that they are dynamic networks of interactions, and their relationships are not aggregations of the individual static entities. They are adaptive; in that, the individual and collective behavior mutate and self-organize corresponding to a change-initiating micro-event or collection of events.
Key concepts
Complex adaptive systems
Organizations can be treated as complex adaptive systems (CAS) as they exhibit fundamental CAS principles like self-organization, complexity, emergence, interdependence, space of possibilities, co-evolution, chaos, and self-similarity.
CAS are contrasted with ordered and chaotic systems by the relationship that exists between the system and the agents which act within it. In an ordered system the level of constraint means that all agent behavior is limited to the rules of the system. In a chaotic system, the agents are unconstrained and susceptible to statistical and other analyses. In a CAS, the system and the agents co-evolve; the system lightly constrains agent behavior, but the agents modify the system by their interaction with it. This self-organizing nature is an important characteristic of CAS; and its ability to learn to adapt, differentia
|
https://en.wikipedia.org/wiki/Hair%20tourniquet
|
Hair tourniquet is a condition where hair or thread becomes tightly wrapped around most commonly a toe, and occasionally a finger, genitals, or other body parts. This results in pain and swelling of the affected part. Complications can include tissue death due to lack of blood flow. It occurs most commonly among children around 4 months of age, though cases have been described in older children and adults.
Most cases occur accidentally. Risk factors may include autism and trichotillomania. The mechanism is believed to involve wet hair become wrapped around a body part and then tightening as it dries. Diagnosis involves examination of the entire child. Prevention is by keeping the parent's hair from contact with the baby such as by the parent keeping their hair brushed and back and washing the baby's clothing separately.
Treatment is with a substance that breaks down hair or cutting through the hair. The condition is rare. Males and females are equally frequently affected. The first medical description dates from 1832. In some cultures thread is tied around the penis of children with bedwetting or for luck.
Signs and symptoms
As this is a condition primarily of young children, symptoms are rarely reported. The child will become suddenly uncomfortable and miserable. As the digit is often inside a sock, the cause may not be clear.
The affected toe can no longer receive an adequate blood supply via the arteries, nor can blood be drained via the veins. The toe will therefore swell and turn blue, indicating ischemia.
The ligature will not stretch in response to the toe swelling and will therefore cut into the skin in more severe cases, like a cheese-wire.
Treatment
The ligature is cut or dissolved as quickly as possible. Often it is possible to lift a portion of it to enable cutting, but in a severe case the ligature must be cut through the skin. This is, of course, injurious to the child, but may prevent loss of the digit. It must take place on the side of t
|
https://en.wikipedia.org/wiki/Waffle%20%28BBS%20software%29
|
Waffle is a bulletin-board system created by Tom Dell for the Dark Side of the Moon BBS which ran under DOS and later UNIX. The software was unique among DOS BBS software in many ways, including the fact that all of the configuration files were in readable text files, and that it fully supported Usenet and UUCP on the DOS platform.
A Usenet news group named comp.bbs.waffle was created for discussion of the Waffle BBS System.
Waffle was first released in 1989. The last version seems to be v1.65. There was a beta version of 1.66 on the main site, but it was never released.
It was possible to link Waffle (under DOS) to Fidonet and WWIV using external gateway utilities.
|
https://en.wikipedia.org/wiki/Landscape%20epidemiology
|
Landscape epidemiology draws some of its roots from the field of landscape ecology. Just as the discipline of landscape ecology is concerned with analyzing both pattern and process in ecosystems across time and space, landscape epidemiology can be used to analyze both risk patterns and environmental risk factors. This field emerges from the theory that most vectors, hosts and pathogens are commonly tied to the landscape as environmental determinants control their distribution and abundance. In 1966, Evgeniy Pavlovsky introduced the concept of natural nidality or focality, defined by the idea that microscale disease foci are determined by the entire ecosystem. With the recent availability of new computing technologies such as geographic information systems, remote sensing, statistical methods including spatial statistics and theories of landscape ecology, the concept of landscape epidemiology has been applied analytically to a variety of disease systems, including malaria, hantavirus, Lyme disease and Chagas' disease.
See also
Tele-epidemiology
|
https://en.wikipedia.org/wiki/Secure%20Network
|
Secure Network is a small offensive security and security research company focusing on Information Security based in Milano, Italy. Besides having notability in Italy, it received international exposure with a research project on Bluetooth security (co-sponsored by F-Secure) codenamed BlueBag, which has been also selected for the Black Hat Briefings conference 2006 in Las Vegas.
In 2009, it also organized SEaCURE.IT, the first international technical security conference ever held in Italy.
Secure Network also offers internet security compliance consulting to private companies.
|
https://en.wikipedia.org/wiki/Curing%20salt
|
Curing salt is used in meat processing to generate a pinkish shade and to extend shelf life. It is both a color agent and a means to facilitate food preservation as it prevents or slows spoilage by bacteria or fungus. Curing salts are generally a mixture of sodium chloride (table salt) and sodium nitrite, and are used for pickling meats as part of the process to make sausage or cured meat such as ham, bacon, pastrami, corned beef, etc. Though it has been suggested that the reason for using nitrite-containing curing salt is to prevent botulism, a 2018 study by the British Meat Producers Association determined that legally permitted levels of nitrite have no effect on the growth of the Clostridium botulinum bacteria that causes botulism, in line with the UK's Advisory Committee on the Microbiological Safety of Food opinion that nitrites are not required to prevent C. botulinum growth and extend shelf life. (see also Sodium Nitrite: Inhibition of microbial growth).
Many curing salts also contain red dye that makes them pink to prevent them from being confused with common table salt. Thus curing salt is sometimes referred to as "pink salt". Curing salts are not to be confused with Himalayan pink salt, a halite which is 97–99% sodium chloride (table salt) with trace elements that give it a pink color.
Types
There are many types of curing salts often specific to a country or region.
Prague Powder #1
One of the most common curing salts. It is also called Insta Cure #1 or Pink curing salt #1. It contains 6.25% sodium nitrite and 93.75% table salt. It is recommended for meats that require short cures and will be cooked and eaten relatively quickly. Sodium nitrite provides the characteristic flavor and color associated with curing.
Prague Powder #2
Also called Pink curing salt #2. It contains 6.25% sodium nitrite, 4% sodium nitrate, and 89.75% table salt. The sodium nitrate found in Prague powder #2 gradually breaks down over time into sodium nitrite, and by the time a
|
https://en.wikipedia.org/wiki/Deborah%20Fouts
|
Deborah Fouts was the co-director of the Chimpanzee and Human Communication Institute (CHCI). CHCI was the home of Washoe, the first non-human to acquire a human language, and three other chimpanzees who use the signs of American Sign Language to communicate with each other and their human caregivers. She is an Adjunct Assistant Professor of Psychology (Research) at Central Washington University in Ellensburg, Washington. She is married to former co-director, now retired Roger Fouts.
Professional background
Roger and Deborah Fouts were co-directors and co-founders of the Chimpanzee and Human Communication Institute.
The Foutses have been a part of Project Washoe since 1967. Project Washoe is the first and longest running project of its kind. Washoe was the very first nonhuman animal to acquire the rudiments of an otherwise fully complex human language, American Sign Language for the Deaf (ASL). The project later focused on the signing of its four sign language using chimpanzees who lived together as a social group: Washoe, Tatu, Dar and Loulis. The chimpanzees used some of the signs of ASL in their interactions with humans and with each other, to comment on their environment, but mostly to make requests, and sometimes answer questions and describe activities and objects. Loulis acquired a total of about five signs from his adoptive mother Washoe and the other chimpanzees, becoming the first chimpanzee to acquire five signs of a human language from chimpanzees thus demonstrating the ability of the chimpanzee to culturally transmit five signs of a language across generations.
Together the Foutses have more than 100 articles published in scientific journals and books. In 1981 they founded Friends of Washoe a non-profit organization dedicated to the welfare of chimpanzees. They have resided in Ellensburg since 1980; and, with hundreds of students, they have conducted their research and enriched the lives of the chimpanzees at Central Washington University. In 1992 t
|
https://en.wikipedia.org/wiki/Mary%20Lee%20Jensvold
|
Mary Lee Jensvold is a senior lecturer at Central Washington University. She was the Director of the Chimpanzee and Human Communication Institute (CHCI) located on the campus of Central Washington University. CHCI was the home of the chimpanzee Washoe and four other chimpanzees who use the signs of American Sign Language to communicate with one another and their human caregivers.
Jensvold studies communication and other behaviors in chimpanzees. She had been working with Washoe and her family since 1986. In 1985 she received a B.A. in Psychology from the University of Oregon, in 1989 a M.S. in Experimental Psychology from Central Washington University, and in 1996 a Ph.D. in Experimental Psychology from the University of Nevada, Reno.
She specializes in ethological studies of great apes, animal intelligence, communication, and language. Her studies include chimpanzee conversational skills, imaginary play in chimpanzees, and environmental enrichment for captive chimpanzees. She is currently Senior Lecturer in the Primate Behavior and Ecology Program and the Anthropology Department at Central Washington University. She also serves on the Advisory Council of METI (Messaging Extraterrestrial Intelligence).
See also
Loulis
Deborah Fouts
Roger Fouts
Koko
|
https://en.wikipedia.org/wiki/List%20of%20mobile%20telephone%20prefixes%20by%20country
|
This is a list of mobile telephone prefixes by country.
International prefixes table
See also
List of country calling codes
Notes
|
https://en.wikipedia.org/wiki/Taxonomic%20sequence
|
Taxonomic sequence (also known as systematic, phyletic or taxonomic order) is a sequence followed in listing of taxa which aids ease of use and roughly reflects the evolutionary relationships among the taxa. Taxonomic sequences can exist for taxa within any rank, that is, a list of families, genera, species can each have a sequence.
Early biologists used the concept of "age" or "primitiveness" of the groups in question to derive an order of arrangement, with "older" or more "primitive" groups being listed first and more recent or "advanced" ones last. A modern understanding of evolutionary biology has brought about a more robust framework for the taxonomic ordering of lists. A list may be seen as a rough one-dimensional representation of a phylogenetic tree. Taxonomic sequences are essentially heuristic devices that help in arrangements of linear systems such as books and information retrieval systems. Since phylogenetic relationships are complex and non-linear, there is no unique way to define the sequence, although they generally have the more basal listed first with species that cluster in a tight group included next to each other.
The organization of field guides and taxonomic monographs may either follow or prescribe the taxonomic sequence; changes in these sequences are often introduced by new publications.
|
https://en.wikipedia.org/wiki/Regular%20number
|
Regular numbers are numbers that evenly divide powers of 60 (or, equivalently, powers of 30). Equivalently, they are the numbers whose only prime divisors are 2, 3, and 5. As an example, 602 = 3600 = 48 × 75, so as divisors of a power of 60 both 48 and 75 are regular.
These numbers arise in several areas of mathematics and its applications, and have different names coming from their different areas of study.
In number theory, these numbers are called 5-smooth, because they can be characterized as having only 2, 3, or 5 as their prime factors. This is a specific case of the more general -smooth numbers, the numbers that have no prime factor greater
In the study of Babylonian mathematics, the divisors of powers of 60 are called regular numbers or regular sexagesimal numbers, and are of great importance in this area because of the sexagesimal (base 60) number system that the Babylonians used for writing their numbers, and that was central to Babylonian mathematics.
In music theory, regular numbers occur in the ratios of tones in five-limit just intonation. In connection with music theory and related theories of architecture, these numbers have been called the harmonic whole numbers.
In computer science, regular numbers are often called Hamming numbers, after Richard Hamming, who proposed the problem of finding computer algorithms for generating these numbers in ascending order. This problem has been used as a test case for functional programming.
Number theory
Formally, a regular number is an integer of the form , for nonnegative integers , , and . Such a number is a divisor of . The regular numbers are also called 5-smooth, indicating that their greatest prime factor is at most 5. More generally, a -smooth number is a number whose greatest prime factor is at
The first few regular numbers are
Several other sequences at the On-Line Encyclopedia of Integer Sequences have definitions involving 5-smooth numbers.
Although the regular numbers appear dense withi
|
https://en.wikipedia.org/wiki/Amylomaize
|
Amylomaize was a term coined in the late 1940s by Robert P. Bear of Bear Hybrids Corn Company in Decatur, Illinois to describe his discovery and commercial breeding of a cornstarch with high (>50%) amylose content, also called high amylose starch. The discovery of amylomaize occurred as a mutation in a normal inbred line; from that one mutation an entire new kind of maize (corn) was developed.
Principal uses of amylomaize starch are in making biodegradable plastics (or bioplastics). It is also used in coatings that are edible and digestible. The food that the first American astronauts used on Apollo flights from 1969 to 1972 were coated with amylomaize film, so that no crumbs would float around in the space capsule.
Several years earlier Robert P. Bear had discovered and reported that waxy corn (100% amylopectin starch) also occurred as a mutation. Once discovered and reported, waxy mutations were found in the order of once every 30,000 observations.
See also
Waxy maize
|
https://en.wikipedia.org/wiki/Anthesis
|
Anthesis is the period during which a flower is fully open and functional. It may also refer to the onset of that period.
The onset of anthesis is spectacular in some species. In Banksia species, for example, anthesis involves the extension of the style far beyond the upper perianth parts. Anthesis of flowers is sequential within an inflorescence, so when the style and perianth are different colours, the result is a striking colour change that gradually sweeps along the inflorescence.
Flowers with diurnal anthesis generally are brightly colored in order to attract diurnal insects, such as butterflies.
Flowers with nocturnal anthesis generally are white or less colorful, and as such, they contrast more strongly with the night. These flowers typically attract nocturnal insects including many moth species.
|
https://en.wikipedia.org/wiki/Divisor%20summatory%20function
|
In number theory, the divisor summatory function is a function that is a sum over the divisor function. It frequently occurs in the study of the asymptotic behaviour of the Riemann zeta function. The various studies of the behaviour of the divisor function are sometimes called divisor problems.
Definition
The divisor summatory function is defined as
where
is the divisor function. The divisor function counts the number of ways that the integer n can be written as a product of two integers. More generally, one defines
where dk(n) counts the number of ways that n can be written as a product of k numbers. This quantity can be visualized as the count of the number of lattice points fenced off by a hyperbolic surface in k dimensions. Thus, for k=2, D(x) = D2(x) counts the number of points on a square lattice bounded on the left by the vertical-axis, on the bottom by the horizontal-axis, and to the upper-right by the hyperbola jk = x. Roughly, this shape may be envisioned as a hyperbolic simplex. This allows us to provide an alternative expression for D(x), and a simple way to compute it in time:
, where
If the hyperbola in this context is replaced by a circle then determining the value of the resulting function is known as the Gauss circle problem.
Sequence of D(n):
0, 1, 3, 5, 8, 10, 14, 16, 20, 23, 27, 29, 35, 37, 41, 45, 50, 52, 58, 60, 66, 70, 74, 76, 84, 87, 91, 95, 101, 103, 111, ...
Dirichlet's divisor problem
Finding a closed form for this summed expression seems to be beyond the techniques available, but it is possible to give approximations. The leading behavior of the series is given by
where is the Euler–Mascheroni constant, and the error term is
Here, denotes Big-O notation. This estimate can be proven using the Dirichlet hyperbola method, and was first established by Dirichlet in 1849. The Dirichlet divisor problem, precisely stated, is to improve this error bound by finding the smallest value of for which
holds true for all . As of today,
|
https://en.wikipedia.org/wiki/Hopcroft%E2%80%93Karp%20algorithm
|
In computer science, the Hopcroft–Karp algorithm (sometimes more accurately called the Hopcroft–Karp–Karzanov algorithm) is an algorithm that takes a bipartite graph as input and produces a maximum-cardinality matching as output — a set of as many edges as possible with the property that no two edges share an endpoint. It runs in time in the worst case, where is set of edges in the graph, is set of vertices of the graph, and it is assumed that . In the case of dense graphs the time bound becomes , and for sparse random graphs it runs in time with high probability.
The algorithm was discovered by and independently by . As in previous methods for matching such as the Hungarian algorithm and the work of , the Hopcroft–Karp algorithm repeatedly increases the size of a partial matching by finding augmenting paths. These paths are sequences of edges of the graph, which alternate between edges in the matching and edges out of the partial matching, and where the initial and final edge are not in the partial matching. Finding an augmenting path allows us to increment the size of the partial matching, by simply toggling the edges of the augmenting path (putting in the partial matching those that were not, and vice versa). Simpler algorithms for bipartite matching, such as the Ford–Fulkerson algorithm‚ find one augmenting path per iteration: the Hopcroft-Karp algorithm instead finds a maximal set of shortest augmenting paths, so as to ensure that only iterations are needed instead of iterations. The same performance of can be achieved to find maximum-cardinality matchings in arbitrary graphs, with the more complicated algorithm of Micali and Vazirani.
The Hopcroft–Karp algorithm can be seen as a special case of Dinic's algorithm for the maximum-flow problem.
Augmenting paths
A vertex that is not the endpoint of an edge in some partial matching is called a free vertex. The basic concept that the algorithm relies on is that of an augmenting path, a path that starts
|
https://en.wikipedia.org/wiki/D-Grid
|
The D-Grid Initiative (German Grid Initiative) was a government project to fund computer infrastructure for education and research (e-Science) in Germany. It uses the term grid computing.
D-Grid started September 1, 2005 with six community projects and an integration project (DGI) as well as several partner projects.
Integration project
The D-Grid integration project intended to integrate community projects. The D-Grid integration project acted as a service provider for the science community in Germany. The project office is located at the Institute for Scientific Computing (IWR) at Forschungszentrum Karlsruhe. The resources to ensure a sustainable Grid infrastructure are provided by four work packages:
D-Grid Base-Software: The major task of this work package is to provide several different middleware packages. These are the Globus Toolkit, UNICORE, LCG/gLite, GridSphere and the Grid Application Toolkit (GAT). The community projects linked together in the D-Grid integration project are supported during the installation, operation and if needed and possible the customisation of the Base-Software.
Deployment and operation of the D-Grid infrastructure: Work package 2 builds up a Core-D-Grid. It was used as a prototype to test the operational functionality of the system. This work package also deals with monitoring, accounting and billing.
Networks and Security: The network infrastructure in D-Grid is based on the DFN Wissenschaftsnetz X-WiN. Work package 3 will provide extensions to the existing network infrastructure according to the needs of Grid middleware used in D-Grid. Further tasks are to build an AA-Infrastructure in D-Grid, develop firewall concepts for Grid environments and set up Grid specific CERT services.
D-Grid project office: The work package is responsible for the integration of community projects into one common D-Grid platform. Work package 4 also deals with sustainability.
Communities
Six community projects participated in the D-Grid In
|
https://en.wikipedia.org/wiki/Fair%20Trade%20USA
|
Fair Trade USA, formerly "TransFair USA", is a 501(c)(3) non-profit organization that sets standards, certifies, and labels products that promote sustainable livelihoods for farmers and workers and protect the environment.
About
Founded in 1998 by the Institute for Agricultural Trade Policy (IATP), Fair Trade USA is an independent, nonprofit organization that sets standards, certifies, and labels products that promote sustainable livelihoods for farmers and workers and protect the environment.
Growth in fair trade
Although coffee remains the most popular fair trade product, Fair Trade USA certifies a variety of product categories, including tea, cocoa, sugar, spices, honey, produce, grains, wine and spirits, flowers, apparel and home goods, and body care. There are more than 12,000 individual Fair Trade Certified products available in North America, and the market is growing rapidly.
Campaigns
Comedian Jimmy Fallon kicked off Fair Trade Month by hosting ice cream makers Ben & Jerry on the Late Night with Jimmy Fallon. The October 4, 2011, episode mentioned the March episode in which Ben Cohen and Jerry Greenfield introduced their new flavor featuring Fair Trade Certified ingredients, Late Night Snack. Introducing the ice cream flavor, which is made with Fair Trade Certified vanilla and cocoa, gave Ben and Jerry the opportunity to discuss Fair Trade and the company's goal to use entirely Fair Trade Certified ingredients by 2013.
Green Mountain Coffee partnered with musical groups Michael Franti & Spearhead and Grace Potter and the Nocturnals to broadcast live concerts promoting Fair Trade. Both concerts were streamed on the Green Mountain Coffee Facebook page. Green Mountain Coffee was recognized in September 2011 as the world's largest purchaser of Fair Trade Certified coffee, having bought 26 million pounds during 2010.
Also in 2011, pastry chef Malika Ameen of Top Chef: Just Desserts Season 1 joined celebrity dietician Ashley Koff to create three dishes usin
|
https://en.wikipedia.org/wiki/AX%20architecture
|
AX (Architecture eXtended) was a Japanese computing initiative starting in around 1986 to allow PCs to handle double-byte (DBCS) Japanese text via special hardware chips, whilst allowing compatibility with software written for foreign IBM PCs.
History
The idea was conceived by Kazuhiko Nishi before he resigned his position as vice president of Microsoft. Microsoft Japan took over the project, and in July 1987 the Preparatory Committee of the AX Consortium started developing its specification.
The AX Consortium officially started in October 1987, including ASCII Corporation, Sony, Hitachi, Sharp, Oki, Casio, Canon, Kyocera, Sanyo, Mitsubishi Electric, etc., but notably excluding Toshiba and Fujitsu (who were hence the 'opposition').
At that time, NEC PC-9801 was the dominant PC architecture in the Japanese PC market because IBM PC/AT and its clone PCs could not display Japanese text. However, NEC did not tolerate PC-9801 compatible machines and was fighting court battles with Epson which was the only PC-9801 compatible machine vendor. Therefore, other vendors desperately needed a standard specification for Japanese capable PCs.
Eventually two standards were developed: JEGA and AX-VGA.
Due to less available software and its higher cost compared to the PC-9801 series, AX failed and was not able to break into the market in Japan. The Nikkei Personal Computing journal reported in 1989 that only 18 out of 36,165 PCs used in 937 companies were AX machines, and 90% of companies had no plan to purchase the AX machine.
In 1990, IBM Japan unveiled DOS/V which enabled IBM PC/AT and its clones to display Japanese text without any additional hardware using a standard VGA card. Soon after, AX disappeared and the decline of NEC PC-9801 began.
AX architecture machines
Several companies released AX computers:
Oki Electric Industry if386AX30 / 50 series
Casio Computer AX-8000D / 8000L
Canon Axi DX-20 / 20P / 10 / 10P
Kyocera AX386 model A
Sanyo Electric MCB-17 /18 se
|
https://en.wikipedia.org/wiki/National%20Centre%20for%20Food%20Manufacturing
|
The National Centre for Food Manufacturing (NCFM) is the food science campus of the University of Lincoln, situated on Park Road at Holbeach in the south of the county of Lincolnshire. It offers part-time apprenticeships and distance learning degrees for individuals working in the food industry.
Apprenticeships, Degree and Foundation Courses
The National Centre for Food Manufacturing offers part time distance learning options to achieve Foundation and BSc (Honours) Food Manufacture degrees and higher degrees through research together with all levels of apprenticeships including Higher Apprenticeships (which includes a Foundation Degree). The Foundation and Undergraduate degrees cover areas including Food and Drink Operations and Manufacturing Management - Food Science and Technology – and Food Supply Chain Management. The Centre also offers part time Masters and PhDs - often progressed by food sector employees and focused on specific Food Manufacturing Industry Challenges.
The Higher and Degree Apprenticeships include the CMDA (Chartered Manager Degree Apprenticeship), Departmental Manager, Laboratory Scientist and Professional Technical degrees.
The Centre provides support to apprentices for Functional Skills development in maths and English as required by their relevant apprenticeship standard and offers employers a complete skills development programme for its employees.
NCFM has apprenticeship partnerships with 250 UK food businesses including Addo Food Group, Bakkavor, Bidfood, Dalehead Foods, Summers Butchery Services, Greencore Group, Tulip, Dovecote Park, Fresttime, Finlays, JDM Food Group, Kerry, Nestle, Worldwide Fruit, University Academy Holbeach, Produce World Group, J.O. Sims Ltd, Greenvale AP, FreshLinc, Ripe Now and Lincolshire Field Products.
Research
NCFM advances food manufacturing and related food supply chain research initiatives via a wide range of industry and academic partnerships. The areas of core research include Robotics and Automatio
|
https://en.wikipedia.org/wiki/Guqin%20notation
|
The notation of the guqin is a unique form of tablature for the Chinese musical instrument, with a history of over 1,500 years, still in use today.
History
Written qin music did not directly tell what notes were played; instead, it was written in a tablature detailing tuning, finger positions, and stroke technique, thus comprising a step by step method and description of how to play a piece. Some tablatures do indicate notes using the gongche system, or indicate rhythm using dots. The earliest example of the modern shorthand tablature survives from around the twelfth century CE. An earlier form of music notation from the Tang era survives in just one manuscript, dated to the seventh century CE, called Jieshi Diao Youlan 《碣石調幽蘭》 (Solitary Orchid in Stone Tablet Mode). It is written in a longhand form called wenzi pu 〔文字譜〕 (literally "written notation"), said to have been created by Yongmen Zhou (雍門周) during the Warring States period, which gives all the details using ordinary written Chinese characters. Later in the Tang dynasty Cao Rou (曹柔) and others simplified the notation, using only the important elements of the characters (like string number, plucking technique, hui number and which finger to stop the string) and combined them into one character notation. This meant that instead of having two lines of written text to describe a few notes, a single character could represent one note, or sometimes as many as nine. This notation form was called jianzi pu 〔減字譜〕 (literally "reduced notation") and it was a great advancement for recording qin pieces. It was so successful that from the Ming dynasty onwards, a great many qinpu 〔琴譜〕 (qin tablature collections) appeared, the most famous and useful being "Shenqi Mipu" (神奇秘譜, lit. The Mysterious and Marvellous Tablature) compiled by Zhu Quan, the 17th son of the founder of the Ming dynasty . In the 1960s, Zha Fuxi discovered more than 130 qinpu that contain well over 3360 pieces of written music. Sadly, many qinpu compi
|
https://en.wikipedia.org/wiki/Unicornuate%20uterus
|
A unicornuate uterus represents a uterine malformation where the uterus is formed from one only of the paired Müllerian ducts while the other Müllerian duct does not develop or only in a rudimentary fashion. The sometimes called hemi-uterus has a single horn linked to the ipsilateral fallopian tube that faces its ovary.
Signs and symptoms
Women with the condition may be asymptomatic and unaware of having a unicornuate uterus; normal pregnancy may occur. In a review of the literature Reichman et al. analyzed the data on pregnancy outcome of 290 women with a unicornuate uterus. 175 women had conceived for a total of 468 pregnancies. They found that about 50% of patients delivered a live baby. The rates for ectopic pregnancy was 2.7%, for miscarriage 34%, and for preterm delivery 20%, while the intrauterine demise rate was 10%. Thus patients with a unicornuate uterus are at a higher risk for pregnancy loss and obstetrical complications.
Cause
The uterus is normally formed during embryogenesis by the fusion of the two Müllerian ducts. If one of the ducts does not develop, only one Müllerian duct contributes to the uterine development. This uterus may or may not be connected to Müllerian structure on the opposite site if the Müllerian duct on that site undergoes some development. A unicornuate uterus has a single cervix and vagina.
Associated defects may affect the renal system, and less common, the skeleton.
The condition is much less common than these other uterine malformations: arcuate uterus, septate uterus, and bicornuate uterus. While the uterus didelphys is estimated to occur in 1/3,000 women, the unicornuate uterus appears to be even more infrequent with an estimated occurrence of about 1/4,000.
Diagnosis
A pelvic examination will typically reveal a single vagina and a single cervix. Investigations are usually prompted on the basis of reproductive problems.
Helpful techniques to investigate the uterine structure are transvaginal ultrasonography and sonohy
|
https://en.wikipedia.org/wiki/Arcuate%20uterus
|
The arcuate uterus is a form of a uterine anomaly or variation where the uterine cavity displays a concave contour towards the fundus. Normally the uterine cavity is straight or convex towards the fundus on anterior-posterior imaging, but in the arcuate uterus the myometrium of the fundus dips into the cavity and may form a small septation. The distinction between an arcuate uterus and a septate uterus is not standardized.
Signs and symptoms
The condition may not be known to the affected individual and not result in any reproductive problems; thus normal pregnancies occur. Indeed, there is no consensus on the relationship of the arcuate uterus and recurrent pregnancy loss. Accordingly, the condition may be a variation or a pathology.
One view maintains that the condition is associated with a higher risk for miscarriage, premature birth, and malpresentation. Thus a study that evaluated women with uterine bleeding by hysteroscopy found that 6.5% of subjects displayed the arcuate uterus and had evidence of reproductive impairments. A study based on hysterosalpingraphic detected arcuate lesions documented increased fetal loss and obstetrical complications as a risk for affected women. Woelfer found that the miscarriage risk is more pronounced in the second trimester. In contrast, a study utilizing 3-D ultrasonography to document the prevalence of the arcuate uterus in a gynecological population found no evidence of increased risk of reproductive loss; in this study 3.1% of women had an arcuate uterus making it the most common uterine anomaly; this prevalence was similar than in women undergoing sterilization and lower than in women with recurrent pregnancy loss.
Cause
The uterus is formed during embryogenesis by the fusion of the two Müllerian ducts. During this fusion a resorption process eliminates the partition between the two ducts to create a single cavity. This process begins caudally and advances cranially, thus an arcuate uterus represents an in the final sta
|
https://en.wikipedia.org/wiki/Uterine%20septum
|
A uterine septum is a congenital uterine malformation where the uterine cavity is partitioned by a longitudinal septum; the outside of the uterus has a normal typical shape. The wedge-like partition may involve only the superior part of the cavity resulting in an incomplete septum or a subseptate uterus, or less frequently the total length of the cavity (complete septum) and the cervix resulting in a double cervix. The septation may also continue caudally into the vagina resulting in a "double vagina".
Signs and symptoms
The condition may not be known to the affected individual and not result in any reproductive problems; thus normal pregnancies may occur. In more serious cases have reported high infertility rates. However, it is associated with a higher risk for miscarriage, premature birth, and malpresentation. According to the classical study by Buttram there is a 60% risk of a spontaneous abortion, this being more common in the second than in the first trimester. However, there is no agreement on this number and other studies show a lower risk. Woelfer found that the miscarriage risk is more pronounced in the first trimester.
The condition is also associated with abnormalities of the renal system. Further, skeletal abnormalities have been linked to the condition.
Cause
The uterus is formed during embryogenesis by the fusion of the two Müllerian ducts. During this fusion a resorption process eliminates the partition between the two ducts to create a single cavity. This process begins caudally and advances cranially, thus a complete septum formation represents an earlier disturbance of this absorption than the incomplete form. Causes for incomplete absorption are not known.
Diagnosis
A pelvic examination may reveal a double vagina or double cervix that should be further investigated and may lead to the discovery of a uterine septum. In most patients, however, the pelvic examination is normal. Investigations are usually prompted on the basis of reproductive pr
|
https://en.wikipedia.org/wiki/Snippet%20%28programming%29
|
Snippet is a programming term for a small region of re-usable source code, machine code, or text. Ordinarily, these are formally defined operative units to incorporate into larger programming modules. Snippet management is a feature of some text editors, program source code editors, IDEs, and related software. It allows the user to avoid repetitive typing in the course of routine edit operations.
Definition
In programming practice, "snippet" refers narrowly to a portion of source code that is literally included by an editor program into a file, and is a form of copy and paste programming. This concrete inclusion is in contrast to abstraction methods, such as functions or macros, which are abstraction within the language. Snippets are thus primarily used when these abstractions are not available or not desired, such as in languages that lack abstraction, or for clarity and absence of overhead.
Snippets are similar to having static preprocessing included in the editor, and do not require support by a compiler. On the flip side, this means that snippets cannot be invariably modified after the fact, and thus is vulnerable to all of the problems of copy and paste programming. For this reason snippets are primarily used for simple sections of code (with little logic), or for boilerplate, such as copyright notices, function prototypes, common control structures, or standard library imports.
Overview
Snippet management is a text editor feature popular among software developers or others who routinely require content from a catalogue of repeatedly entered text (such as with source code or boilerplate). Often this feature is justified because the content varies only slightly (or not at all) each time it is entered.
Snippets in text editors
Text editors that include this feature ordinarily provide a mechanism to manage the catalogue, and separate "snippets" in the same manner that the text editor and operating system allow management of separate files. These basic manag
|
https://en.wikipedia.org/wiki/Welch%E2%80%93Satterthwaite%20equation
|
In statistics and uncertainty analysis, the Welch–Satterthwaite equation is used to calculate an approximation to the effective degrees of freedom of a linear combination of independent sample variances, also known as the pooled degrees of freedom, corresponding to the pooled variance.
For sample variances , each respectively having degrees of freedom, often one computes the linear combination.
where is a real positive number, typically . In general, the probability distribution of {{math|χ}} cannot be expressed analytically. However, its distribution can be approximated by another chi-squared distribution, whose effective degrees of freedom are given by the Welch–Satterthwaite equation'''
There is no assumption that the underlying population variances are equal. This is known as the Behrens–Fisher problem.
The result can be used to perform approximate statistical inference tests. The simplest application of this equation is in performing Welch's t-test.
See also
Pooled variance
|
https://en.wikipedia.org/wiki/Hjulstr%C3%B6m%20curve
|
The Hjulström curve, named after Filip Hjulström (1902–1982), is a graph used by hydrologists and geologists to determine whether a river will erode, transport, or deposit sediment. It was originally published in his doctoral thesis "Studies of the morphological activity of rivers as illustrated by the River Fyris." in 1935. The graph takes sediment particle size and water velocity into account.
The upper curve shows the critical erosion velocity in cm/s as a function of particle size in mm, while the lower curve shows the deposition velocity as a function of particle size. Note that the axes are logarithmic.
The plot shows several key concepts about the relationships between erosion, transportation, and deposition. For particle sizes where friction is the dominating force preventing erosion, the curves follow each other closely and the required velocity increases with particle size. However, for cohesive sediment, mostly clay but also silt, the erosion velocity increases with decreasing grain size, as the cohesive forces are relatively more important when the particles get smaller. The critical velocity for deposition, on the other hand, depends on the settling velocity, and that decreases with decreasing grainsize. The Hjulström curve shows that sand particles of a size around 0.1 mm require the lowest stream velocity to erode.
The curve was expanded by Åke Sundborg in 1956. He significantly improved the level of detail in the cohesive part of the diagram, and added lines for different modes of transportation. The result is called the Sundborg diagram, or the Hjulström-Sundborg Diagram, in the academic literature.
This curve dates back to early 20th century research on river geomorphology and has no more than a historical value nowadays, although its simplicity is still attractive. Among the drawbacks of this curve are that it does not take the water depth into account and more importantly, that it does not show that sedimentation is caused by flow velocity
|
https://en.wikipedia.org/wiki/Apollo%20PRISM
|
PRISM (Parallel Reduced Instruction Set Multiprocessor) was Apollo Computer's high-performance CPU used in their DN10000 series workstations. It was for some time the fastest microprocessor available, a high fraction of a Cray-1 in a workstation. Hewlett-Packard purchased Apollo in 1989, ending development of PRISM, although some of PRISM's ideas were later used in HP's own HP-PA Reduced instruction set computer (RISC) and Itanium processors.
PRISM was based on what would be known today as a VLIW-design, while most efforts of the era, 1988, were based on a more "pure" RISC approach. In early RISC designs, the core processor was simplified as much as possible in order to allow more of the chip's real-estate to be used for registers and simplifying the addition of instruction pipelines for improved performance.
Compilers
The compilers used with the systems were expected to dedicate more time during compilation to making effective use of the registers and cleaning the instruction stream. By doing instruction scheduling in the compiler, this design avoided the problems and complexity of dynamic instruction scheduling (where instructions for multiple functional units must be selected carefully in order to avoid interdependencies between intermediate values) encountered in superscalar designs such as Digital Equipment Corporation's Alpha.
In some respects, the VLIW design can be thought of as "super-RISCy", as it offloads the instruction selection process to the compiler as well. In the VLIW design, the compiler examines the code and selects instructions that are known to be "safe", and then packages them into longer instruction words. For instance, for a CPU with two functional units, like the PRISM, the compiler would find pairs of safe instructions and stuff them into a single larger word. Inside the CPU, the instructions are simply split apart again, and fed into the selected units.
This design minimizes logical changes to the CPU as functional units are added,
|
https://en.wikipedia.org/wiki/CompuCell3D
|
CompuCell3D (CC3D) is a three-dimensional C++ and Python software problem solving environment for simulations of biocomplexity problems, integrating multiple mathematical [morphogenesis] models. These include the cellular Potts model (CPM) which can model cell clustering, growth, division, death, adhesion, and volume and surface area constraints; as well as partial differential equation solvers for modeling reaction–diffusion of external chemical fields and cell type automata for differentiation. By integrating these models CompuCell3D enables modeling of cellular reactions to external chemical fields such as secretion or resorption, and responses such as chemotaxis and haptotaxis.
CompuCell3D is conducive for experimentation and testing biological models by providing a flexible and extensible package, with many different levels of control. High-level steering is possible through CompuCell Player, an interactive GUI built upon Qt threads which execute in parallel with the computational back end. Functionality such as zooming, rotation, playing and pausing simulations, setting colors and viewing cross sections is available through the player, with a sample screenshot shown below.
Extending the back end is possible through an XML-based domain-specific language Biologo, which after lexical analysis and generation transparently converts to C++ extensions which can be compiled and dynamically loaded at runtime. The back end uses object-oriented design patterns which contribute to extensibility, reducing coupling between independently operating modules. Optional functionality can be encapsulated through plugins, which are dynamically loaded at runtime through an XML configuration file reference.
CompuCell3D can model several different phenomena, including avian limb development, in vitro capillary development, adhesion-driven cell sorting, Dictyostelium discoideum, and fluid flows. The binaries and source code, as well as documentation and examples, are available at th
|
https://en.wikipedia.org/wiki/Pontecorvo%E2%80%93Maki%E2%80%93Nakagawa%E2%80%93Sakata%20matrix
|
In particle physics, the Pontecorvo–Maki–Nakagawa–Sakata matrix (PMNS matrix), Maki–Nakagawa–Sakata matrix (MNS matrix), lepton mixing matrix, or neutrino mixing matrix is a unitary
mixing matrix which contains information on the mismatch of quantum states of neutrinos when they propagate freely and when they take part in weak interactions. It is a model of neutrino oscillation. This matrix was introduced in 1962 by Ziro Maki, Masami Nakagawa, and Shoichi Sakata,
to explain the neutrino oscillations predicted by Bruno Pontecorvo.
The PMNS matrix
The Standard Model of particle physics contains three generations or "flavors" of neutrinos,
,
, and
,
each labeled with a subscript showing the charged lepton that it partners with in the charged-current weak interaction. These three eigenstates of the weak interaction form a complete, orthonormal basis for the Standard Model neutrino. Similarly, one can construct an eigenbasis out of three neutrino states of definite mass, , , and , which diagonalize the neutrino's free-particle Hamiltonian. Observations of neutrino oscillation established experimentally that for neutrinos, as for quarks, these two eigenbases are different – they are 'rotated' relative to each other.
Consequently, each flavor eigenstate can be written as a combination of mass eigenstates, called a "superposition", and vice versa. The PMNS matrix, with components corresponding to the amplitude of mass eigenstate in terms of flavor "", "", ""; parameterizes the unitary transformation between the two bases:
The vector on the left represents a generic neutrino expressed in the flavor-eigenstate basis, and on the right is the PMNS matrix multiplied by a vector representing that same neutrino in the mass-eigenstate basis. A neutrino of a given flavor is thus a "mixed" state of neutrinos with distinct mass: If one could measure directly that neutrino's mass, it would be found to have mass with probability .
The PMNS matrix for antineutrinos is identical
|
https://en.wikipedia.org/wiki/GgNMOS
|
Grounded-gate NMOS, commonly known as ggNMOS, is an electrostatic discharge (ESD) protection device used within CMOS integrated circuits (ICs). Such devices are used to protect the inputs and outputs of an IC, which can be accessed off-chip (wire-bonded to the pins of a package or directly to a printed circuit board) and are therefore subject to ESD when touched. An ESD event can deliver a large amount of energy to the chip, potentially destroying input/output circuitry; a ggNMOS device or other ESD protective devices provide a safe path for current to flow, instead of through more sensitive circuitry. ESD protection by means of such devices or other techniques is important to product reliability: 35% of all IC failures in the field are associated with ESD damage.
Structure
As the name implies, a ggNMOS device consists of a relatively wide NMOS device in which the gate, source, and body are tied together to ground. The drain of the ggNMOS is connected to the I/O pad under protection. A parasitic NPN bipolar junction transistor (BJT) is thus formed with the drain (n-type) acting as the collector, the base/source combination (n-type) as the emitter, and the substrate (p-type) as the base. As is explained below, a key element to the operation of the ggNMOS is the parasitic resistance present between the emitter and base terminals of the parasitic npn BJT. This resistance is a result of the finite conductivity of the p-type doped substrate.
Operation
When a positive ESD event appears upon the I/O pad (drain), the collector-base junction of the parasitic NPN BJT becomes reverse biased to the point of avalanche breakdown. At this point, the positive current flowing from the base to ground induces a voltage potential across the parasitic resistor, causing a positive voltage to appear across the base-emitter junction. The positive VBE forward biases this junction, triggering the parasitic NPN BJT.
|
https://en.wikipedia.org/wiki/Nautilus%20%28secure%20telephone%29
|
Nautilus is a program which allows two parties to securely communicate using modems or TCP/IP. It runs from a command line and is available for the Linux and Windows operating systems. The name was based upon Jules Verne's Nautilus and its ability to overcome a Clipper ship as a play on Clipper chip.
The program was originally developed by
Bill Dorsey, Andy Fingerhut, Paul Rubin, Bill Soley, and David Miller.
Nautilus is historically significant in the realm of secure communications because it was one of the first programs which were released as open source to the general public which used strong encryption. It was created as a response to the Clipper chip in which the US government planned to use a key escrow scheme on all products which used the chip. This would allow them to monitor "secure" communications. Once this program and another similar program PGPfone were available on the internet, the proverbial cat was "out of the bag" and it would have been nearly impossible to stop the use of strong encryption for telephone communications.
The project had to move their web presence by the end of May 2014 due to the decision of to shut down the developer platform that hosted the project.
External links
new Nautilus homepage from May 1 2014 on
"Can Nautilus Sink Clipper?" Article in Wired, Aug 1995
Secure communication
Cryptographic software
VoIP software
|
https://en.wikipedia.org/wiki/History%20of%20electromagnetic%20theory
|
The history of electromagnetic theory begins with ancient measures to understand atmospheric electricity, in particular lightning. People then had little understanding of electricity, and were unable to explain the phenomena. Scientific understanding into the nature of electricity grew throughout the eighteenth and nineteenth centuries through the work of researchers such as Coulomb, Ampère, Faraday and Maxwell.
In the 19th century it had become clear that electricity and magnetism were related, and their theories were unified: wherever charges are in motion electric current results, and magnetism is due to electric current. The source for electric field is electric charge, whereas that for magnetic field is electric current (charges in motion).
Ancient and classical history
The knowledge of static electricity dates back to the earliest civilizations, but for millennia it remained merely an interesting and mystifying phenomenon, without a theory to explain its behavior, and it was often confused with magnetism. The ancients were acquainted with rather curious properties possessed by two minerals, amber (, ) and magnetic iron ore ( , "the Magnesian stone, lodestone"). Amber, when rubbed, attracts lightweight objects, such as feathers; magnetic iron ore has the power of attracting iron.
Based on his find of an Olmec hematite artifact in Central America, the American astronomer John Carlson has suggested that "the Olmec may have discovered and used the geomagnetic lodestone compass earlier than 1000 BC". If true, this "predates the Chinese discovery of the geomagnetic lodestone compass by more than a millennium". Carlson speculates that the Olmecs may have used similar artifacts as a directional device for astrological or geomantic purposes, or to orient their temples, the dwellings of the living or the interments of the dead. The earliest Chinese literature reference to magnetism lies in a 4th-century BC book called Book of the Devil Valley Master (鬼谷子): "The lo
|
https://en.wikipedia.org/wiki/Antigenic%20variation
|
Antigenic variation or antigenic alteration refers to the mechanism by which an infectious agent such as a protozoan, bacterium or virus alters the proteins or carbohydrates on its surface and thus avoids a host immune response, making it one of the mechanisms of antigenic escape. It is related to phase variation. Antigenic variation not only enables the pathogen to avoid the immune response in its current host, but also allows re-infection of previously infected hosts. Immunity to re-infection is based on recognition of the antigens carried by the pathogen, which are "remembered" by the acquired immune response. If the pathogen's dominant antigen can be altered, the pathogen can then evade the host's acquired immune system. Antigenic variation can occur by altering a variety of surface molecules including proteins and carbohydrates. Antigenic variation can result from gene conversion, site-specific DNA inversions, hypermutation, or recombination of sequence cassettes. The result is that even a clonal population of pathogens expresses a heterogeneous phenotype. Many of the proteins known to show antigenic or phase variation are related to virulence.
In bacteria
Antigenic variation in bacteria is best demonstrated by species of the genus Neisseria (most notably, Neisseria meningitidis and Neisseria gonorrhoeae, the gonococcus); species of the genus Streptococcus and the Mycoplasma. The Neisseria species vary their pili (protein polymers made up of subunits called pilin which play a critical role in bacterial adhesion, and stimulate a vigorous host immune response) and the Streptococci vary their M-protein.
In the bacterium Borrelia burgdorferi, the cause of Lyme disease, the surface lipoprotein VlsE can undergo recombination which results in antigenic diversity. The bacterium carries a plasmid that contains fifteen silent vls cassettes and one functional copy of vlsE. Segments of the silent cassettes recombine with the vlsE gene, generating variants of the surfa
|
https://en.wikipedia.org/wiki/Voice%20inversion
|
Voice inversion scrambling is an analog method of obscuring the content of a transmission. It is sometimes used in public service radio, automobile racing, cordless telephones and the Family Radio Service. Without a descrambler, the transmission makes the speaker "sound like Donald Duck". Despite the term, the technique operates on the passband of the information and so can be applied to any information being transmitted.
Forms and details
There are various forms of voice inversion which offer differing levels of security. Overall, voice inversion scrambling offers little true security as software and even hobbyist kits are available from kit makers for scrambling and descrambling. The cadence of the speech is not changed. It is often easy to guess what is happening in the conversation by listening for other audio cues like questions, short responses and other language cadences.
In the simplest form of voice inversion, the frequency of each component is replaced with , where is the frequency of a carrier wave. This can be done by amplitude modulating the speech signal with the carrier, then applying a low-pass filter to select the lower sideband. This will make the low tones of the voice sound like high ones and vice versa. This process also occurs naturally if a radio receiver is tuned to a single sideband transmission but set to decode the wrong sideband.
There are more advanced forms of voice inversion which are more complex and require more effort to descramble. One method is to use a random code to choose the carrier frequency and then change this code in real time. This is called Rolling Code voice inversion and one can often hear the "ticks" in the transmission which signal the changing of the inversion point.
Another method is split band voice inversion. This is where the band is split and then each band is inverted separately. A rolling code can also be added to this method for variable split band inversion (VSB).
Common carrier frequencies are: 2.
|
https://en.wikipedia.org/wiki/Cliff%20effect
|
In telecommunications, the (digital) cliff effect or brickwall effect is a sudden loss of digital signal reception. Unlike analog signals, which gradually fade when signal strength decreases or electromagnetic interference or multipath increases, a digital signal provides data which is either perfect or non-existent at the receiving end. It is named for a graph of reception quality versus signal quality, where the digital signal "falls off a cliff" instead of having a gradual rolloff. This is an example of an EXIT chart.
The phenomenon is primarily seen in broadcasting, where signal strength is liable to vary, rather than in recorded media, which generally have a good signal. However, it may be seen in significantly damaged media, which is at the edge of readability.
Broadcasting
Digital television
This effect can most easily be seen on digital television, including both satellite TV and over-the-air terrestrial TV. While forward error correction is applied to the broadcast, when a minimum threshold of signal quality (a maximum bit error rate) is reached it is no longer enough for the decoder to recover. The picture may break up (macroblocking), lock on a freeze frame, or go blank. Causes include rain fade or solar transit on satellites, and temperature inversions and other weather or atmospheric conditions causing anomalous propagation on the ground.
Three particular issues particularly manifest the cliff effect. Firstly, anomalous conditions will cause occasional signal degradation. Secondly, if one is located in a fringe area, where the antenna is just barely strong enough to receive the signal, then usual variation in signal quality will cause relatively frequent signal degradation, and a very small change in overall signal quality can have a dramatic impact on the frequency of signal degradation – one incident per hour (not significantly affecting watchability) versus problems every few seconds or continuous problems. Thirdly, in some cases, where the sign
|
https://en.wikipedia.org/wiki/Russula%20cyanoxantha
|
Russula cyanoxantha, synonymous with R. xyanoxantha, commonly known as the charcoal burner or variegated russula, is a basidiomycete mushroom, distinguished from most other members of the genus Russula by the fact that its gills do not split, but are soft and flexible. It is one of the most common species of Russula in Europe.
It is an edible mushroom. It was designated "Mushroom of the Year" in 1997 by the German Association of Mycology.
Description
The most salient characteristic is the weak gills, which feel greasy to the touch, are flexible and do not break. The cap is wide, convex at first and later flattened, and greenish to bright brown; they vary considerably in color. The stipe is pure white, slightly convex underneath, from in height and in diameter. The spores are pure white. The stipe will give a green reaction when rubbed with iron salts (ferrous sulphate). Most other (but not all) Russula species give a salmon reaction. Coupled with the gill flexibility this is a good diagnostic clue to species level.
Distribution and habitat
Russula cyanoxantha grows in slightly acidic, but nutrient-rich soil. Like all Russulas, it is a mycorrhizal fungus. It is found most commonly in beech forests, and often in deciduous or mixed forests, appearing from May to November, with the highest concentration in July to September.
Use
The edible mushroom is suitable for many kinds of preparation; the flesh is not as hard as that of many other edible Russulas. It has a mild, nutty taste.
Similar species
The cap of the grey-green Russula grisea is more blue-grey but has violet or green hues with light cream gills; it also grows in mixed forests, particularly under beech, and more rarely in coniferous forests. Russula olivacea also may have a variegated cap, but produces yellow spores.
See also
List of Russula species
|
https://en.wikipedia.org/wiki/C-list%20%28computer%20security%29
|
In capability-based computer security, a C-list is an array of capabilities, usually associated with a process and maintained by the kernel. The program running in the process does not manipulate capabilities directly, but refers to them via C-list indexes—integers indexing into the C-list.
The file descriptor table in Unix is an example of a C-list. Unix processes do not manipulate file descriptors directly, but refer to them via file descriptor numbers, which are C-list indexes.
In the KeyKOS and EROS operating systems, a process's capability registers constitute a C-list.
See also
Access-control list
|
https://en.wikipedia.org/wiki/BTRON
|
BTRON (Business TRON) is one of the subprojects of the TRON Project proposed by Ken Sakamura, which is responsible for the business phase. It refers to the operating systems (OS), keyboards, peripheral interface specifications, and other items related to personal computers (PCs) that were developed there.
Originally, it refers to specifications rather than specific products, but in reality, the term "BTRON" is often used to refer to implementations. Currently, Personal Media Corporation's B-right/V is an implementation of BTRON3, and a software product called "" that includes it has been released.
Specifications
As with other TRON systems, only the specification of BTRON has been formulated, and the implementation method is not specified. Implementation is mentioned in this section to the extent necessary to explain the specification, but please refer to the Implementation section for details.
BTRON1, BTRON2, BTRON3
The BTRON project began with Matsushita Electric Industrial and Personal Media prototyping "BTRON286," an implementation on a 16-bit CPU 286 for the CEC machine described below. BTRON1 specifications include the BTRON1 Programming Standard Handbook, which describes the OS API, and the BTRON1 Specification Software Specification. which describes the OS API.
BTRON2 is planned to be implemented on , and only the specification has been created and published. It is planned to be implemented on evaluation machines equipped with TRON chips made by Fujitsu and named "2B". One of its features is that all OS-managed computing resources such as memory, processes, and threads are handled in a real/pseudomorphic model, a feature of BTRON.
SIGBTRON's TRON chip machine MCUBE implemented "3B," which is 32-bit and uses an ITRON-specification RTOS (modified from "ItIs") for the microkernel. 3B and The B-right specification used in , etc. is "BTRON3" (currently, the microkernel is I-right); the specification that B-right/V conforms to is published as the BTRON3 spec
|
https://en.wikipedia.org/wiki/Solar%20transit
|
In astronomy, a solar transit is a movement of any object passing between the Sun and the Earth. This includes the planets Mercury and Venus (see Transit of Mercury and Transit of Venus). A solar eclipse is also a solar transit of the Moon, but technically only if it does not cover the entire disc of the Sun (an annular eclipse), as "transit" counts only objects that are smaller than what they are passing in front of. Solar transit is only one of several types of astronomical transit
A solar transit (also called a solar outage, sometimes solar fade, sun outage, or sun fade) also occurs to communications satellites, which pass in front of the Sun for several minutes each day for several days straight for a period in the months around the equinoxes, the exact dates depending on where the satellite is in the sky relative to its earth station. Because the Sun also produces a great deal of microwave radiation in addition to sunlight, it overwhelms the microwave radio signals coming from the satellite's transponders. This enormous electromagnetic interference causes interruptions in fixed satellite services that use satellite dishes, including TV networks and radio networks, as well as VSAT and DBS.
Only downlinks from the satellite are affected, uplinks from the Earth are normally not, as the planet "shades" the Earth station when viewed from the satellite. Satellites in geosynchronous orbit are irregularly affected based on their inclination. Reception from satellites in other orbits are frequently but only momentarily affected by this, and by their nature the same signal is usually repeated or relayed on another satellite, if a tracking dish is used at all. Satellite radio and other services like GPS are not affected, as they use no receiving dish, and therefore do not concentrate the interference. (GPS and certain satellite radio systems use non-geosynchronous satellites.)
Solar transit begins with only a brief degradation in signal quality for a few moments.
|
https://en.wikipedia.org/wiki/Cleft%20palate%20incidence%20by%20population
|
Cleft lip and/or palate is a congenital abnormality that is seen frequently around the world. On average, about 1 in every 500-750 live births result in a cleft (Hardin-Jones, Karnell, & Peterson-Falzone, 2001). Furthermore, in the U.S., the prevalence for cleft lip with or without cleft palate (CL +/- P) is 2.2 to 11.7 per 10,000 births. Cleft palate alone (CP) results in a prevalence rate of 5.5 to 6.6 per 10,000 births (Forrester & Merz, 2004). Cleft of the lip, palate, or both is one of the most common congenital abnormalities and has a birth prevalence rate ranging from 1/1000 to 2.69/1000 amongst different parts of the world (McLeod, Saeed, & Arana- Urioste, 2004).
Africans and African Americans
A look into the prevalence rates of different cultures in the U.S. when compared to country of origin begins with Africans and African Americans. One per 2500 African Americans are born with a cleft (Suleiman, Hamzah, Abusalab, & Samaan, 2005). African Americans have a lower prevalence rate of CL +/- P when compared to Caucasians. A prevalence rate of 0.61 per 1,000 and 1.05 per 1,000 live births respectively was reported by Croen, Shaw, Wasserman and Tolarova (1998). In Malawi there is a reported low prevalence rate for cleft lip and/or palate, 0.7 per 1,000 live births (Chisi, Igbibi, & Msamati, 2000). Suleiman et al. (2005) found that the prevalence rate of clefting among a group of Sudanese hospital new-borns in the city of Khartoum is 0.9 per 1,000 live births.
Mestizo Americans
Mestizo Americans are people of mixed European and Native American origins from Mexico, Central America and South America, and the Caribbean (Meyerson, 1990). The prevalence of Mestizo Americans is lower than that of Caucasians and Native Americans, yet it is still higher than African Americans (Croen et al., 1998). Mestizo have a prevalence of clefting of 9.7 per 10,000 live births (Kirby, Petrini & Alter, 2000). In Sucre, Bolivia the prevalence rate of CL +/- P is 1.23 per 1,000 live b
|
https://en.wikipedia.org/wiki/The%20Fairylogue%20and%20Radio-Plays
|
The Fairylogue and Radio-Plays was an early attempt to bring L. Frank Baum's Oz books to the motion picture screen. It was a mixture of live actors, hand-tinted magic lantern slides, and film. Baum himself would appear as if he were giving a lecture, while he interacted with the characters (both on stage and on screen). Although acclaimed throughout its tour, the show experienced budgetary problems (with the show costing more to produce than the money that sold-out houses could bring in) and folded after two months of performances. It opened in Grand Rapids, Michigan on September 24, 1908. It then opened in Orchestra Hall in Chicago on October 1, toured the country and ended its run in New York City. There, it was scheduled to run through December 31, and ads for it continued to run in The New York Times until then, but it reportedly closed on December 16.
After First National Pictures acquired Selig Polyscope, the film was re-released on September 24, 1925.
Although today seen mostly as a failed first effort to adapt the Oz books, The Fairylogue and Radio-Plays is notable in film history because it contains the earliest original film score to be documented.
The film is lost, but the script for Baum's narration and production stills survive.
Michael Radio Color
The films were colored (credited as "illuminations") by Duval Frères of Paris, in a process known as "Radio-Play", and were noted for being the most lifelike hand-tinted imagery of the time. Baum once claimed in an interview that a "Michael Radio" was a Frenchman who colored the films, though no evidence of such a person, even with the more proper French spelling "Michel", as second-hand reports unsurprisingly revise it, has been documented. It did not refer to the contemporary concept of radio (or, for that matter, a radio play), but played on notions of the new and fantastic at the time, similar to the way "high-tech" or sometimes "cyber" would be used later in the century. The "Fairylogue" part of
|
https://en.wikipedia.org/wiki/Mock%20modular%20form
|
In mathematics, a mock modular form is the holomorphic part of a harmonic weak Maass form, and a mock theta function is essentially a mock modular form of weight . The first examples of mock theta functions were described by Srinivasa Ramanujan in his last 1920 letter to G. H. Hardy and in his lost notebook. Sander Zwegers discovered that adding certain non-holomorphic functions to them turns them into harmonic weak Maass forms.
History
Ramanujan's 12 January 1920 letter to Hardy listed 17 examples of functions that he called mock theta functions, and his lost notebook contained several more examples. (Ramanujan used the term "theta function" for what today would be called a modular form.) Ramanujan pointed out that they have an asymptotic expansion at the cusps, similar to that of modular forms of weight , possibly with poles at cusps, but cannot be expressed in terms of "ordinary" theta functions. He called functions with similar properties "mock theta functions". Zwegers later discovered the connection of the mock theta function with weak Maass forms.
Ramanujan associated an order to his mock theta functions, which was not clearly defined. Before the work of Zwegers, the orders of known mock theta functions included
3, 5, 6, 7, 8, 10.
Ramanujan's notion of order later turned out to correspond to the conductor of the Nebentypus character of the weight harmonic Maass forms which admit Ramanujan's mock theta functions as their holomorphic projections.
In the next few decades, Ramanujan's mock theta functions were studied by Watson, Andrews, Selberg, Hickerson, Choi, McIntosh, and others, who proved Ramanujan's statements about them and found several more examples and identities. (Most of the "new" identities and examples were already known to Ramanujan and reappeared in his lost notebook.) In 1936, Watson found that under the action of elements of the modular group, the order 3 mock theta functions almost transform like modular forms of weight (multiplied by
|
https://en.wikipedia.org/wiki/TORCH%20syndrome
|
TORCH syndrome is a cluster of symptoms caused by congenital infection with toxoplasmosis, rubella, cytomegalovirus, herpes simplex, and other organisms including syphilis, parvovirus, and Varicella zoster. Zika virus is considered the most recent member of TORCH infections.
TORCH is an acronym for , Agents, , , and Simplex.
Signs and symptoms
Though caused by different infections, the signs and symptoms of TORCH syndrome are consistent. They include hepatosplenomegaly (enlargement of the liver and spleen), fever, lethargy, difficulty feeding, anemia, petechiae, purpurae, jaundice, and chorioretinitis. The specific infection may cause additional symptoms.
TORCH syndrome may develop before birth, causing stillbirth, in the neonatal period, or later in life.
Pathophysiology
TORCH syndrome is caused by in-utero infection with one of the TORCH agents, disrupting fetal development.
Diagnosis
Presence of IgM is diagnostic and persistence of IgG beyond 6–9 months is diagnostic.
Prevention
TORCH syndrome can be prevented by treating an infected pregnant woman, thereby preventing the infection from affecting the fetus.
Treatment
The treatment of TORCH syndrome is mainly supportive and depends on the symptoms present; medication is an option for herpes and cytomegalovirus infections.
Epidemiology
Developing countries are more severely affected by TORCH syndrome.
|
https://en.wikipedia.org/wiki/Digital%20delay%20generator
|
A digital delay generator (also known as digital-to-time converter) is a piece of electronic test equipment that provides precise delays for triggering, syncing, delaying, and gating events. These generators are used in many experiments, controls, and processes where electronic timing of a single event or multiple events to a standard timing reference is needed. The digital delay generator may initiate a sequence of events or be triggered by an event. What differentiates it from ordinary electronic timing is the synchronicity of its outputs to each other and the initiating event.
A time-to-digital converter does the inverse function.
Equipment
The digital delay generator is similar to a pulse generator in function, but the timing resolution is much finer, and the delay and width jitter much less.
Some manufacturers, calling their units "digital delay and pulse generators", have added independent amplitude polarity and level control to each of their outputs in addition to delay and width control. Now each channel provides its delay, width, and amplitude control, with the triggering synchronized to an external source or internal rep rate generator - like a general-purpose pulse generator.
Some delay generators provide precise delays (edges) to trigger devices. Others provide accurate delays and widths also to allow a gating function. Some delay generators provide a single timing channel, while others provide multiple timing channels.
Digital delay generator outputs are typically logic level, but some offer higher voltages to cope with electromagnetic interference environments. For very harsh environments, optical outputs and/or inputs with fiber optic connectors are also offered as options by some manufacturers. In general, a delay generator operates in a 50 Ω transmission line environment with the line terminated in its characteristic impedance to minimize reflections and timing ambiguities.
Historically, digital delay generators were single channel devices wit
|
https://en.wikipedia.org/wiki/Pointcheval%E2%80%93Stern%20signature%20algorithm
|
In cryptography, the Pointcheval–Stern signature algorithm is a digital signature scheme based on the closely related ElGamal signature scheme. It changes the ElGamal scheme slightly to produce an algorithm which has been proven secure in a strong sense against adaptive chosen-message attacks, assuming the discrete logarithm problem is intractable in a strong sense.
David Pointcheval and Jacques Stern developed the forking lemma technique in constructing their proof for this algorithm. It has been used in other security investigations of various cryptographic algorithms.
|
https://en.wikipedia.org/wiki/Bourbaki%20dangerous%20bend%20symbol
|
The dangerous bend or caution symbol ☡ () was created by the Nicolas Bourbaki group of mathematicians and appears in the margins of mathematics books written by the group. It resembles a road sign that indicates a "dangerous bend" in the road ahead, and is used to mark passages tricky on a first reading or with an especially difficult argument.
Variations
Others have used variations of the symbol in their books. The computer scientist Donald Knuth introduced an American-style road-sign depiction in his Metafont and TeX systems, with a pair of adjacent signs indicating doubly dangerous passages.
Typography
In the LaTeX typesetting system, Knuth's dangerous bend symbol can be produced by
first loading the font manfnt (a font with extra symbols used in Knuth's TeX manual) with
\usepackage{manfnt}
and then typing
\dbend
There are several variations given by \lhdbend, \reversedvideodbend, \textdbend, \textlhdbend, and \textreversedvideodbend.
See also
Halmos box
|
https://en.wikipedia.org/wiki/Series%20of%20tubes
|
"A series of tubes" is a phrase used originally as an analogy by then-United States Senator Ted Stevens (R-Alaska) to describe the Internet in the context of opposing network neutrality. On June 28, 2006, he used this metaphor to criticize a proposed amendment to a committee bill. The amendment would have prohibited Internet service providers such as AT&T, Comcast, Time Warner Cable and Verizon Communications from charging fees to give some companies' data a higher priority in relation to other traffic. The metaphor was widely ridiculed, because Stevens was perceived to have displayed an extremely limited understanding of the Internet, despite his leading the Senate committee responsible for regulating it.
Partial text of Stevens's comments
Media commentary
On June 28, 2006, Public Knowledge government affairs manager Alex Curtis wrote a brief blog entry introducing the senator's speech and posted an MP3 recording. The next day, the Wired magazine blog 27B Stroke 6 featured a lengthier post by Ryan Singel, which included Singel's transcriptions of some parts of Stevens's speech considered the most humorous. Within days, thousands of other blogs and message boards posted the story.
Most writers and commentators derisively cited several of Stevens's misunderstandings of Internet technology, arguing that the speech showed that he had formed a strong opinion on a topic which he understood poorly (e.g., referring to an e-mail message as "an Internet," and blaming bandwidth issues for an e-mail problem much more likely to be caused by mail server or routing issues). The story sparked mainstream media attention, including a mention in The New York Times. The technology podcast This Week in Tech also discussed the incident.
According to The Wall Street Journal, as summarized by MediaPost commentator Ross Fadner, "'The Internet is a Series of Tubes!' spawned a new slogan that became a rallying cry for Net neutrality advocates. ... Stevens's overly simplistic description
|
https://en.wikipedia.org/wiki/Portal%20%28video%20game%29
|
Portal is a 2007 puzzle-platform game developed and published by Valve. It was released in a bundle, The Orange Box, for Windows, Xbox 360 and PlayStation 3, and has been since ported to other systems, including Mac OS X, Linux, Android (via Nvidia Shield), and Nintendo Switch.
Portal consists primarily of a series of puzzles that must be solved by teleporting the player's character and simple objects using "the Aperture Science Handheld Portal Device", often referred to as the "portal gun", a device that can create inter-spatial portals between two flat planes. The player-character, Chell, is challenged and taunted by an artificial intelligence named GLaDOS (Genetic Lifeform and Disk Operating System) to complete each puzzle in the Aperture Science Enrichment Center using the portal gun with the promise of receiving cake when all the puzzles are completed. The game's unique physics allows kinetic energy to be retained through portals, requiring creative use of portals to maneuver through the test chambers. This gameplay element is based on a similar concept from the game Narbacular Drop; many of the team members from the DigiPen Institute of Technology who worked on Narbacular Drop were hired by Valve for the creation of Portal, making it a spiritual successor to the game.
Portal was acclaimed as one of the most original games of 2007, despite some criticism for its short duration. It received praise for its originality, unique gameplay and dark story with a humorous series of dialogue. GLaDOS, voiced by Ellen McLain in the English-language version, received acclaim for her unique characterization, and the end credits song "Still Alive", written by Jonathan Coulton for the game, was praised for its original composition and humorous twist. Portal is often cited as one of the greatest video games ever made. Excluding Steam download sales, over four million copies of the game have been sold since its release, spawning official merchandise from Valve including plush
|
https://en.wikipedia.org/wiki/Indeterminate%20system
|
In mathematics, particularly in algebra, an indeterminate system is a system of simultaneous equations (e.g., linear equations) which has more than one solution (sometimes infinitely many solutions). In the case of a linear system, the system may be said to be underspecified, in which case the presence of more than one solution would imply an infinite number of solutions (since the system would be describable in terms of at least one free variable), but that property does not extend to nonlinear systems (e.g., the system with the equation ).
An indeterminate system by definition is consistent, in the sense of having at least one solution. For a system of linear equations, the number of equations in an indeterminate system could be the same as the number of unknowns, less than the number of unknowns (an underdetermined system), or greater than the number of unknowns (an overdetermined system). Conversely, any of those three cases may or may not be indeterminate.
Examples
The following examples of indeterminate systems of equations have respectively, fewer equations than, as many equations as, and more equations than unknowns:
Conditions giving rise to indeterminacy
In linear systems, indeterminacy occurs if and only if the number of independent equations (the rank of the augmented matrix of the system) is less than the number of unknowns and is the same as the rank of the coefficient matrix. For if there are at least as many independent equations as unknowns, that will eliminate any stretches of overlap of the equations' surfaces in the geometric space of the unknowns (aside from possibly a single point), which in turn excludes the possibility of having more than one solution. On the other hand, if the rank of the augmented matrix exceeds (necessarily by one, if at all) the rank of the coefficient matrix, then the equations will jointly contradict each other, which excludes the possibility of having any solution.
Finding the solution set of an indeterminate lin
|
https://en.wikipedia.org/wiki/Independent%20equation
|
An independent equation is an equation in a system of simultaneous equations which cannot be derived algebraically from the other equations. The concept typically arises in the context of linear equations. If it is possible to duplicate one of the equations in a system by multiplying each of the other equations by some number (potentially a different number for each equation) and summing the resulting equations, then that equation is dependent on the others. But if this is not possible, then that equation is independent of the others.
If an equation is independent of the other equations in its system, then it provides information beyond that which is provided by the other equations. In contrast, if an equation is dependent on the others, then it provides no information not contained in the others collectively, and the equation can be dropped from the system without any information loss.
The number of independent equations in a system equals the rank of the augmented matrix of the system—the system's coefficient matrix with one additional column appended, that column being the column vector of constants.
The number of independent equations in a system of consistent equations (a system that has at least one solution) can never be greater than the number of unknowns. Equivalently, if a system has more independent equations than unknowns, it is inconsistent and has no solutions.
See also
Linear algebra
Indeterminate system
Independent variable
|
https://en.wikipedia.org/wiki/Ribonomics
|
Ribonomics is the study of ribonucleic acids (RNAs) associated with RNA-binding proteins (RBPs). The term was introduced by Robert Cedergren and colleagues who used a bioinformatic search tool to discover novel ribozymes and RNA motifs originally found in HIV.
Ribonomics, like genomics or proteomics, is the large-scale, high-throughput approach to identifying subsets of RNAs by their association with proteins in cells. Since many messenger RNAs (mRNAs) are linked with multiple processes, this technique offers a facile mechanism to study the relationship of various intracellular systems.
Prokaryotes co-regulate genes common to cellular processes via a polycistronic operon. Since eukaryotic transcription produces mRNA encoding proteins in a monocistronic fashion, many gene products must be concomitantly expressed (see gene expression) and translated in a timed fashion. RBPs are thought to be the molecules which physically and biochemically organize these messages to different cellular locales where they may be translated, degraded or stored. The study of transcripts associated with RBPs is therefore thought to be important in eukaryotes as a mechanism for coordinated gene regulation. The likely biochemical processes which account for this regulation are the expedited/delayed degradation of RNA. In addition to the influence on RNA half-life, translation rates are also thought to be altered by RNA-protein interactions.
The Drosophila ELAV family, the Puf family in yeast, and the human La, Ro, and FMR proteins are known examples of RBPs, showing the diverse species and processes with which post-transcriptional gene regulation is associated.
See also
ELAVL1
ELAVL2
ELAVL3
Transcript of unknown function
|
https://en.wikipedia.org/wiki/A%C3%A7a%C3%AD%20oil
|
Açaí oil is obtained from the fruit of Euterpe oleracea (açaí palm), which grows in the Amazon rainforest. The oil is rich in phenolic compounds similar in profile to the pulp itself, such as vanillic acid, syringic acid, p-hydroxybenzoic acid, protocatechuic acid and ferulic acid as well as (+)-catechin and numerous procyanidin oligomers.
Açai oil is green in color and has a bland aroma. It is high in oleic and palmitic fatty acids (table).
Uses
Açai oil is widely used for cooking and as a salad dressing. In cosmetics, it is used in shampoos, soaps and skin moisturizers.
Source:
See also
Açaí palm
|
https://en.wikipedia.org/wiki/Trimetrexate
|
Trimetrexate is a quinazoline derivative. It is a dihydrofolate reductase inhibitor.
Uses
It has been used with leucovorin in treating pneumocystis pneumonia.
It has been investigated for use in treating leiomyosarcoma.
It is a methotrexate (MTX) analog that is active against transport-deficient MTX-resistant tumor cells that overcome the acquired and natural resistance to methotrexate. Other uses include skin lymphoma.
|
https://en.wikipedia.org/wiki/Local%20shared%20object
|
A local shared object (LSO), commonly called a Flash cookie (due to its similarity with an HTTP cookie), is a piece of data that websites that use Adobe Flash may store on a user's computer. Local shared objects have been used by all versions of Flash Player (developed by Macromedia, which was later acquired by Adobe Systems) since version 6.
Flash cookies, which can be stored or retrieved whenever a user accesses a page containing a Flash application, are a form of local storage. Similar to cookies, they can be used to store user preferences, save data from Flash games, or track users' Internet activity. LSOs have been criticised as a breach of browser security, but there are now browser settings and addons to limit the duration of their storage.
Storage
Local shared objects contain data stored by individual websites. Data is stored in the Action Message Format. With the default settings, the Flash Player does not seek the user's permission to store local shared objects on the hard disk. By default, an SWF application running in Flash Player from version 9 to 11 (as of Sept 1, 2011) may store up to of data to the user's hard drive. If the application attempts to store more, a dialog asks the user whether to allow or deny the request.
Adobe Flash Player does not allow third-party local shared objects to be shared across domains. For example, a local shared object from "www.example.com" cannot be read by the domain "www.example.net". However, the first-party website can always pass data to a third-party via some settings found in the dedicated XML file and passing the data in the request to the third party. Also, third-party LSOs are allowed to store data by default. By default, LSO data is shared across browsers on the same machine. As an example:
A visitor accesses a site using their Firefox browser, then views a page displaying a specific product, then closes the Firefox browser, the information about that product can be stored in the LSO.
If that same v
|
https://en.wikipedia.org/wiki/Vector%20notation
|
In mathematics and physics, vector notation is a commonly used notation for representing vectors, which may be Euclidean vectors, or more generally, members of a vector space.
For representing a vector, the common typographic convention is lower case, upright boldface type, as in . The International Organization for Standardization (ISO) recommends either bold italic serif, as in , or non-bold italic serif accented by a right arrow, as in .
In advanced mathematics, vectors are often represented in a simple italic type, like any variable.
History
In 1835 Giusto Bellavitis introduced the idea of equipollent directed line segments which resulted in the concept of a vector as an equivalence class of such segments.
The term vector was coined by W. R. Hamilton around 1843, as he revealed quaternions, a system which uses vectors and scalars to span a four-dimensional space. For a quaternion q = a + bi + cj + dk, Hamilton used two projections: S q = a, for the scalar part of q, and V q = bi + cj + dk, the vector part. Using the modern terms cross product (×) and dot product (.), the quaternion product of two vectors p and q can be written pq = –p.q + p×q. In 1878, W. K. Clifford severed the two products to make the quaternion operation useful for students in his textbook Elements of Dynamic. Lecturing at Yale University, Josiah Willard Gibbs supplied notation for the scalar product and vector products, which was introduced in Vector Analysis.
In 1891, Oliver Heaviside argued for Clarendon to distinguish vectors from scalars. He criticized the use of Greek letters by Tait and Gothic letters by Maxwell.
In 1912, J.B. Shaw contributed his "Comparative Notation for Vector Expressions" to the Bulletin of the Quaternion Society. Subsequently, Alexander Macfarlane described 15 criteria for clear expression with vectors in the same publication.
Vector ideas were advanced by Hermann Grassmann in 1841, and again in 1862 in the German language. But German mathematicians wer
|
https://en.wikipedia.org/wiki/Amazon%20S3
|
Amazon S3 or Amazon Simple Storage Service is a service offered by Amazon Web Services (AWS) that provides object storage through a web service interface. Amazon S3 uses the same scalable storage infrastructure that Amazon.com uses to run its e-commerce network. Amazon S3 can store any type of object, which allows uses like storage for Internet applications, backups, disaster recovery, data archives, data lakes for analytics, and hybrid cloud storage. AWS launched Amazon S3 in the United States on March 14, 2006, then in Europe in November 2007.
Technical details
Design
Amazon S3 manages data with an object storage architecture which aims to provide scalability, high availability, and low latency with high durability. The basic storage units of Amazon S3 are objects which are organized into buckets. Each object is identified by a unique, user-assigned key. Buckets can be managed using the console provided by Amazon S3, programmatically with the AWS SDK, or the REST application programming interface. Objects can be up to five terabytes in size. Requests are authorized using an access control list associated with each object bucket and support versioning which is disabled by default. Since buckets are typically the size of an entire file system mount in other systems, this access control scheme is very coarse-grained. In other words, unique access controls cannot be associated with individual files. Amazon S3 can be used to replace static web-hosting infrastructure with HTTP client-accessible objects, index document support, and error document support.
The Amazon AWS authentication mechanism allows the creation of authenticated URLs, valid for a specified amount of time. Every item in a bucket can also be served as a BitTorrent feed. The Amazon S3 store can act as a seed host for a torrent and any BitTorrent client can retrieve the file. This can drastically reduce the bandwidth cost for the download of popular objects. A bucket can be configured to save HTTP log i
|
https://en.wikipedia.org/wiki/Trace%20fossil%20classification
|
Trace fossils are classified in various ways for different purposes. Traces can be classified taxonomically (by morphology), ethologically (by behavior), and toponomically, that is, according to their relationship to the surrounding sedimentary layers. Except in the rare cases where the original maker of a trace fossil can be identified with confidence, phylogenetic classification of trace fossils is an unreasonable proposition.
Taxonomic classification
The taxonomic classification of trace fossils parallels the taxonomic classification of organisms under the International Code of Zoological Nomenclature. In trace fossil nomenclature a Latin binomial name is used, just as in animal and plant taxonomy, with a genus and specific epithet. However, the binomial names are not linked to an organism, but rather just a trace fossil. This is due to the rarity of association between a trace fossil and a specific organism or group of organisms. Trace fossils are therefore included in an ichnotaxon separate from Linnaean taxonomy. When referring to trace fossils, the terms ichnogenus and ichnospecies parallel genus and species respectively.
The most promising cases of phylogenetic classification are those in which similar trace fossils show details complex enough to deduce the makers, such as bryozoan borings, large trilobite trace fossils such as Cruziana, and vertebrate footprints. However, most trace fossils lack sufficiently complex details to allow such classification.
Ethologic classification
The Seilacherian System
Adolf Seilacher was the first to propose a broadly accepted ethological basis for trace fossil classification. He recognized that most trace fossils are created by animals in one of five main behavioural activities, and named them accordingly:
Cubichnia are the traces of organisms left on the surface of a soft sediment. This behaviour may simply be resting as in the case of a starfish, but might also evidence the hiding place of prey, or even the ambus
|
https://en.wikipedia.org/wiki/European%20Conference%20on%20Wireless%20Sensor%20Networks
|
The European Conference on Wireless Sensor Networks (EWSN) is an annual academic conference on wireless sensor networks.
Although there is no official ranking of academic conferences on wireless sensor networks, EWSN is widely regarded as the top European event in sensor networks.
EWSN Events
EWSN started in year 2004:
EWSN 2015, Porto, Portugal, February 9–11, 2015
EWSN 2014, Oxford, UK, February 17–19, 2014
EWSN 2013 , Ghent, Belgium, February 13–15, 2013
EWSN 2012, Trento, Italy, February 15–17, 2012
EWSN 2011, Bonn, Germany, February 23.25, 2011
EWSN 2010, Coimbra, Portugal, February 17–19, 2010
EWSN 2009, Cork, Ireland, February 11–13, 2009
EWSN 2008, Bologna, Italy, January 30–31, February 1, 2008
EWSN 2007, Delft, The Netherlands, January 29–31, 2007
EWSN 2006, Zurich, Switzerland, February 13–15, 2006
EWSN 2005, Istanbul, Turkey, January 31 - February 2, 2005
EWSN 2004, Berlin, Germany, January 19–21, 2004
History
EWSN started in year 2004 and the prime motivation behind EWSN was to provide the European researchers working in sensor networks a venue to disseminate their research results. However, over the years EWSN has grown into a truly International event with participants and authors coming from all over the world. In 2006 it was decided to silently upgrade the event from a workshop to a conference. With this change in effect the acronym (i.e. EWSN) remains the same. Therefore, when giving a reference to EWSN 2004 to 2006 use European Workshop on Wireless Sensor Networks, and when giving a reference to EWSN 2007 onwards use European Conference on Wireless Sensor Networks.
See also
Wireless sensor network
External links
EWSN Bibliography (from DBLP)
Wireless sensor network
Computer networking conferences
|
https://en.wikipedia.org/wiki/Cervical%20canal
|
The cervical canal is the spindle-shaped, flattened canal of the cervix, the neck of the uterus.
Anatomy
The cervical canal communicates with the uterine cavity via the internal orifice of the uterus (or internal os) and with the vagina via the external orifice of the uterus (ostium of uterus or external os). The internal orifice of the uterus is an interior narrowing of the uterine cavity. It corresponds to a slight constriction known as the isthmus that can be seen on the surface of the uterus about midway between the apex and base. The external orifice of the uterus is a small, depressed, somewhat circular opening on the rounded extremity of the cervix, opening to the vagina. Through this aperture, the cervical cavity communicates with that of the vagina.
The external orifice is bounded by two lips, an anterior and a posterior. The anterior is shorter and thicker, though it projects lower than the posterior because of the slope of the cervix. Normally, both lips are in contact with the posterior vaginal wall. Prior to pregnancy the external orifice has a rounded shape when viewed through the vaginal canal (as through a speculum). Following childbirth, the orifice takes on an appearance more like a transverse slit or is "H-shaped".
The wall of the canal presents an anterior and a posterior longitudinal ridge, from each of which proceed a number of small oblique columns, the palmate folds, giving the appearance of branches from the stem of a tree; to this arrangement the name arbor vitae uteri is applied.
The folds on the two walls are not exactly opposed, but fit between one another so as to close the cervical canal.
Histology
The cervical canal is generally lined by "endocervical mucosa" which consists of a single layer of mucinous columnar epithelium. However, after menopause, the functional squamocolumnar junction moves into the cervical canal, and hence the distal part of the cervical canal may be lined by stratified squamous epithelium (conforming to a
|
https://en.wikipedia.org/wiki/Uterine%20cavity
|
The uterine cavity is the inside of the uterus. It is triangular in shape, the base (broadest part) being formed by the internal surface of the body of the uterus between the openings of the fallopian tubes, the apex by the internal orifice of the uterus through which the cavity of the body communicates with the canal of the cervix. The uterine cavity where it enters the openings of the fallopian tubes is a mere slit, flattened antero-posteriorly.
|
https://en.wikipedia.org/wiki/144%2C000
|
144,000 is a natural number. It has significance in various religious movements and ancient prophetic belief systems.
Religion
Christianity
Book of Revelation
The number 144,000 appears three times in the Book of Revelation:
Revelation 7:3–8:
Revelation 14:1:
Revelation 14:3–5:
The numbers 12,000 and 144,000 are variously interpreted in traditional Christianity. Some, taking the numbers in Revelation to be symbolic, believe it represents all of God's people throughout history in the heavenly Church. One suggestion is that the number comes from 12, a symbol for totality, which is squared and multiplied by one thousand for more emphasis. Others insist the numbers 12,000 and 144,000 are literal numbers and representing either descendants of Jacob (also called Israel in the Bible) or others to whom God has given a superior destiny with a distinct role at the time of the end of the world.
One understanding is that the 144,000 are recently converted Jewish evangelists sent out to bring sinners to Jesus Christ during the seven year tribulation period. Preterists believe they are Jewish Christians, sealed for deliverance from the destruction of Jerusalem in 70 A.D. Dispensationalist Tim LaHaye, in his commentary Revelation: Illustrated and Made Plain (Zondervan, 1975), considers the 144,000 in Revelation 7 to refer to Jews and those in Revelation 14 to refer to Christians.
Jehovah's Witnesses
Jehovah's Witnesses believe that exactly 144,000 faithful Christians from Pentecost of 33 AD until the present day will be resurrected to heaven as immortal spirit beings to spend eternity with God and Christ. They believe that these people are "anointed" by God to become part of the spiritual "Israel of God". They believe the 144,000 (which they consider to be synonymous with the "little flock" of Luke 12:32) will serve with Christ as king-priests for a thousand years, while all other people accepted by God (the "other sheep" of John 10:16, composed of "the great crowd" o
|
https://en.wikipedia.org/wiki/Mash%20ingredients
|
Mash ingredients, mash bill, mashbill, or grain bill are the materials that brewers use to produce the wort that they then ferment into alcohol. Mashing is the act of creating and extracting fermentable and non-fermentable sugars and flavor components from grain by steeping it in hot water, and then letting it rest at specific temperature ranges to activate naturally occurring enzymes in the grain that convert starches to sugars. The sugars separate from the mash ingredients, and then yeast in the brewing process converts them to alcohol and other fermentation products.
A typical primary mash ingredient is grain that has been malted. Modern-day malt recipes generally consist of a large percentage of a light malt and, optionally, smaller percentages of more flavorful or highly colored types of malt. The former is called "base malt"; the latter is known as "specialty malts" .
The grain bill of a beer or whisky may vary widely in the number and proportion of ingredients. For example, in beer-making, a simple pale ale might contain a single malted grain, while a complex porter may contain a dozen or more ingredients. In whisky production, Bourbon uses a mash made primarily from maize (often mixed with rye or wheat and a small amount of malted barley), and single malt Scotch exclusively uses malted barley.
Variables
Each particular ingredient has its own flavor that contributes to the final character of the beverage. In addition, different ingredients carry other characteristics, not directly relating to the flavor, which may dictate some of the choices made in brewing: nitrogen content, diastatic power, color, modification, and conversion.
Nitrogen content
The nitrogen content of a grain relates to the mass fraction of the grain that is made up of protein, and is usually expressed as a percentage; this fraction is further refined by distinguishing what fraction of the protein is water-soluble, also usually expressed as a percentage; 40% is typical for most beermakin
|
https://en.wikipedia.org/wiki/Conference%20on%20Embedded%20Networked%20Sensor%20Systems
|
SenSys, the ACM Conference on Embedded Networked Sensor Systems, is an annual academic conference in the area of embedded networked sensors.
About SenSys
ACM SenSys is a selective, single-track forum for the presentation of research results on systems issues in the area of embedded networked sensors. The conference provides a venue to address the research challenges facing the design, deployment, use, and fundamental limits of these systems. Sensor networks require contributions from many fields, from wireless communication and networking, embedded systems and hardware, distributed systems, data management, and applications. SenSys welcomes cross-disciplinary work.
Ranking
Although there is no official ranking of academic conferences on wireless sensor networks, SenSys is widely regarded by researchers as one of the two (along with IPSN) most prestigious conferences focusing on sensor network research. SenSys focuses more on system issues while IPSN on algorithmic and theoretical considerations. The acceptance rate for 2017 was 17.2% (26 out of 151 papers accepted for publication).
SenSys Events
SenSys started in year 2003 and following is a list of SenSys events from 2003 to 2017:
SenSys 2017, Delft, The Netherlands, November 5–8, 2017
SenSys 2016, Stanford, CA, USA, November 14–16, 2016
SenSys 2015, Seoul, South Korea, November 1–4, 2015
SenSys 2014, Memphis, Tennessee, USA, November 3–6, 2014
SenSys 2013, Rome, Italy, November 11–14, 2013
SenSys 2012, Toronto, Canada, November 6–9, 2012
SenSys 2011, Seattle, WA, USA, November 1–4, 2011
SenSys 2010, Zurich, Switzerland, November 3–5, 2010
SenSys 2009, Berkeley, California, USA, November 4–6, 2009
SenSys 2008, Raleigh, North Carolina, USA, November 5–7, 2008
SenSys 2007, Sydney, Australia, November 6–9, 2007
SenSys 2006, Boulder, Colorado, USA, November 1–3, 2006
SenSys 2005, San Diego, CA, USA, November 2–4, 2005
SenSys 2004, Baltimore, MD, USA, November 3–5, 2004
SenSys 2003, Los Angeles, Calif
|
https://en.wikipedia.org/wiki/Human-body%20model
|
The human-body model (HBM) is the most commonly used model for characterizing the susceptibility of an electronic device to damage from electrostatic discharge (ESD). The model is a simulation of the discharge which might occur when a human touches an electronic device.
The HBM definition most widely used is the test model defined in the United States military standard, MIL-STD-883, Method 3015.9, Electrostatic Discharge Sensitivity Classification. This method establishes a simplified equivalent electrical circuit and the necessary test procedures required to model an HBM ESD event.
An internationally widely used standard is JEDEC standard JS-001.
HBM is used primarily for manufacturing environments to quantify an integrated circuit to survive the manufacturing process. A similar standard, IEC 61000-4-2, is used for system level testing and quantifies protection levels for a real world ESD event in an uncontrolled environment.
Model
In both JS-001-2012 and MIL-STD-883H the charged human body is modeled by a 100 pF capacitor and a 1500 ohm discharging resistance. During testing, the capacitor is fully charged to several kilovolts (2 kV, 4 kV, 6 kV and 8 kV are typical standard levels) and then discharged through the resistor connected in series to the device under test.
See also
Charged-device model
Transmission-line pulse
|
https://en.wikipedia.org/wiki/Hacker%20II%3A%20The%20Doomsday%20Papers
|
Hacker II: The Doomsday Papers is computer game written by Steve Cartwright and published by Activision in 1986. It is the sequel to the 1985 game Hacker. Hacker II was released for the Amiga, Apple II, Apple IIGS, Amstrad CPC, Atari ST, Commodore 64, IBM PC, Macintosh, and ZX Spectrum.
Plot
Hacker II is more difficult and involved than the first game. In Hacker II, the player is actually recruited based upon their (assumed) success with the activities in the original game. Once again, they are tasked with controlling a robot, this time to infiltrate a secure facility in order to retrieve documents known only as "The Doomsday Papers" from a well guarded vault to ensure the security of the United States.
Eventually, as they escape with the papers, the player is confronted by agents of the United States who reveal that they have actually been working for a former Magma employee, who wanted the papers in revenge for what had happened to the company the player had presumably exposed in the first game. The building that the player had unwittingly broken into was a government facility. The player then has to go back into the facility as part of a gambit to expose the Magma agent, avoiding the same security that had threatened the player before.
Gameplay
Gameplay is considerably changed from the previous game, and the packaging notably includes a "manual" describing the function of a four-way monitor system provided to the player. It is hooked into the camera security network of the facility the player is asked to infiltrate. A handful of robots are available, hidden in the facility, in case some are lost. By using the camera system and in-game map that helps track guard patrols and the location of the robots, the player must explore the one floor facility and find the codes needed to open the vault and escape with the papers. To aid the player there is also a pre-recorded security tape of a typical day for every camera in the facility, which the player can bypass t
|
https://en.wikipedia.org/wiki/Thomas%20Jefferson%20Jackson%20See
|
Thomas Jefferson Jackson (T. J. J.) See (February 19, 1866 – July 4, 1962) was an American astronomer whose promulgated theories in astronomy and physics were eventually disproven. His educational and professional career were dogged by plagiarism and conflict, including his attacks on relativity. He was fired from his position at two observatories, eventually serving out his professional years at a naval shipyard in California.
Early life
See was born near Montgomery City, Missouri. He attended the University of Missouri, graduating in 1889 with an undergraduate career that was outwardly stellar. See achieved honors distinction in nearly every subject, became his class valedictorian and was the recipient of the Laws Astronomical Medal for an original thesis on an astronomical subject. However, his speech "The Spirit of the Age" was a plagiarized version of an earlier speech given by another student, and his "original thesis" for the Laws Astronomical Medal was claimed to be original work but was just from prior work by Sir George Darwin. See was also a critical player in the academic insurgency aimed at ousting university president Samuel Laws (in favor of See's mentor William Benjamin Smith). This plagiarism and bitter in-fighting "set the scene for a career perhaps unrivalled as an example of wasted talent". Nevertheless, with the outwardly strong credentials, See went to the University of Berlin where he received a PhD in mathematics in 1892. With a European doctorate, See returned to America with enviable credentials and a career of great promise.
Scientific work
See specialized in the study of binary stars, particularly in determining their orbits. See initially found work at the University of Chicago, where he worked as an instructor under George Ellery Hale. See left Chicago in 1896 after failing to receive a promotion. He next worked at Lowell Observatory until he was fired in 1898 for his arrogant attitude towards the staff. See's arrogance and overconfid
|
https://en.wikipedia.org/wiki/Aperture%20%28computer%20memory%29
|
In computing, an aperture is a portion of physical address space (i.e. physical memory) that is associated with a particular peripheral device or a memory unit. Apertures may reach external devices such as ROM or RAM chips, or internal memory on the CPU itself.
Typically, a memory device attached to a computer accepts addresses starting at zero, and so a system with more than one such device would have ambiguous addressing. To resolve this, the memory logic will contain several aperture selectors, each containing a range selector and an interface to one of the memory devices.
The set of selector address ranges of the apertures are disjoint. When the CPU presents a physical address within the range recognized by an aperture, the aperture unit routes the request (with the address remapped to a zero base) to the attached device. Thus, apertures form a layer of address translation below the level of the usual virtual-to-physical mapping.
See also
Address bus
AGP aperture
Memory-mapped I/O
External links
Flash Memory Solutions
Computer memory
Computer architecture
|
https://en.wikipedia.org/wiki/GNU%20Bazaar
|
GNU Bazaar (formerly Bazaar-NG, command line tool bzr) is a distributed and client–server revision control system sponsored by Canonical.
Bazaar can be used by a single developer working on multiple branches of local content, or by teams collaborating across a network.
Bazaar is written in the Python programming language, with packages for major Linux distributions, and Microsoft Windows. Bazaar is free software and part of the GNU Project.
Features
Bazaar commands are similar to those found in CVS or Subversion. A new project can be started and maintained without a remote repository server by invoking bzr init in a directory which a person wishes to version.
In contrast to purely distributed version control systems which do not use a central server, Bazaar supports working with or without a central server. It is possible to use both methods at the same time with the same project. The websites Launchpad and SourceForge provide free hosting service for projects managed with Bazaar.
Bazaar has support for working with some other revision control systems. This allows users to branch from another system (such as Subversion), make local changes and commit them into a Bazaar branch, and then later merge them back into the other system. Read-only access is also available for Git and Mercurial. Bazaar also allows for interoperation with many other systems (including CVS, Darcs, Git, Perforce, Mercurial) by allowing one to import/export the history.
Bazaar supports files with names from the complete Unicode set. It also allows commit messages, committer names, etc. to be in Unicode.
History
Baz: an earlier Canonical version control system
The name "Bazaar" was originally used by a fork of the GNU arch client tla. This fork is called Baz to distinguish it from the current Bazaar software. Baz was announced in October 2004 by Canonical employee Robert Collins and maintained until 2005, when the project then called Bazaar-NG (the present Bazaar) was announced a
|
https://en.wikipedia.org/wiki/Supraglenoid%20tubercle
|
The supraglenoid tubercle is a region of the scapula from which the long head of the biceps brachii muscle originates. It is a small, rough projection superior to the glenoid cavity near the base of the coracoid process. The term supraglenoid is from the Latin supra, meaning above, and glenoid, meaning socket or cavity.
Clinical relevance
Biceps tendonitis
Biceps tendonitis originates on the long head of the biceps brachii at the supraglenoid tubercle in 30% of cases. The main symptom is generally anterior biceps instability, but the disease can also be characterized by chronic anterior shoulder pain which radiates towards the lateral part of the elbow. In cases of biceps tendinitis, steroids can be injected fluoroscopically at the supraglenoid tubercle to reduce pain associated with the pathology.
Avulsion
The supraglenoid tubercle ossifies separately from the rest of the scapula, so may not be as strong as the rest of the bone. It can be pulled off (avulsed), often after an excessively strong contraction of the biceps brachii. This may also cause a fracture of the surrounding parts of the scapula, particularly the glenoid cavity. This may also occur in horses. This type of bone fracture is quite rare.
Additional images
See also
Infraglenoid tubercle
|
https://en.wikipedia.org/wiki/Infraglenoid%20tubercle
|
The infraglenoid tubercle is the part of the scapula from which the long head of the triceps brachii muscle originates. The infraglenoid tubercle is a tubercle located on the lateral part of the scapula, inferior to (below) the glenoid cavity. The name infraglenoid tubercle refers to its location below the glenoid cavity.
Function
The infraglenoid tubercle is the origin of the long head of the triceps brachii muscle. It helps to stabilise the muscle origin.
Additional images
See also
Supraglenoid tubercle
|
https://en.wikipedia.org/wiki/Greater%20sciatic%20notch
|
The greater sciatic notch is a notch in the ilium, one of the bones that make up the human pelvis. It lies between the posterior inferior iliac spine (above), and the ischial spine (below). The sacrospinous ligament changes this notch into an opening, the greater sciatic foramen.
The notch holds the piriformis, the superior gluteal vein and artery, and the superior gluteal nerve; the inferior gluteal vein and artery and the inferior gluteal nerve; the sciatic and posterior femoral cutaneous nerves; the internal pudendal artery and veins, and the nerves to the internal obturator and quadratus femoris muscles.
Of these, the superior gluteal vessels and nerve pass out above the piriformis, and the other structures below it.
The greater sciatic notch is wider in females (about 74.4 degrees on average) than in males (about 50.4 degrees).
See also
Greater sciatic foramen
Lesser sciatic notch
|
https://en.wikipedia.org/wiki/Lesser%20sciatic%20notch
|
Below the ischial spine is a small notch, the lesser sciatic notch; it is smooth, coated in the recent state with cartilage, the surface of which presents two or three ridges corresponding to the subdivisions of the tendon of the obturator internus, which winds over it.
It is converted into a foramen, the lesser sciatic foramen, by the sacrotuberous and sacrospinous ligaments, and transmits the tendon of the obturator internus, the nerve which supplies that muscle, and the internal pudendal vessels and pudendal nerve.
See also
Lesser sciatic foramen
Greater sciatic notch
|
https://en.wikipedia.org/wiki/Physical-to-Virtual
|
In computing. Physical-to-Virtual ("P2V" or "p-to-v") involves the process of decoupling and migrating a physical server's operating system (OS), applications, and data from that physical server to a virtual-machine guest hosted on a virtualized platform.
Methods of P2V migration
Manual P2V
User manually creates a virtual machine in a virtual host environment and copies all the files from OS, applications and data from the source machine.
Semi-automated P2V
Performing a P2V migration using a tool that assists the user in moving the servers from physical state to virtual machine.
Microsoft's Virtual Server 2005 Migration Toolkit, HOWTO: Guideline for use Virtual Server Migration ToolKit (KB555306)
VMware provides a semi-automated tool called VMware vCenter Converter for moving physical servers running Windows or Linux into virtual environments while they are powered on. VMware vCenter Converter replaces two older utilities: Importer (bundled with VMware Workstation) and P2V Assistant.
Oracle's Virtual Box has a Linux-based tool which allows the conversion of a dd image of an existing hard drive
Microsoft provides the SysInternals disk2vhd utility for making images from Windows XP or later systems to be used with Windows Virtual PC, Microsoft Virtual Server or Hyper-V.
openQRM, an open-source datacenter management platform, does P2V (and V2P, V2V or P2P).
Storix's bare-metal recovery product (System Backup Administrator) provides P2V and V2P capabilities for Linux, AIX, and Solaris.
Fully automated P2V
Performing a P2V migration using a tool that migrates the server over the network without any assistance from the user.
Veritas Backup Exec has Physical to Virtual (P2V) conversion (and V2P) feature build into the backup engine which can be used for migrations or instant recovery
vContinuum vContinuum by InMage systems is an automated P2V data protection/migration tool
Symantec System Recovery enables fast, automated P2V and V2P conversions
Quest vConverte
|
https://en.wikipedia.org/wiki/Position%20tolerance
|
Position Tolerance (symbol: ⌖) is a geometric dimensioning and tolerancing (GD&T) location control used on engineering drawings to specify desired location, as well as allowed deviation to the position of a feature on a part.
Position tolerance must only be applied to features of size, which requires that the feature have at least two opposable points.
See also
Circle
Miscellaneous Technical
Technical drawing
|
https://en.wikipedia.org/wiki/Fallout%20Protection
|
Fallout Protection: What To Know And Do About Nuclear Attack was an official United States federal government booklet released in December 1961 by the United States Department of Defense and the Office of Civil Defense.
The first page of the book is a note from then-U.S. Secretary of Defense Robert McNamara explaining that the booklet is the result of the first task he was given when he assumed responsibility for the Federal Civil Defense Program in August 1961. The task, assigned by President John F. Kennedy, was to "give the American people the facts they need to know about the dangers of a thermonuclear attack and what they can do to protect themselves."
Subject matter
Much more straightforward about the dangers of atomic weapons than Survival Under Atomic Attack, and published in the era of the hydrogen bomb and the ICBM, the book first explains general information and hazards of nuclear weapons, fallout and radiation. Second, it covers community fallout shelters, improvised fallout shelters and supplies that one ought to have ready in case of a nuclear attack. There is a brief section outlining "emergency housekeeping" which covers water and food conservation and first aid for the time spent in the shelter. Last, the booklet mentions recovery and how to clean up the fallout.
See also
Fallout: An American Nuclear Tragedy
List of books about nuclear issues
Protect and Survive
Survivalism
Nuclear War Survival Skills
Duck and Cover (film)
External links
Fallout Protection Booklet History
Fallout Protection in various formats from the Internet Archive
1961 books
Disaster preparedness in the United States
Publications of the United States government
Cold War history of the United States
Nuclear warfare
Books about nuclear issues
Cold War documents
United States civil defense
Nuclear fallout
|
https://en.wikipedia.org/wiki/George%20Lane%20%28mental%20calculator%29
|
George Lane (born 1964) is a British mental calculator and author. He is a five-time gold medalist in the Mental Calculations event at the annual Mind Sports Olympiad, and is one of only five Grandmasters of Mental Calculation, as recognised by the Mind Sports Organisation. George is a regular facilitator at the annual Junior Mental Calculation World Championship. He has written two books.
|
https://en.wikipedia.org/wiki/Kerckhoff%20Marine%20Laboratory
|
The William G. Kerckhoff Marine Laboratory is owned and operated by the California Institute of Technology (Caltech). It is located 101 Dahlia Street, in the Corona del Mar district of Newport Beach, in Orange County, California.
History
The marine laboratory was established by biologist Thomas Hunt Morgan in 1928 to replicate the facilities at the Stazione Zoologica in Naples, Italy. Caltech made the decision to purchase the facility in 1929. It is one of the oldest marine laboratories on the West Coast of the United States. From 1962 until his death in 2002, Dr. Wheeler J. North conducted numerous studies on the ecology of the California kelp forests while based at this laboratory. During the 1990s and 2000s investigators included members of the Eric Davidson lab working on various marine biology related projects. The current Director of the Kerckhoff Marine Lab is Prof. Victoria J. Orphan, who maintains an active research program there.
See also
Marine biology
Marine ecosystem
Notes
California Institute of Technology buildings and structures
Marine biological stations
Biological research institutes in the United States
Buildings and structures in Orange County, California
Newport Beach, California
|
https://en.wikipedia.org/wiki/Bang%E2%80%93bang%20control
|
In control theory, a bang–bang controller (hysteresis, 2 step or on–off controller), is a feedback controller that switches abruptly between two states. These controllers may be realized in terms of any element that provides hysteresis. They are often used to control a plant that accepts a binary input, for example a furnace that is either completely on or completely off. Most common residential thermostats are bang–bang controllers. The Heaviside step function in its discrete form is an example of a bang–bang control signal. Due to the discontinuous control signal, systems that include bang–bang controllers are variable structure systems, and bang–bang controllers are thus variable structure controllers.
Bang–bang solutions in optimal control
In optimal control problems, it is sometimes the case that a control is restricted to be between a lower and an upper bound. If the optimal control switches from one extreme to the other (i.e., is strictly never in between the bounds), then that control is referred to as a bang–bang solution.
Bang–bang controls frequently arise in minimum-time problems. For example, if it is desired for a car starting at rest to arrive at a certain position ahead of the car in the shortest possible time, the solution is to apply maximum acceleration until the unique switching point, and then apply maximum braking to come to rest exactly at the desired position.
A familiar everyday example is bringing water to a boil in the shortest time, which is achieved by applying full heat, then turning it off when the water reaches a boil. A closed-loop household example is most thermostats, wherein the heating element or air conditioning compressor is either running or not, depending upon whether the measured temperature is above or below the setpoint.
Bang–bang solutions also arise when the Hamiltonian is linear in the control variable; application of Pontryagin's minimum or maximum principle will then lead to pushing the control to its upper or l
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.