You are not logged in.
2522) Automobile
Gist
An automobile is a wheeled, self-propelled motor vehicle designed primarily for transporting passengers on roads, typically featuring four wheels and an engine. Developed in the late 19th century, these vehicles have evolved from early internal combustion models to include modern electric powertrains. There are over 1.6 billion cars in use worldwide, profoundly influencing global transportation and industry.
An automobile is a self-propelled, wheeled motor vehicle primarily designed for transporting people on roads, commonly featuring four wheels and powered by an internal combustion engine or electric motor, though the term also covers vehicles like trucks and motorcycles. Essentially, it's a "self-moving" vehicle, combining the Greek auto (self) and Latin mobilis (movable).
Summary
A car, or an automobile, is a motor vehicle with wheels. Most definitions of cars state that they run primarily on roads, seat 1-8 people, have four wheels, and mainly transport people rather than cargo. There are over 1.6 billion cars in use worldwide as of 2025.
The French inventor Nicolas-Joseph Cugnot built the first steam-powered road vehicle in 1769, while the Swiss inventor François Isaac de Rivaz designed and constructed the first internal combustion-powered automobile in 1808. The modern car—a practical, marketable automobile for everyday use—was invented in 1886, when the German inventor Carl Benz patented his Benz Patent-Motorwagen. Commercial cars became widely available during the 20th century. The 1901 Oldsmobile Curved Dash and the 1908 Ford Model T, both American cars, are widely considered the first mass-produced and mass-affordable cars, respectively. Cars were rapidly adopted in the US, where they replaced horse-drawn carriages. In Europe and other parts of the world, demand for automobiles did not increase until after World War II. In the 21st century, car usage is still increasing rapidly, especially in China, India, and other newly industrialised countries.
Cars have controls for driving, parking, passenger comfort, and a variety of lamps. Over the decades, additional features and controls have been added to vehicles, making them progressively more complex. These include rear-reversing cameras, air conditioning, navigation systems, and in-car entertainment. Most cars in use in the early 2020s are propelled by an internal combustion engine, fueled by the combustion of fossil fuels. Electric cars, which were invented early in the history of the car, became commercially available in the 2000s and widespread in the 2020s. The transition from fossil-fuel-powered cars to electric cars is a central feature of most climate change mitigation scenarios.
There are costs and benefits to car use. The costs to the individual include acquiring the vehicle, interest payments (if the car is financed), repairs and maintenance, fuel, depreciation, driving time, parking fees, taxes, and insurance. The costs to society include resources used to produce cars and fuel, maintaining roads, land-use, road congestion, air pollution, noise pollution, public health, and disposing of the vehicle at the end of its life. Traffic collisions are the largest cause of injury-related deaths worldwide. Personal benefits include on-demand transportation, mobility, independence, and convenience. Societal benefits include economic benefits, such as job and wealth creation from the automotive industry, transportation provision, and societal wellbeing from leisure and travel opportunities. People's ability to move flexibly from place to place has far-reaching implications for society.
Details
An automobile is a usually four-wheeled vehicle designed primarily for passenger transportation and commonly propelled by an internal-combustion engine using a volatile fuel.
Automotive design
The modern automobile is a complex technical system employing subsystems with specific design functions. Some of these consist of thousands of component parts that have evolved from breakthroughs in existing technology or from new technologies such as electronic computers, high-strength plastics, and new alloys of steel and nonferrous metals. Some subsystems have come about as a result of factors such as air pollution, safety legislation, and competition between manufacturers throughout the world.
Passenger cars have emerged as the primary means of family transportation, with an estimated 1.4 billion in operation worldwide. About one-quarter of these are in the United States, where more than three trillion miles (almost five trillion kilometres) are traveled each year. In recent years, Americans have been offered hundreds of different models, about half of them from foreign manufacturers. To capitalize on their proprietary technological advances, manufacturers introduce new designs ever more frequently. With some 70 million new units built each year worldwide, manufacturers have been able to split the market into many very small segments that nonetheless remain profitable.
New technical developments are recognized to be the key to successful competition. Research and development engineers and scientists have been employed by all automobile manufacturers and suppliers to improve the body, chassis, engine, drivetrain, control systems, safety systems, and emission-control systems.
These outstanding technical advancements are not made without economic consequences. According to a study by Ward’s Communications Incorporated, the average cost for a new American car increased $4,700 (in terms of the value of the dollar in 2000) between 1980 and 2001 because of mandated safety and emission-control performance requirements (such as the addition of air bags and catalytic converters). New requirements continued to be implemented in subsequent years. The addition of computer technology was another factor driving up car prices, which increased by 29 percent between 2009 and 2019. This is in addition to the consumer costs associated with engineering improvements in fuel economy, which may be offset by reduced fuel purchases.
Vehicle design depends to a large extent on its intended use. Automobiles for off-road use must be durable, simple systems with high resistance to severe overloads and extremes in operating conditions. Conversely, products that are intended for high-speed, limited-access road systems require more passenger comfort options, increased engine performance, and optimized high-speed handling and vehicle stability. Stability depends principally on the distribution of weight between the front and rear wheels, the height of the centre of gravity and its position relative to the aerodynamic centre of pressure of the vehicle, suspension characteristics, and the selection of which wheels are used for propulsion. Weight distribution depends principally on the location and size of the engine. The common practice of front-mounted engines exploits the stability that is more readily achieved with this layout. The development of aluminum engines and new manufacturing processes has, however, made it possible to locate the engine at the rear without necessarily compromising stability.
Body
Automotive body designs are frequently categorized according to the number of doors, the arrangement of seats, and the roof structure. Automobile roofs are conventionally supported by pillars on each side of the body. Convertible models with retractable fabric tops rely on the pillar at the side of the windshield for upper body strength, as convertible mechanisms and glass areas are essentially nonstructural. Glass areas have been increased for improved visibility and for aesthetic reasons.
The high cost of new factory tools makes it impractical for manufacturers to produce totally new designs every year. Completely new designs usually have been programmed on three- to six-year cycles with generally minor refinements appearing during the cycle. In the past, as many as four years of planning and new tool purchasing were needed for a completely new design. Computer-aided design (CAD), testing by use of computer simulations, and computer-aided manufacturing (CAM) techniques may now be used to reduce this time requirement by 50 percent or more. See machine tool: Computer-aided design and computer-aided manufacturing (CAD/CAM).
Automotive bodies are generally formed out of sheet steel. The steel is alloyed with various elements to improve its ability to be formed into deeper depressions without wrinkling or tearing in manufacturing presses. Steel is used because of its general availability, low cost, and good workability. For certain applications, however, other materials, such as aluminum, fibreglass, and carbon-fibre reinforced plastic, are used because of their special properties. Polyamide, polyester, polystyrene, polypropylene, and ethylene plastics have been formulated for greater toughness, dent resistance, and resistance to brittle deformation. These materials are used for body panels. Tooling for plastic components generally costs less and requires less time to develop than that for steel components and therefore may be changed by designers at a lower cost.
To protect bodies from corrosive elements and to maintain their strength and appearance, special priming and painting processes are used. Bodies are first dipped in cleaning baths to remove oil and other foreign matter. They then go through a succession of dip and spray cycles. Enamel and acrylic lacquer are both in common use. Electrodeposition of the sprayed paint, a process in which the paint spray is given an electrostatic charge and then attracted to the surface by a high voltage, helps assure that an even coat is applied and that hard-to-reach areas are covered. Ovens with conveyor lines are used to speed the drying process in the factory. Galvanized steel with a protective zinc coating and corrosion-resistant stainless steel are used in body areas that are more likely to corrode.
Chassis
In most passenger cars through the middle of the 20th century, a pressed-steel frame—the vehicle’s chassis—formed a skeleton on which the engine, wheels, axle assemblies, transmission, steering mechanism, brakes, and suspension members were mounted. The body was flexibly bolted to the chassis during a manufacturing process typically referred to as body-on-frame construction. This process is used today for heavy-duty vehicles, such as trucks, which benefit from having a strong central frame, subjected to the forces involved in such activities as carrying freight, including the absorption of the movements of the engine and axle that is allowed by the combination of body and frame.
In modern passenger-car designs, the chassis frame and the body are combined into a single structural element. In this arrangement, called unit-body (or unibody) construction, the steel body shell is reinforced with braces that make it rigid enough to resist the forces that are applied to it. Separate frames or partial “stub” frames have been used for some cars to achieve better noise-isolation characteristics. The heavier-gauge steel present in modern component designs also tends to absorb energy during impacts and limit intrusion in accidents.
Engine
A wide range of engines has been used experimentally and in automotive production. The most successful for automobiles has been the gasoline-fueled reciprocating-piston internal-combustion engine, operating on a four-stroke cycle, while diesel engines are widely used for trucks and buses. The gasoline engine was originally selected for automobiles because it could operate more flexibly over a wide range of speeds, and the power developed for a given weight engine was reasonable; it could be produced by economical mass-production methods; and it used a readily available, moderately priced fuel. Reliability, compact size, exhaust emissions, and range of operation later became important factors.
There has been an ongoing reassessment of these priorities with new emphasis on the reduction of greenhouse gases (see greenhouse effect) or pollution-producing characteristics of automotive power systems. This has created new interest in alternate power sources and internal-combustion engine refinements that previously were not close to being economically feasible. Several limited-production battery-powered electric vehicles are marketed today. In the past they had not proved to be competitive, because of costs and operating characteristics. The gasoline engine, with new emission-control devices to improve emission performance, has been challenged in recent years by hybrid power systems that combine gasoline or diesel engines with battery systems and electric motors. Such designs are, however, more complex and therefore more costly.
The evolution of higher-performance engines in the United States led the industry away from long, straight engine cylinder layouts to compact six- and eight-cylinder V-type layouts for larger cars (with horsepower ratings up to about 350). Smaller cars depend on smaller four-cylinder engines. European automobile engines were of a much wider variety, ranging from 1 to 12 cylinders, with corresponding differences in overall size, weight, piston displacement, and cylinder bores. A majority of the models had four cylinders and horsepower ratings up to 120. Most engines had straight or in-line cylinders. There were, however, several V-type models and horizontally opposed two- and four-cylinder makes. Overhead camshafts were frequently employed. The smaller engines were commonly air-cooled and located at the rear of the vehicle; compression ratios were relatively low. Increased interest in improved fuel economy brought a return to smaller V-6 and four-cylinder layouts, with as many as five valves per cylinder to improve efficiency. Variable valve timing to improve performance and lower emissions has been achieved by manufacturers in all parts of the world. Electronic controls automatically select the better of two profiles on the same cam for higher efficiency when engine speeds and loads change.
Fuel
Specially formulated gasoline is essentially the only fuel used for automobile operation, although diesel fuels are used for many trucks and buses and a few automobiles, and compressed liquefied hydrogen has been used experimentally. The most important requirements of a fuel for automobile use are proper volatility, sufficient antiknock quality, and freedom from polluting by-products of combustion. The volatility is reformulated seasonally by refiners so that sufficient gasoline vaporizes, even in extremely cold weather, to permit easy engine starting. Antiknock quality is rated by the octane number of the gasoline. The octane number requirement of an automobile engine depends primarily on the compression ratio of the engine but is also affected by combustion-chamber design, the maintenance condition of engine systems, and chamber-wall deposits. In the 21st century regular gasoline carried an octane rating of 87 and high-test in the neighbourhood of 93.
Automobile manufacturers have lobbied for regulations that require the refinement of cleaner-burning gasolines, which permit emission-control devices to work at higher efficiencies. Such gasoline was first available at some service stations in California, and from 2017 the primary importers and refiners of gasoline throughout the United States were required to remove sulfur particles from fuel to an average level of 10 parts per million (ppm).
Vehicle fleets fueled by natural gas have been in operation for several years. Carbon monoxide and particulate emissions are reduced by 65 to 90 percent. Natural-gas fuel tanks must be four times larger than gasoline tanks for equivalent vehicles to have the same driving range. This compromises cargo capacity.
Ethanol (ethyl alcohol) is often blended with gasoline (15 parts to 85 parts) to raise its octane rating, which results in a smoother-running engine. Ethanol, however, has a lower energy density than gasoline, which results in decreased range per tankful.
Lubrication
All moving parts of an automobile require lubrication. Without it, friction would increase power consumption and damage the parts. The lubricant also serves as a coolant, a noise-reducing cushion, and a sealant between engine piston rings and cylinder walls. The engine lubrication system incorporates a gear-type pump that delivers filtered oil under pressure to a system of drilled passages leading to various bearings. Oil spray also lubricates the cams and valve lifters.
Wheel bearings and universal joints require a fairly stiff grease; other chassis joints require a soft grease that can be injected by pressure guns. Hydraulic transmissions require a special grade of light hydraulic fluid, and manually shifted transmissions use a heavier gear oil similar to that for rear axles to resist heavy loads on the gear teeth. Gears and bearings in lightly loaded components, such as generators and window regulators, are fabricated from self-lubricating plastic materials. Hydraulic fluid is also used in other vehicle systems in conjunction with small electric pumps and motors.
Cooling system
Almost all automobiles employ liquid cooling systems for their engines. A typical automotive cooling system comprises (1) a series of channels cast into the engine block and cylinder head, surrounding the combustion chambers with circulating water or other coolant to carry away excessive heat, (2) a radiator, consisting of many small tubes equipped with a honeycomb of fins to radiate heat rapidly, which receives and cools hot liquid from the engine, (3) a centrifugal-type water pump with which to circulate coolant, (4) a thermostat, which maintains constant temperature by automatically varying the amount of coolant passing into the radiator, and (5) a fan, which draws fresh air through the radiator.
For operation at temperatures below 0 °C (32 °F), it is necessary to prevent the coolant from freezing. This is usually done by adding some compound, such as ethylene glycol, to depress the freezing point of the coolant. By varying the amount of additive, it is possible to protect against freezing of the coolant down to any minimum temperature normally encountered. Coolants contain corrosion inhibitors designed to make it necessary to drain and refill the cooling system only every few years.
Air-cooled cylinders operate at higher, more efficient temperatures, and air cooling offers the important advantage of eliminating not only freezing and boiling of the coolant at temperature extremes but also corrosion damage to the cooling system. Control of engine temperature is more difficult, however, and high-temperature-resistant ceramic parts are required when design operating temperatures are significantly increased.
Pressurized cooling systems have been used to increase effective operating temperatures. Partially sealed systems using coolant reservoirs for coolant expansion if the engine overheats were introduced in the early 1970s. Specially formulated coolants that do not deteriorate over time eliminate the need for annual replacement.
Electrical system
The electrical system comprises a storage battery, generator, starting (cranking) motor, lighting system, ignition system, and various accessories and controls. Originally, the electrical system of the automobile was limited to the ignition equipment. With the advent of the electric starter on a 1912 Cadillac model, electric lights and horns began to replace the kerosene and acetylene lights and the bulb horns. Electrification was rapid and complete, and, by 1930, 6-volt systems were standard everywhere.
Increased engine speeds and higher cylinder pressures made it increasingly difficult to meet high ignition voltage requirements. The larger engines required higher cranking torque. Additional electrically operated features—such as radios, window regulators, and multispeed windshield wipers—also added to system requirements. To meet these needs, 12-volt systems replaced the 6-volt systems in the late 1950s around the world.
The ignition system provides the spark to ignite the air-fuel mixture in the cylinders of the engine. The system consists of the spark plugs, coil, distributor, and battery. In order to jump the gap between the electrodes of the spark plugs, the 12-volt potential of the electrical system must be stepped up to about 20,000 volts. This is done by a circuit that starts with the battery, one side of which is grounded on the chassis and leads through the ignition switch to the primary winding of the ignition coil and back to the ground through an interrupter switch. Interrupting the primary circuit induces a high voltage across the secondary terminal of the coil. The high-voltage secondary terminal of the coil leads to a distributor that acts as a rotary switch, alternately connecting the coil to each of the wires leading to the spark plugs.
Solid-state or transistorized ignition systems were introduced in the 1970s. These distributor systems provided increased durability by eliminating the frictional contacts between breaker points and distributor cams. The breaker point was replaced by a revolving magnetic-pulse generator in which alternating-current pulses trigger the high voltage needed for ignition by means of an amplifier electronic circuit. Changes in engine ignition timing are made by vacuum or electronic control unit (microprocessor) connections to the distributor.
The source of energy for the various electrical devices of the automobile is a generator, or alternator, that is belt-driven from the engine crankshaft. The design is usually an alternating-current type with built-in rectifiers and a voltage regulator to match the generator output to the electric load and also to the charging requirements of the battery, regardless of engine speed.
A lead-acid battery serves as a reservoir to store excess output of the generator. This provides energy for the starting motor and power for operating other electric devices when the engine is not running or when the generator speed is not sufficiently high for the load.
The starting motor drives a small spur gear so arranged that it automatically moves in to mesh with gear teeth on the rim of the flywheel as the starting-motor armature begins to turn. When the engine starts, the gear is disengaged, thus preventing damage to the starting motor from overspeeding. The starting motor is designed for high current consumption and delivers considerable power for its size for a limited time.
Headlights must satisfactorily illuminate the highway ahead of the automobile for driving at night or in inclement weather without temporarily blinding approaching drivers. This was achieved in modern cars with double-filament bulbs with a high and a low beam, called sealed-beam units. Introduced in 1940, these bulbs found widespread use following World War II. Such units could have only one filament at the focal point of the reflector. Because of the greater illumination required for high-speed driving with the high beam, the lower beam filament was placed off centre, with a resulting decrease in lighting effectiveness. Separate lamps for these functions can also be used to improve illumination effectiveness.
Dimming is automatically achieved on some cars by means of a photocell-controlled switch in the lamp circuit that is triggered by the lights of an oncoming car. Lamp clusters behind aerodynamic plastic covers permitted significant front-end drag reduction and improved fuel economy. In this arrangement, steerable headlights became possible with an electric motor to swivel the lamp assembly in response to steering wheel position. The regulations of various governments dictate brightness and field of view requirements for vehicle lights.
Signal lamps and other special-purpose lights have increased in usage since the 1960s. Amber-coloured front and red rear signal lights are flashed as a turn indication; all these lights are flashed simultaneously in the “flasher” (hazard) system for use when a car is parked along a roadway or is traveling at a low speed on a high-speed highway. Marker lights that are visible from the front, side, and rear also are widely required by law. Red-coloured rear signals are used to denote braking, and cornering lamps, in connection with turning, provide extra illumination in the direction of an intended turn. Backup lights provide illumination to the rear and warn anyone behind the vehicle when the driver is backing up. High-voltage light-emitting diodes (LEDs) have been developed for various signal and lighting applications.
Transmission
The gasoline engine must be disconnected from the driving wheels when it is started and when idling. This characteristic necessitates some type of unloading and engaging device to permit gradual application of load to the engine after it has been started. The torque, or turning effort, that the engine is capable of producing is low at low crankshaft speeds, increasing to a maximum at some fairly high speed representing the maximum, or rated, horsepower.
The efficiency of an automobile engine is highest when the load on the engine is high and the throttle is nearly wide open. At moderate speeds on level pavement, the power required to propel an automobile is only a fraction of this. Under normal driving conditions at constant moderate speed, the engine may operate at an uneconomically light load unless some means is provided to change its speed and power output.
The transmission is such a speed-changing device. Installed in the power train between the engine and the driving wheels, it permits the engine to operate at a higher speed when its full power is needed and to slow down to a more economical speed when less power is needed. Under some conditions, as in starting a stationary vehicle or in ascending steep grades, the torque of the engine is insufficient, and amplification is needed. Most devices employed to change the ratio of the speed of the engine to the speed of the driving wheels multiply the engine torque by the same factor by which the engine speed is increased.
The simplest automobile transmission is the sliding-spur gear type with three or more forward speeds and reverse. The desired gear ratio is selected by manipulating a shift lever that slides a spur gear into the proper position to engage the various gears. A clutch is required to engage and disengage gears during the selection process. The necessity of learning to operate a clutch is eliminated by an automatic transmission. Most automatic transmissions employ a hydraulic torque converter, a device for transmitting and amplifying the torque produced by the engine. Each type provides for manual selection of reverse and low ranges that either prevent automatic upshifts or employ lower gear ratios than are used in normal driving. Grade-retard provisions are also sometimes included to supply dynamic engine braking on hills. Automatic transmissions not only require little skill to operate but also make possible better performance than is obtainable with designs that require clutch actuation.
In hydraulic transmissions, shifting is done by a speed-sensitive governing device that changes the position of valves that control the flow of hydraulic fluid. The vehicle speeds at which shifts occur depend on the position of the accelerator pedal, and the driver can delay upshifts until higher speed is attained by depressing the accelerator pedal further. Control is by hydraulically engaged bands and multiple-disk clutches running in oil, either by the driver’s operation of the selector lever or by speed- and load-sensitive electronic control in the most recent designs. Compound planetary gear trains with multiple sun gears and planet pinions have been designed to provide a low forward speed, intermediate speeds, a reverse, and a means of locking into direct drive. This unit is used with various modifications in almost all hydraulic torque-converter transmissions. All transmission control units are interconnected with vehicle emission control systems that adjust engine timing and air-to-fuel ratios to reduce exhaust emissions.
Oil in the housing is accelerated outward by rotating vanes in the pump impeller and, reacting against vanes in the turbine impeller, forces them to rotate, as shown schematically in the figure. The oil then passes into the stator vanes, which redirect it to the pump. The stator serves as a reaction member providing more torque to turn the turbine than was originally applied to the pump impeller by the engine. Thus, it acts to multiply engine torque by a factor of up to 2 1/2 to 1.
Blades in all three elements are specially contoured for their specific function and to achieve particular multiplication characteristics. Through a clutch linkage, the stator is allowed gradually to accelerate until it reaches the speed of the pump impeller. During this period torque multiplication gradually drops to approach 1 to 1.
The hydraulic elements are combined with two or more planetary gear sets, which provide further torque multiplication between the turbine and the output shaft.
Continuously (or infinitely) variable transmissions provide a very efficient means of transferring engine power and, at the same time, automatically changing the effective input-to-output ratio to optimize economy by keeping the engine running within its best power range. Most designs employ two variable-diameter pulleys connected by either a steel or high-strength rubber V-belt. The pulleys are split so that effective diameters may be changed by an electrohydraulic actuator to change the transmission ratio. This permits the electronic control unit to select the optimum ratio possible for maximum fuel economy and minimum emissions at all engine speeds and loads. Originally these units were limited to small cars, but belt improvements have made them suitable for larger cars.
Additional Information
The four-wheeled transportation vehicle symbolizes the promise and the pitfalls of the modern age.
An automobile is a self-propelled motor vehicle intended for passenger transportation on land. It usually has four wheels and an internal combustion engine fueled most often by gasoline, a liquid petroleum product. Known more commonly as a car, formerly as a motorcar, it is one of the most universal of modern technologies, manufactured by one of the world’s largest industries. More than 73 million new automobiles were produced worldwide in the year 2017.
The scientific and technical building blocks of the automobile go back several hundred years. For example, in the late 1600s, Dutch scientist Christiaan Huygens invented a type of internal combustion engine sparked by gunpowder. The “horseless carriage” in its modern form had been developed by the end of the 19th century. At that time, it was not clear which of three fuel sources would become most commercially successful: steam, electric power, or gasoline. Cars run by steam engines could go at high speeds but had a short range and were inconvenient to start. Battery-powered electric cars had a 38 percent share of the United States automobile market in 1900, but they also had a limited range and recharging stations were hard to find.
The gasoline-powered automobile won the competition. By 1920, it had overtaken the streets and byways of Europe and the United States. The manufacturing methods introduced by U.S. carmaker Henry Ford revolutionized industrial manufacturing. Ford was the first to install assembly lines in his factory to speed up production. Such techniques reduced the price of Ford’s Model T until it became affordable for most middle-class families. As the 20th century progressed, modern life came to seem increasingly inconceivable, or at least highly inconvenient, without access to a car. Nowadays, the U.S. population drives more than 4.8 trillion kilometers (three trillion miles) every year on average.
But this fundamental component of industrial and consumer society has played a major role in destabilizing Earth’s atmosphere, on which all living things depend. The average automobile emits between four and nine tons (3,629 to 8,165 kilograms; 8,000 to 18,000 pounds) of carbon dioxide and other greenhouse gases per year. Every gallon of gasoline burned to operate a car emits just under 9.1 kilograms (20 pounds) of carbon dioxide. The transportation sector as a whole, including cars, trucks, trains, and aircraft, became the largest source of U.S. greenhouse gas emissions in 2017. Air pollution from automobile exhaust is also a major problem, as are car accidents, which killed more than 100 people per day in the United States in 2016, according to the National Highway Traffic Safety Administration.
:max_bytes(150000):strip_icc():format(webp)/GettyImages-1186091457-0c53b6749ceb49b79053f61f583eef03.jpg)
Q: How do you catch a squirrel interested in ornithology?
A: Climb a tree and act like a nuthatch (Sitta carolinensis).
* * *
Q: How do you catch a Polynesian squirrel?
A: Climb a tree and act like a coconut.
* * *
Q: How can you catch a little squirrel?
A: Climb a tree and pretend to be an almond (botanically speaking, almonds are fruits).
* * *
Q: How do you catch a squirrel with a Katy Perry fixation?
A: Climb a tree and act like a chestnut.
* * *
Q: How do you catch a mechanically inclined squirrel?
A: Climb a tree and act like a 9/16 12N nut.
* * *
2459) Edward Mills Purcell
Gist:
Life
Edward Mills Purcell was born in Taylorville, Illinois. His father worked for a telephone company. Purcell studied electrical engineering at Purdue University in Indiana and physics at Harvard University. During World War II Purcell worked on the development of radar at MIT, but he returned afterwards to Harvard, where he did his Nobel Prize-awarded work and continued to work for the rest of his career. Purcell was married and had two sons.
Work
Protons and neutrons in nuclei act like small, rotating magnets. Atoms and molecules therefore align in a magnetic field. Radio waves can disturb their direction of rotation, but only in certain stages, in accordance with quantum mechanics. When the atoms return to their original positions, they emit electromagnetic radio waves with frequencies characteristic of different elements and isotopes. In 1946, Edward Purcell and Felix Bloch developed methods for precise measurement, making it possible to study different materials’ compositions.
Summary
E.M. Purcell (born Aug. 30, 1912, Taylorville, Ill., U.S.—died March 7, 1997, Cambridge, Mass.) was an American physicist who shared, with Felix Bloch of the United States, the Nobel Prize for Physics in 1952 for his independent discovery (1946) of nuclear magnetic resonance in liquids and in solids. Nuclear magnetic resonance (NMR) has become widely used to study the molecular structure of pure materials and the composition of mixtures.
During World War II Purcell headed a group studying radar problems at the Radiation Laboratory of the Massachusetts Institute of Technology, Cambridge. In 1946 he developed his NMR detection method, which was extremely accurate and a major improvement over the atomic-beam method devised by the American physicist Isidor I. Rabi.
Purcell became professor of physics at Harvard University in 1949 and in 1952 detected the 21-centimetre-wavelength radiation emitted by neutral atomic hydrogen in interstellar space. Such radio waves had been predicted by the Dutch astronomer H.C. van de Hulst in 1944, and their study enabled astronomers to determine the distribution and location of hydrogen clouds in galaxies and to measure the rotation of the Milky Way. In 1960 Purcell became Gerhard Gade professor at Harvard, and in 1979 he received the National Medal of Science. In 1980 he became professor emeritus.
Details
Edward Mills Purcell (August 30, 1912 – March 7, 1997) was an American physicist who shared the 1952 Nobel Prize for Physics for his independent discovery (published 1946) of nuclear magnetic resonance in liquids and in solids. Nuclear magnetic resonance (NMR) has become widely used to study the molecular structure of pure materials and the composition of mixtures. Friends and colleagues knew him as Ed Purcell.
Biography
Born and raised in Taylorville, Illinois, Purcell received his BSEE in electrical engineering from Purdue University, followed by his M.A. and Ph.D. in physics from Harvard University. He was a member of the Alpha Xi chapter of the Phi Kappa Sigma fraternity while at Purdue. After spending the years of World War II working at the MIT Radiation Laboratory on the development of microwave radar, Purcell returned to Harvard to do research. In December 1945, he discovered nuclear magnetic resonance (NMR) with his colleagues Robert Pound and Henry Torrey. NMR provides scientists with an elegant and precise way of determining chemical structure and properties of materials, and is widely used in physics and chemistry. It also is the basis of magnetic resonance imaging (MRI), one of the most important medical advances of the 20th century. For his discovery of NMR, Purcell shared the 1952 Nobel Prize in physics with Felix Bloch of Stanford University.
Purcell also made contributions to astronomy as the first to detect radio emissions from neutral galactic hydrogen (the famous 21 cm line due to hyperfine splitting), affording the first views of the spiral arms of the Milky Way. This observation helped launch the field of radio astronomy, and measurements of the 21 cm line are still an important technique in modern astronomy. He has also made seminal contributions to solid state physics, with studies of spin-echo relaxation, nuclear magnetic relaxation, and negative spin temperature (important in the development of the laser). With Norman F. Ramsey, he was the first to question the CP symmetry of particle physics.
Purcell was the recipient of many awards for his scientific, educational, and civic work. He served as science advisor to Presidents Dwight D. Eisenhower, John F. Kennedy, and Lyndon B. Johnson. He was president of the American Physical Society, and a member of the American Philosophical Society, the National Academy of Sciences, and the American Academy of Arts and Sciences. He was awarded the National Medal of Science in 1979, and the Jansky Lectureship before the National Radio Astronomy Observatory. Purcell was also inducted into his Fraternity's (Phi Kappa Sigma) Hall of Fame as the first Phi Kap ever to receive a Nobel Prize.
Purcell was the author of the innovative introductory text Electricity and Magnetism. The book, a Sputnik-era project funded by an NSF grant, was influential for its use of relativity in the presentation of the subject at this level. The 1965 edition, now freely available due to a condition of the federal grant, was originally published as a volume of the Berkeley Physics Course. The book is also in print as a commercial third edition, as Purcell and Morin. Purcell is also remembered by biologists for his famous lecture "Life at Low Reynolds Number", in which he explained forces and effects dominating in limiting flow regimes (often at the micro scale). He also emphasized the time-reversibility of low Reynolds number flows with a principle referred to as the Scallop theorem.
Purcell died on March 7, 1997, in Cambridge, Massachusetts, aged 84.

Comedy Quotes - III
1. The male is always the pawn in a romantic comedy. Come together, break up, go chase her, get her, roll credits. That's what happens in all of them. - Matthew McConaughey
2. When I tried to branch out into comedy, I didn't do very well at it, so I went back to doing what I do naturally well, or what the audience expects from me - action pictures. - Sylvester Stallone
3. As for doing more dramatic work over comedy, I do whatever turns me on at the moment. - Sandra Bullock
4. You know, if you look all my stuff... If you go back to 'Saturday Night Live,' my stuff always has music, even a bunch of my comedy stuff - like in 'Shrek,' the donkey is always singing. Music is always there. - Eddie Murphy
5. All I need to make a comedy is a park, a policeman and a pretty girl. - Charlie Chaplin.
Silver Nitrate
Gist
Silver nitrate is a colorless, odorless inorganic compound commonly used as a versatile chemical precursor in photography, electroplating, and for manufacturing mirrors. It is widely used in medicine as a topical antiseptic and cauterizing agent for wound care and warts. It is prepared by dissolving silver in nitric acid and must be handled with care due to its corrosive nature.
Silver nitrate is used medically as an antiseptic, antibacterial, and cauterizing agent for wound care, wart removal, and controlling bleeding by releasing silver ions that kill microbes and form protective scabs, while industrially it serves as a precursor for silver compounds, in photographic emulsions, and for analytical chemistry (like testing for halides). It is applied topically to treat wounds, burns, and skin tags, though its use in newborns for eye infections is largely replaced by other treatments in the U.S.
Summary
Details
Silver nitrate is an inorganic compound with chemical formula AgNO3. It is a versatile precursor to many other silver compounds, such as those used in photography. It is far less sensitive to light than the halides.[citation needed] It was once called lunar caustic because silver was called luna by ancient alchemists who associated silver with the moon. In solid silver nitrate, the silver ions are three-coordinated in a trigonal planar arrangement.
Uses:
Precursor to other silver compounds
Silver nitrate is the least expensive salt of silver; it offers several other advantages as well. It is non-hygroscopic, in contrast to silver fluoroborate and silver perchlorate. In addition, it is relatively stable to light, and it dissolves in numerous solvents, including water. The nitrate can be easily replaced by other ligands, rendering AgNO3 versatile. Treatment with solutions of halide ions gives a precipitate of AgX (X = Cl, Br, I). When making photographic film, silver nitrate is treated with halide salts of sodium or potassium to form insoluble silver halide in situ in photographic gelatin, which is then applied to strips of tri-acetate or polyester. Similarly, silver nitrate is used to prepare some silver-based explosives, such as the fulminate, azide, or acetylide, through a precipitation reaction.
Halide abstraction
The silver cation, Ag+, reacts quickly with halide sources to produce the insoluble silver halide, which is a cream precipitate if Br− is used, a white precipitate if Cl− is used and a yellow precipitate if I− is used. This reaction is commonly used in inorganic chemistry to abstract halides.
Other silver salts with non-coordinating anions, namely silver tetrafluoroborate and silver hexafluorophosphate are used for more demanding applications.
Similarly, this reaction is used in analytical chemistry to confirm the presence of chloride, bromide, or iodide ions. Samples are typically acidified with dilute nitric acid to remove interfering ions, e.g. carbonate ions and sulfide ions. This step avoids confusion of silver sulfide or silver carbonate precipitates with that of silver halides. The color of precipitate varies with the halide: white (silver chloride), pale yellow/cream (silver bromide), yellow (silver iodide). AgBr and especially AgI photo-decompose to the metal, as evidenced by a grayish color on exposed samples.
The same reaction was used on steamships in order to determine whether or not boiler feedwater had been contaminated with seawater. It is still used to determine if moisture on formerly dry cargo is a result of condensation from humid air, or from seawater leaking through the hull.
Organic synthesis
Silver nitrate is used in many ways in organic synthesis, e.g. for deprotection and oxidations. Ag+ binds alkenes reversibly, and silver nitrate has been used to separate mixtures of alkenes by selective absorption. The resulting adduct can be decomposed with ammonia to release the free alkene. Silver nitrate is highly soluble in water but is poorly soluble in most organic solvents, except acetonitrile (111.8 g/100 g, 25 °C).
Biology
In histology, silver nitrate is used for silver staining, for demonstrating reticular fibers, proteins and nucleic acids. For this reason it is also used to demonstrate proteins in polyacrylamide gel electrophoresis (PAGE) gels. It can be used as a stain in scanning electron microscopy.
Cut flower stems can be placed in a silver nitrate solution, which prevents the production of ethylene. This delays ageing of the flower.
Indelible ink
Silver nitrate produces long-lasting stain when applied to skin and is one of indelible ink’s ingredients. An electoral stain makes use of this to mark a finger of people who have voted in an election, allowing easy identification to prevent double-voting.
In addition to staining skin, silver nitrate has a history of use in stained glass. In the 14th century, artists began using a "silver stain" (also known as a yellow stain) made from silver nitrate to create a yellow effect on clear glass. The stain would produce a stable color that could range from pale lemon to deep orange or gold. Silver stain was often used with glass paint, and was applied to the opposite side of the glass as the paint. It was also used to create a mosaic effect by reducing the number of pieces of glass in a window. Despite the age of the technique, this process of creating stained glass remains almost entirely unchanged.
Medicine:
Silver salts have antiseptic properties. In 1881 Credé introduced a method known as Credé's prophylaxis, which used of dilute (2%) solutions of silver nitrate in newborn babies' eyes at birth to prevent contraction of gonorrhea from the mother, which could cause blindness via ophthalmia neonatorum. (Modern antibiotics are now used instead).
Fused silver nitrate, shaped into sticks, was traditionally called "lunar caustic". It is used as a cauterizing agent, for example to remove granulation tissue around a stoma. General Sir James Abbott noted in his journals that in India in 1827 it was infused by a British surgeon into wounds in his arm resulting from the bite of a mad dog to cauterize the wounds and prevent the onset of rabies.
Silver nitrate is used to cauterize superficial blood vessels in the nose to help prevent nosebleeds.
Dentists sometimes use silver nitrate-infused swabs to heal oral ulcers. Silver nitrate is used by some podiatrists to kill cells located in the nail bed.
The Canadian physician C. A. Douglas Ringrose researched the use of silver nitrate for sterilization procedures, believing that silver nitrate could be used to block and corrode the fallopian tubes. The technique was ineffective.
Disinfection
Much research has been done in evaluating the ability of the silver ion at inactivating Escherichia coli, a microorganism commonly used as an indicator for fecal contamination and as a surrogate for pathogens in drinking water treatment. Concentrations of silver nitrate evaluated in inactivation experiments range from 10–200 micrograms per liter as Ag+. Silver's antimicrobial activity saw many applications prior to the discovery of modern antibiotics, when it fell into near disuse. Its association with argyria made consumers wary and led them to turn away from it when given an alternative.
Against warts
Repeated daily application of silver nitrate can induce adequate destruction of cutaneous warts, but occasionally pigmented scars may develop. In a placebo-controlled study of 70 patients, silver nitrate given over nine days resulted in clearance of all warts in 43% and improvement in warts in 26% one month after treatment compared to 11% and 14%, respectively, in the placebo group.
Safety
As an oxidant, silver nitrate should be properly stored away from organic compounds. It reacts explosively with ethanol. Despite its common usage in extremely low concentrations to prevent gonorrhea and control nosebleeds, silver nitrate is still very toxic and corrosive. Brief exposure will not produce any immediate side effects other than the purple, brown or black stains on the skin, but upon constant exposure to high concentrations, side effects will be noticeable, which include burns. Long-term exposure may cause eye damage. Silver nitrate is known to be a skin and eye irritant. Silver nitrate has not been thoroughly investigated for potential carcinogenic effect.
Silver nitrate is currently unregulated in water sources by the United States Environmental Protection Agency. However, if more than 1 gram of silver is accumulated in the body, a condition called argyria may develop. Argyria is a permanent cosmetic condition in which the skin and internal organs turn a blue-gray color. The United States Environmental Protection Agency used to have a maximum contaminant limit for silver in water until 1990, when it was determined that argyria did not impact the function of any affected organs despite the discolouration. Argyria is more often associated with the consumption of colloidal silver solutions rather than with silver nitrate, since it is only used at extremely low concentrations to disinfect the water. However, it is still important to be wary before ingesting any sort of silver-ion solution.
Additional Information
Silver nitrate is a chemical compound with the formula AgNO3. It consists of an ionic bond between the silver cation (Ag+) and the nitrate anion (NO3–). Due to the ionic nature of this compound, it readily dissolves in water and dissociates into its constituent ions.
Silver nitrate is a precursor to many compounds of silver, including the silver compounds used in photography. When compared to silver halides, which are used in photography due to their sensitivity to light, AgNO3 is quite stable when exposed to light.
Structure of AgNO3
The nitrate ion described above consists of one nitrogen atom which is surrounded by three oxygen atoms. The nitrogen-oxygen bonds in this ion are similar to each other. The formal charge assigned to the nitrogen atom is -1, whereas each oxygen atom holds a charge of -⅔. The net charge associated with the nitrate ion is -1, which is quenched by the +1 charge held by the Ag+ ion via an ionic bond in AgNO3. It can be noted that the structure of the nitrate ion is stabilized by resonance.
Properties of Silver Nitrate
Some important physical and chemical properties of silver nitrate are listed in this subsection.
Physical Properties
* The molar mass of silver nitrate is 169.872 grams per mole.
* AgNO3 has a colourless appearance in its solid-state and is odourless.
* In its solid state, it has a density of 4.35 grams per cubic centimetre. Its density in the liquid state at a temperature of 210 degrees C corresponds to 3.97 g/{cm}^{3}.
* The melting and boiling points of silver nitrate are 482.8 K and 713 K respectively. However, this compound tends to decompose at temperatures approaching its boiling point.
* Silver nitrate, like most ionic compounds, dissolves readily in water. Its solubility in water corresponds to 122 g /100mL at 0 oC and 256g / 100mL at a temperature of 25 degrees centigrade.
* The crystal structure of AgNO3 is orthorhombic.
Chemical Properties
* The hazards of AgNO3 include its toxic and corrosive nature.
* The reaction between silver nitrate and ethanol is explosive.
* The silver present in the silver nitrate compound is displaced by copper, which forms copper nitrate. The chemical equation for this reaction is given by
2AgNO3 + Cu → Cu(NO3)2 + 2Ag
* When heated to 440 oC, this compound completely decomposes to give oxygen, nitrogen dioxide, and silver.
* Silver nitrate on decomposition gives silver, oxygen gas and nitrite.
* It can be noted that even though metal nitrates generally decompose to yield metal oxides, the decomposition reaction of silver nitrate gives rise to elemental silver because silver oxide decomposes at an even lower temperature than AgNO3.
Uses of Silver Nitrate
Silver nitrate has a wide range of applications in many fields such as biology, chemical synthesis, and medicine. Some of these uses of AgNO3 are listed below.
* Silver nitrate is a very versatile compound because the nitrate ion can be replaced by other ligands that can bind to the silver ion.
* Due to the ability of this compound to form a precipitate of silver halides when treated with halide ions, it is used while making photographic films.
* Many silver-based explosives can be prepared with a precipitation reaction of silver nitrate.
* In the field of inorganic chemistry, halides are extracted with the help of this compound.
* The branch of chemistry known as analytical chemistry uses this reaction to check for the presence of halide anions such as the iodide, bromide, or chloride ions.
* Mixtures of alkenes can be separated with the help of this compound since the silver cation binds with alkenes in a reversible fashion.
* When diluted with water to a concentration of 0.5%, silver nitrate can serve as an antiseptic in many medical setups.
* A diluted solution of AgNO3 can be administered to the eyes of a baby which is born to a mother suffering from gonorrhea, which combats the gonorrhoea bacteria and protects the baby from the onset of blindness.
* This compound is also known to be used for the treatment and the removal of unwanted warts in human beings.
Frequently Asked Questions:
What are the uses of silver nitrate?
Silver nitrate is widely used in many organic synthesis reactions in several ways. For example, for the deprotection and oxidation reactions. The Ag+ ion reversibly binds alkenes, and selectively adsorbing silver nitrate can be used to isolate alkene mixtures. The resulting adduct can be decomposed (in order to release the free alkene) with ammonia. Silver nitrate has been, in the past, used for silver staining (a process that employs silver or silver compounds to selectively change the appearance of a specific object). This compound is also used in medicine owing to its antiseptic qualities.
Is silver nitrate dangerous?
Silver nitrate is an oxidant and must, therefore, be kept away from organic compounds. Despite its widespread usage (especially in extremely low amounts) for the prevention of gonorrhoea and to stop bleeding from the nose, silver nitrate is often highly toxic and corrosive. Short-term exposure to this compound does not cause any immediate side effects apart from the development of a violet, brown or black stain on the part of the skin that was in contact with the silver nitrate. However, exposure to this compound over long periods of time is often accompanied by damage to the eyes. This compound is widely classified as an irritant to the skin and the eyes.
How is silver nitrate prepared?
Silver nitrate is usually prepared by combining silver with nitric acid. Common silver objects used in these reactions include silver bullions and silver foils. The products formed in this reaction include silver nitrate, water, and nitrogen oxides. The by-products of this chemical reaction depend on the nitric acid concentration that is used. It is important to note that this reaction must be carried out under a fume hood because of the evolution of poisonous oxides of nitrogen during the chemical reaction.

Hi,
#10791. What does the term in Biology Immunity mean?
#10792. What does the term in Biology Immunoglobulin mean?
Hi,
#5997. What does the noun lagoon mean?
#5998. What does the noun laissez faire mean?
Hi,
#2595. In which part of human body is Cuneus situated?
Ferric Sulfate
Gist
Ferric sulfate is a yellowish-brown or grayish-white inorganic salt used primarily as a coagulant in water treatment, a pigment, and a hemostatic agent in dentistry. It is highly acidic, produced by oxidizing iron(II) sulfate with sulfuric acid. It acts as an astringent.
The main function of ferric sulfate is as a hemostatic agent in different medical practices. This hemostatic function is achieved when ferric sulfate is applied directly in the damaged tissue. Once applied, ferric sulfate forms ferric ion-protein complex which helps the sealing of the damaged vessels mechanically.
Summary
Iron(III) sulfate or ferric sulfate (British English: sulphate instead of sulfate) is a family of inorganic compounds with the formula Fe2(SO4)3(H2O)n. A variety of hydrates are known, including the most commonly encountered form of "ferric sulfate". Solutions are used in dyeing as a mordant and as a coagulant for industrial wastes. Solutions of ferric sulfate are also used in the processing of aluminum and steel.
Production
Ferric sulfate solutions are usually generated from iron wastes. The actual identity of the iron species is often vague, but many applications do not demand high-purity materials. It is produced on a large scale by treating sulfuric acid, a hot solution of ferrous sulfate, and an oxidizing agent. Typical oxidizing agents include chlorine, nitric acid, and hydrogen peroxide.
Details
Ferric sulfate has the molecular formula of Fe2SO4, and it is a dark brown or yellow chemical agent with acidic properties. It is produced by the reaction of sulfuric acid and an oxidizing agent. It is used in different fields such as dermatology, dentistry and it is thought to present hemostatic properties by interacting chemically with blood proteins. By the FDA, ferric sulfate is a direct food substance affirmed in the GRAS category (Generally Recognized As Safe).
Indication
Ferric sulfate was first used in dermatology as part of the Monsel's solution. This solution is an antihemorrhagic agent used in skin and mucosal biopsies. The use of ferric sulfate in dermatology is under review as ferric sulfate is corrosive and injurious and it can cause degenerative changes that are not observed with other alternatives like collagen.
Ferric sulfate is also used as a coagulative and hemostatic agent. It is a mechanic hemostatic agent used directly on the damaged tissue.
In dentistry, ferric sulfate is used as a pulpotomy medicament to control pulpal bleeding, as an antibacterial agent and as a hemostatic reagent for restorative dentistry, for postextraction hemorrhage and for periradicular and endodontic surgery.
Pharmacodynamics
The administration of ferric sulfate as a dermatologic agent has showed delayed reepithelialization and dyspigmentation. Some studies have reported the generation of inflammation in the sites of administration of ferric sulfate.
Mechanism of action
The main function of ferric sulfate is as a hemostatic agent in different medical practices. This hemostatic function is achieved when ferric sulfate is applied directly in the damaged tissue. Once applied, ferric sulfate forms ferric ion-protein complex which helps the sealing of the damaged vessels mechanically. The formation of agglutinated protein complexes produces the generation of occlusion in the capillary orifices. The formation of the ferric protein complex is thought to be due to a chemical reaction between the acidic form of ferric sulfate and the blood proteins.
Absorption
Pharmacokinetic studies related to the absorption of ferric sulfate have not been performed.
Volume of distribution
Pharmacokinetic studies related to the volume of distribution of ferric sulfate have not been performed.
Protein binding
Ferric sulfate presents very high protein binding properties, this property is thought to be due to its acidic profile.
Metabolism
Pharmacokinetic studies related to the metabolism of ferric sulfate have not been performed.
Additional Information
Ferric sulfate appears as a yellow crystalline solid or a grayish-white powder. The primary hazard is the threat to the environment. Immediate steps should be taken to limit its spread to the environment. It is used for water purification, and as a soil conditioner.
Iron(3+) sulfate is a compound of iron and sulfate in which the ratio of iron(3+) to sulfate ions is 3:2. It has a role as a catalyst, a mordant and an astringent. It is an iron molecular entity and a metal sulfate. It contains an iron(3+).
Ferric sulfate has the molecular formula of Fe2SO4, and it is a dark brown or yellow chemical agent with acidic properties. It is produced by the reaction of sulfuric acid and an oxidizing agent. It is used in different fields such as dermatology, dentistry and it is thought to present hemostatic properties by interacting chemically with blood proteins. By the FDA, ferric sulfate is a direct food substance affirmed in the GRAS category (Generally Recognized As Safe).
The main function of ferric sulfate is as a hemostatic agent in different medical practices. This hemostatic function is achieved when ferric sulfate is applied directly in the damaged tissue. Once applied, ferric sulfate forms ferric ion-protein complex which helps the sealing of the damaged vessels mechanically.
Iron(III) Sulfate is an inorganic compound that is also termed ferric sulfate. Its chemical formula is Fe2(SO4)3. In iron III sulfate, each iron atom has ionic bonds with the sulfate. A variety of hydrates of iron III sulfate are known, such as nonahydrate, anhydrous monohydrate, etc. In fact, they are the most commonly encountered form of "Iron III Sulfate". It is slightly soluble in water and very hygroscopic. It is sparingly soluble in alcohol and negligibly soluble in acetone and ethyl acetate. It is not soluble in sulfuric acid and ammonia. As iron III sulfate is insoluble in sulfuric acid, it is used for producing iron III sulfate. It emits toxic fumes of iron and sulfur oxide when heated to decomposition. It is a threat to the environment and immediate steps should be taken to control its spread in the environment. It is used as a coagulant in water purification, as an astringent, and as a soil conditioner. It is corrosive to copper, copper alloys, mild steel, and galvanized steel.
Uses of Iron III Sulfate
* Iron III Sulfate is mainly used as a coagulant in water purification and sewage treatment.
* The solution of ferric sulfate is used as a mordant in dyeing and calico printing.
* It is also used in the preparation of iron salts and pigments, in the ferric salt leaching process, in soil conditioners, and in the coal conversion process.
* It acts as a disinfectant, polymerization catalyst, and hemostatic agent for endodontic surgery.
* It is also used in etching aluminium and in pickling stainless steel and copper.
* It is also used as a solids removal agent and oxidizing agent.
Hazards
* Iron III sulfate is a threat to the environment and immediate steps should be taken to control its spread in the environment.
* Though Iron III Sulfate is a stable, not flammable compound, it emits toxic fumes of iron and sulfur oxide when heated to decomposition.
* Prolonged exposure to Iron III Sulfate is toxic to the lungs and mucous membranes and may cause damage to them.
* Contact with this causes skin irritation and may also cause an allergic skin reaction.
* Inhalation of its dust irritates the nose and throat, and its ingestion irritates the mouth and stomach.
* It is harmful to aquatic life with long-lasting effects.

Hi,
#9881.
Hi,
#6374.
Hi,
2735.
2521) Odometer
Gist
An odometer is an instrument on a vehicle's dashboard that measures and displays the total distance traveled, typically in miles or kilometers. It works via mechanical gears or electronic sensors (on modern cars) tracking wheel rotations to determine distance, crucial for maintenance tracking and resale value.
Odometer includes the root from the Greek word hodos, meaning "road" or "trip". An odometer shares space on your dashboard with a speedometer, a tachometer, and maybe a "tripmeter". The odometer is what crooked car salesmen tamper with when they want to reduce the mileage a car registers as having traveled.
Summary
An odometer is a device that registers the distance traveled by a vehicle. Modern digital odometers use a computer chip to track mileage. They make use of a magnetic or optical sensor that tracks pulses of a wheel that connects to a vehicle’s tires. This data is stored in the engine control module (ECM). Odometers use these stored values to determine the total distance traveled by a vehicle.
Analog or mechanical odometers consist of a train of gears (with a gear ratio of 1,000:1) that causes a drum, classified in tenths of a mile or kilometre, to make one turn per mile or kilometre. A series of usually six such drums is arranged in such a way that one of the numerals on each drum is visible in a rectangular window. The drums are coupled so that 10 revolutions of the first cause one revolution of the second, and so forth, with the numbers appearing in the window representing the vehicle’s accumulated mileage.
The Roman architect and engineer Vitruvius is credited with inventing the initial version of an odometer in 15 bce. The concept consisted of a chariot wheel that turned 400 times to show one Roman mile. This wheel was mounted in a frame with a 400-tooth cogwheel. For every 400 rotations of the chariot wheel, the cogwheel would drop one pebble. In 1642 the French mathematician Blaise Pascal used the same principle to create an apparatus that used gears and wheels. For every 10 rotations of a gear, a second gear advanced one place. The modern odometer was invented about 1847 by pioneers William Clayton and Orson Pratt, members of the Church of Jesus Christ of Latter-day Saints. They attached their apparatus to a wagon wheel while they traversed the plains from Nebraska to the Great Salt Lake valley.
Details
An automobile’s most prominent yet unexplored part is the odometer. It is placed behind the steering wheel on the dashboard. It displays the distance the car has run. Odometer readings are beneficial to car owners when selling the vehicle. It helps evaluate the mileage or plans for car service.
Odometers can be mechanical, electrical, or a combination of the two. They are also known as mileometer or milometers in countries with imperial units or US customary units. Odometer is the most widely used name, especially in the UK and the Commonwealth countries.
Meaning
An odometer is a device used to measure the displacement of an object. It measures the distance travelled between the start point and the endpoint. Odometer is derived from two Greek words that mean path and measure.
Who invented the odometer?
Vitruvius, a Roman architect and engineer, is credited for the invention of the odometer in the 15th century. He used a standard chariot wheel, mounted on a frame with a 400-teeth cogwheel, and turned it 400 times in a Roman mile. The cogwheel employed a gear that slipped a stone into the box for every mile. Thus, it helped learn the miles covered by counting the pebbles.
In the 16th century, Blaise Pascal invented a calculating machine called Pascaline. It was a prototype of an odometer—the Pascaline comprised gears and wheels, where each gear had ten teeth. Every time a tooth completed a revolution, the second gear was engaged. This principle is used in the mechanical odometer.
English military engineer Thomas Savery invented an odometer for ships. In 1775, Ben Franklin, a statesman and a writer, created a simple odometer that measured the mileage of the routes. He attached it to his carriage.
In 1847, the Mormon Pioneers invented an odometer while crossing the plains from Missouri to Utah. Also known as a roadometer, they attached it to the wagon’s wheel, and when the wagon started the journey, it counted the wheel revolutions. Orson Pratt and William Clayton designed the odometer, and Appleton Milo Harmon, the carpenter, built it.
In 1854, Nova Scotia’s Samuel McKeen designed another early version of the odometer. The device measured driven mileage. He attached the device to the carriage side and measured the miles with wheels turning.
Types of odometer:
There are two types of odometers.
1) Mechanical odometers
2) Electronic odometers
Mechanical odometers
Mechanical odometers start with the transmission. The transmission system contains a small gear that measures the odometer advancing. This small gear is connected to the speedometer drive cable. The other end of this cable is connected to the instrument cluster.
The internal transmission gear turns when the engine is turned on, and the car starts moving. This internal transmission gear motion is conveyed to another set of gears linked to changeable digits by the connected drive cable. Thus, the counting begins from the right side of the group of numbers.
The process continues till the distance travelled by automobile compels the left side digits to roll over. This counting process repeats until all the adjacent numbers touch their apex values. Then, all the digits are set back to zero, and it starts again.
Mechanical odometers are not always precise and a hundred per cent accurate.
Electronic odometers
After the mechanical odometers came the electronic ones. They are also known as digital odometers. They depend on the automobile’s electronics for establishing accurate mileage.
Electronic odometers, like mechanical ones, employ a special gear for changing the count seen on the dashboard. In addition, a magnetic sensor replaces the drive cable to track the gear turns in the transmission. The wires conduct the obtained signal to the car’s onboard computer that interprets and converts the data into mileage count.
The advantage of electronic odometers over mechanical ones is that they provide better accuracy. In addition, no one can manipulate electronic odometers easily, hence giving an accurate count of the vehicle’s mileage.
Odometers come with an additional trip meter called a trip odometer. It helps the car owners determine the mileage for any particular distance without interfering with the primary odometer reading.
Conclusion
The primary purpose of an odometer is to measure the distance travelled by the vehicle. In addition, odometer readings help determine various maintenance milestones such as tyre rotations, oil changes etc. Dealers use odometer readings to estimate the vehicle’s valuation in the used car market. Resetting odometer values require changing the entire transmission system of a car. Hence, odometer readings are very difficult to reset. Also, tampering with the reading is considered a fraud and punishable by law.
Additional Information
Mechanical odometers have been counting the miles for centuries. Although they are a dying breed, they are incredibly cool because they are so simple! A mechanical odometer is nothing more than a gear train with an incredible gear ratio.
The odometer we took apart for this article has a 1690:1 gear reduction! That means the input shaft of this odometer has to spin 1,690 times before the odometer will register 1 mile.
Odometers like this are being replaced by digital odometers that provide more features and cost less, but they aren't nearly as cool. In this article, we'll take a look inside a mechanical odometer, and then we'll talk about how digital odometers work.
Mechanical Odometers
Mechanical odometers are turned by a flexible cable made from a tightly wound spring. The cable usually spins inside a protective metal tube with a rubber housing. On a bicycle, a little wheel rolling against the bike wheel turns the cable, and the gear ratio on the odometer has to be calibrated to the size of this small wheel. On a car, a gear engages the output shaft of the transmission, turning the cable.
The cable snakes its way up to the instrument panel, where it is connected to the input shaft of the odometer.
The Gearing
This odometer uses a series of three worm gears to achieve its 1690:1 gear reduction. The input shaft drives the first worm, which drives a gear. Each full revolution of the worm only turns the gear one tooth. That gear turns another worm, which turns another gear, which turns the last worm and finally the last gear, which is hooked up to the tenth-of-a-mile indicator.
Each indicator has a row of pegs sticking out of one side, and a single set of two pegs on the other side. When the set of two pegs comes around to the white plastic gears, one of the teeth falls in between the pegs and turns with the indicator until the pegs pass. This gear also engages one of the pegs on the next bigger indicator, turning it a tenth of a revolution.
On the white wheel between the "3" and the "4," there are two pegs. One time per revolution, one of the gear teeth on the white gear falls in between these two pegs, causing the black gear next to it to move one-tenth of a revolution.
You can now see why, when your odometer "rolls over" a large number of digits (say from 19,999 to 20,000 miles), the "2" at the far left side of the display may not line up perfectly with the rest of the digits. A tiny amount of gear lash in the white helper gears prevents perfect alignment of all the digits. Usually, the display will have to get to 21,000 miles before the digits line up well again.
You can also see that mechanical odometers like this one are rewindable. In many older vehicles, driving in reverse could cause the mechanical odometer to run backward due to the straightforward gear mechanism. However, some mechanical odometers were equipped with mechanisms to prevent reverse counting, ensuring the mileage only increased regardless of the driving direction.
In the movie "Ferris Bueller's Day Off," in the scene where they have the car up on blocks with the wheels spinning in reverse -- that should've worked! In real life, the odometer would've turned back. Another trick is to hook the odometer's cable up to a drill and run it backwards to rewind the miles.
Computerized Odometers
If you make a trip to the bike shop, you most likely won't find any cable-driven odometers or speedometers. Instead, you will find bicycle computers. Bicycles with computers like these have a magnet attached to one of the wheels and a pickup attached to the frame. Once per revolution of the wheel, the magnet passes by the pickup, generating a voltage in the pickup. The computer counts these voltage spikes, or pulses, and uses them to calculate the distance traveled.
If you have ever installed one of these bike computers, you know that you have to program them with the circumference of the wheel. The circumference is the distance traveled when the wheel makes one full revolution. Each time the computer senses a pulse, it adds another wheel circumference to the total distance and updates the digital display.
Many modern cars use a system like this, too. Instead of a magnetic pickup on a wheel, they use a toothed wheel mounted to the output of the transmission and a magnetic sensor that counts the pulses as each tooth of the wheel goes by. Some cars use a slotted wheel and an optical pickup, like a computer mouse does. Just like on the bicycle, the computer in the car knows how much distance the car travels with each pulse, and uses this to update the odometer reading.
One of the most interesting things about car odometers is how the information is transmitted to the dashboard. Instead of a spinning cable transmitting the distance signal, the distance (along with a lot of other data) is transmitted over a single wire communications bus from the engine control unit (ECU) to the dashboard. The car is like a local area network with many different devices connected to it. Here are some of the devices that may be connected to the computer network in a car:
* Engine control unit (ECU)
* Climate control system
* Dashboard
* Power window controls
* Radio
* Anti-lock braking system
* Air bag control module
* Body control module (operates the interior lights, etc.)
* Transmission control module
Many vehicles use a standardized communication protocol, called SAE J1850, to enable all of the different electronics modules to communicate with each other.
The engine control unit counts all of the pulses and keeps track of the overall distance traveled by the car. This means that if someone tries to "roll back" the odometer, the value stored in the ECU will disagree. This value can be read using a diagnostic computer, which all car-dealership service departments have.
Several times per second, the ECU sends out a packet of information consisting of a header and the data. The header is just a number that identifies the packet as a distance reading, and the data is a number corresponding to the distance traveled. The instrument panel contains another computer that knows to look for this particular packet, and whenever it sees one it updates the odometer with the new value. In cars with digital odometers, the dashboard simply displays the new value. Cars with analog odometers have a small stepper motor that turns the dials on the odometer.

2458) Felix Bloch
Gist:
Life
Felix Bloch was born in Zurich, Switzerland, the son of a merchant, and studied at ETH and elsewhere. When the Nazis took power in 1933, he left Europe to work at Stanford University. After becoming an American citizen, he worked on atomic energy in Los Alamos during World War II and later on radar at Harvard University. Immediately after the war, he did his Nobel Prize-awarded work at Stanford. He became the first head of CERN outside Geneva in 1954-1955. Bloch was married and had four children.
Work
Protons and neutrons in nuclei act like small, rotating magnets. Atoms and molecules therefore align in a magnetic field. Radio waves can disturb their direction of rotation, but only in certain stages, in accordance with quantum mechanics. When the atoms return to their original positions, they emit electromagnetic radio waves with frequencies characteristic of different elements and isotopes. In 1946, Felix Bloch and Edward Purcell developed methods for precise measurement, making it possible to study different materials’ compositions.
Summary
Felix Bloch (born Oct. 23, 1905, Zürich, Switz.—died Sept. 10, 1983, Zürich) was a Swiss-born American physicist who shared (with E.M. Purcell) the Nobel Prize for Physics in 1952 for developing the nuclear magnetic resonance method of measuring the magnetic field of atomic nuclei.
Bloch’s doctoral dissertation (University of Leipzig, 1928) promulgated a quantum theory of solids that provided the basis for understanding electrical conduction. Bloch taught at the University of Leipzig until 1933; when Adolf Hitler came to power he emigrated to the United States and was naturalized in 1939. After joining the faculty of Stanford University, Palo Alto, Calif., in 1934, he proposed a method for splitting a beam of neutrons into two components that corresponded to the two possible orientations of a neutron in a magnetic field. In 1939, using this method, he and Luis Alvarez (winner of the Nobel Prize for Physics in 1968) measured the magnetic moment of the neutron (a property of its magnetic field). Bloch worked on atomic energy at Los Alamos, N.M., and radar countermeasures at Harvard University during World War II.
Bloch returned to Stanford in 1945 to develop, with physicists W.W. Hansen and M.E. Packard, the principle of nuclear magnetic resonance, which helped establish the relationship between nuclear magnetic fields and the crystalline and magnetic properties of various materials. It later became useful in determining the composition and structure of molecules. Nuclear magnetic resonance techniques have become increasingly important in diagnostic medicine.
Bloch was the first director general of the European Organization for Nuclear Research (1954–55; CERN).
Details
Felix Bloch (23 October 1905 – 10 September 1983) was a Swiss-American theoretical physicist who shared the 1952 Nobel Prize in Physics with Edward Mills Purcell "for their development of new methods for nuclear magnetic precision measurements and discoveries in connection therewith".
He was the first Stanford University Nobel laureate.
Bloch made fundamental theoretical contributions to the understanding of ferromagnetism and electron behavior in crystal lattices. He is also considered one of the developers of nuclear magnetic resonance.
Education
Bloch was born on 23 October 1905 in Zurich, Switzerland, to Jewish parents, Gustav Bloch and Agnes Mayer. Gustav was financially unable to attend university and worked as a wholesale grain dealer in Zurich. Gustav moved to Zurich from Moravia in 1890 to become a Swiss citizen. Their first child was a girl born in 1902, while Felix was born three years later.
Bloch entered public elementary school at the age of six and is said to have been teased, in part because he "spoke Swiss German with a somewhat different accent than most members of the class". He received support from his older sister during much of this time, but she died at the age of 12, devastating Felix, who is said to have lived a "depressed and isolated life" in the following years. Bloch learned to play the piano by the age of 8 and was drawn to arithmetic for its "clarity and beauty". Bloch graduated from elementary school at twelve and enrolled in the Cantonal Gymnasium in Zurich for secondary school in 1918. He was placed on a six-year curriculum here to prepare him for university. He continued his curriculum through 1924, even through his study of engineering and physics in other schools, though it was limited to mathematics and languages after the first three years.
After these first three years at the Gymnasium, at the age of 15, Bloch began to study at the ETH Zurich. Although he initially studied engineering, he soon changed to physics. During this time, he attended lectures and seminars given by Peter Debye and Hermann Weyl at the ETH Zurich and Erwin Schrödinger at the neighboring University of Zurich. A fellow student in these seminars was John von Neumann.
Bloch graduated in 1927, and was encouraged by Debye to go to the University of Leipzig to study under Werner Heisenberg. Bloch became Heisenberg's first graduate student, and gained his doctorate in 1928. His doctoral thesis established the quantum theory of solids, using waves to describe electrons in periodic lattices.
Career and research
Bloch remained in European academia, working on superconductivity with Wolfgang Pauli in Zurich; with Hans Kramers and Adriaan Fokker in the Netherlands; with Heisenberg on ferromagnetism, where he developed a description of boundaries between magnetic domains, now known as Bloch walls, and theoretically proposed a concept of spin waves, excitations of magnetic structure; with Niels Bohr in Copenhagen, where he worked on a theoretical description of the stopping of charged particles traveling through matter; and with Enrico Fermi in Rome.
In 1932, Bloch returned to Leipzig to assume a position as Privatdozent (lecturer). In 1933, immediately after Adolf Hitler came to power, Bloch left Germany out of fear of anti-Jewish persecution, returning to Zurich before traveling to Paris to lecture at the Institut Henri Poincaré.
In 1934, the chairman of Stanford Physics invited Bloch to join the faculty. Bloch accepted the offer and emigrated to the United States. In the fall of 1938, Bloch began working with the 37 inch cyclotron at the University of California, Berkeley, to determine the magnetic moment of the neutron. Bloch went on to become the first professor of theoretical physics at Stanford. In 1939, he became a naturalized citizen of the United States.
During World War II, Bloch briefly worked on the atomic bomb project at Los Alamos. Disliking the military atmosphere of the laboratory and uninterested in the theoretical work there, Bloch left to join the radar project at Harvard University.
After the war, he concentrated on investigations into nuclear induction and nuclear magnetic resonance, which are the underlying principles of MRI. In 1946, he proposed the Bloch equations, which determine the time evolution of nuclear magnetization. He was elected to the National Academy of Sciences in 1948. Along with Edward Purcell, Bloch was awarded the Nobel Prize in Physics in 1952 for his work on nuclear magnetic induction.
When CERN was being set up in the early 1950s, its founders were searching for someone of stature and international prestige to head the fledgling international laboratory, and in 1954 Professor Bloch became CERN's first director-general, at the time when construction was getting under way on the present Meyrin site and plans for the first machines were being drawn up. After leaving CERN, he returned to Stanford University, where he in 1961 was made Max Stein Professor of Physics.
In 1964, he was elected a foreign member of the Royal Netherlands Academy of Arts and Sciences. He was also a member of the American Academy of Arts and Sciences and the American Philosophical Society.
Family
On 14 March 1940, Bloch married Lore Clara Misch (1911–1996), a fellow physicist working on X-ray crystallography, whom he had met at an American Physical Society meeting. They had four children, twins George Jacob Bloch and Daniel Arthur Bloch (born 15 January 1941), son Frank Samuel Bloch (born 16 January 1945), and daughter Ruth Hedy Bloch (born 15 September 1949).
Bloch died on 10 September 1983 in Zurich at the age of 77. In 2025 Bloch's family donated his Nobel Prize medal to CERN.

Q: What did the bird say to the racing squirrel?
A: You walnut beat that!
* * *
Q: How many squirrels does it take to change a light bulb?
A: Actually, none because squirrels only change bulbs that are NUT broken.
* * *
Q: Why does it take more than one squirrel to screw in a lightbulb?
A: Because they're so darn stupid!
* * *
Q: Why was the squirrel late for work?
A: Traffic was NUTS.
* * *
Q: How do you catch a carpenter squirrel (definition: a squirrel that likes power tools)?
A: Go to Home Depot and pretend to be nut-wood.
* * *
Comedy Quotes - II
1. We participate in a tragedy; at a comedy we only look. - Aldous Huxley
2. Friends applaud, the comedy is over. - Ludwig van Beethoven
3. I am completely open to doing a romantic comedy, but I will never do something just for the sake of doing a specific genre or because it's the time or place to do a different type of movie. I think that would be a huge mistake. - Leonardo DiCaprio
4. I will do comedy until the day I die: inappropriate comedy, funny comedy, gender-bending, twisting comedy, whatever comedy is out there. - Sandra Bullock
5. Even actresses that you really admire, like Reese Witherspoon, you think, 'Another romantic comedy?' You see her in something like 'Walk the Line' and think, 'God, you're so great!' And then you think, 'Why is she doing these stupid romantic comedies?' But of course, it's for money and status. - Gwyneth Paltrow
6. When I tried to branch out into comedy, I didn't do very well at it, so I went back to doing what I do naturally well, or what the audience expects from me - action pictures. - Sylvester Stallone
7. As for doing more dramatic work over comedy, I do whatever turns me on at the moment. - Sandra Bullock
8. I think that you can fall into bad habits with comedy... It's a tightrope to stay true to the character, true to the irony, and allow the irony to happen. - Ben Kingsley
Field Vision Test
Gist
A visual field test (perimetry) maps your peripheral and central vision to detect blind spots (scotomas). It is essential for diagnosing and managing glaucoma, neurological conditions (e.g., MS, tumors, strokes), and monitoring medication side effects. Typically lasting 5–10 minutes per eye, patients click a button when they see light flashes while staring at a central point.
A visual field test can determine if you have blind spots, known as scotomas, in your vision and where they are. A blind spot's size and shape can show how eye disease or a brain disorder is affecting your vision.
Summary
A visual field test is an eye examination that can detect dysfunction in central and peripheral vision which may be caused by various medical conditions such as glaucoma, stroke, pituitary disease, brain tumours or other neurological deficits. Visual field testing can be performed clinically by keeping the subject's gaze fixed while presenting objects at various places within their visual field. Simple manual equipment can be used such as in the tangent screen test or the Amsler grid. When dedicated machinery is used it is called a perimeter.
The exam may be performed by a technician in one of several ways. The test may be performed by a technician directly, with the assistance of a machine, or completely by an automated machine. Machine-based tests aid diagnostics by allowing a detailed printout of the patient's visual field.
Details
A visual field test measures your peripheral vision, or how well you can see above, below and to the sides of something you’re looking at. It’s also called a perimetry test. Visual field testing is important for many conditions, including glaucoma.
Overview:
What is a visual field test?
A visual field test is a simple and painless test an eye care provider gives you to diagnose or monitor various eye conditions.
A visual field test measures two things:
* How far up, down, left and right your eye sees without moving (when you’re looking straight ahead).
* How sensitive your vision is in different parts of the visual field, which is the name for the entire area that you can see.
Your eyes normally see a wide area of the space in front of you. Without moving your eyes, you can see not only what’s straight ahead, but also some of what’s above, below and off to either side. Providers call all of the area you can see that isn’t right in front of you “peripheral vision.” This surrounds the area that’s right in front of you that you can see (central vision).
Vision is usually best right in the middle of the visual field, so you probably turn your eyes toward the things you want to see more clearly. The farther away from the center of your vision an object is, the less clearly you can see it. When an object moves far enough to the side, it disappears from your vision completely.
When is a visual field test performed?
When you visit an optometrist or ophthalmologist, a visual field test is part of a routine eye exam. Visual field testing can help your eye care provider find early signs of diseases like glaucoma that gradually damage vision. Some people with glaucoma don’t notice any problems with their vision, but the visual field test shows a loss of peripheral vision.
A visual field test can also help your provider find out more about the part of your nervous system that allows you to see. The visual part of your nervous system includes:
* Your retina, the part of your eye that’s like a translator that changes light energy into an electrical signal.
* Your optic nerve, the nerve that carries the signals to your brain so they can become images.
* Your brain, the place where the signals become the images you see.
Issues with any part of this system can change your visual field. There are well-known patterns in the test results that help providers recognize certain types of injury or disease.
By repeating visual field tests at regular intervals, providers also can tell whether your condition is getting better or worse.
Medical conditions that might cause a provider to order a visual field test
Your healthcare provider may want you to have a visual field test if you have (or they think you may have) certain conditions. Providers use the results to both diagnose and monitor conditions such as:
* Glaucoma.
* Stroke.
* Macular degeneration.
* Multiple sclerosis (MS).
* Graves’ disease.
* Pituitary gland disease.
* Blind spot (scotoma).
Why do some people need to have visual field tests many times?
Sometimes your eye care provider will want to repeat the visual field test right away to make sure the results are accurate. If you’re tired, for example, the test results can be unreliable.
Your provider might also recommend that you take a visual field test again in a few weeks, a few months or a year. This might be necessary to make sure that they find any new problems early. When you have certain eye conditions, your provider will do visual field tests regularly to find out how well the treatment is working.
Visual field tests are especially important in the treatment of glaucoma. These tests will tell the provider if you’re losing vision even before you notice. That’s just one of the reasons why people who have glaucoma should keep all of their appointments with their provider.
Test Details:
What happens during a visual field test?
You don’t have to prepare for a visual field test. It’s not invasive, so you aren’t likely to have any side effects.
There are several types of visual field tests, but they all have one thing in common: you look straight ahead at one point and signal when you see an object or a light somewhere off to the side.
Your provider will explain to you exactly where to look so that the test is accurate.
The two most basic types of visual field tests are very simple:
* Amsler grid: The Amsler grid is a pattern of straight lines that make perfect squares. You look at a large dot in the middle of the grid and describe any areas where the lines look blurry, wavy or broken. The Amsler grid is a quick test that only measures the middle of the visual field (your central vision) and provides your doctor with a small amount of information.
* Confrontation visual field: The term “confrontation” in this test just means that the person giving the test sits facing the person having the test, about 3 or 4 feet (around 1 meter) away. The provider holds their arms straight out to the sides. You look straight ahead, and the tester moves one hand and then the other inward toward you. You give a signal as soon as you see their hand.
The confrontation visual field test measures only the outer edge of the visual field. It’s not very exact.
Other types of visual field tests
You may hear about different types of or terms for visual field tests, including static and kinetic perimetry tests. (Perimetry test is another way of saying peripheral vision test.)
* Kinetic perimetry tests: A kinetic perimetry test is one in which the person giving the test moves an object around, and you tell them when you can see it. Providers often use the Goldmann perimetry test.
* Static perimetry tests: Automated peripheral vision tests are static perimetry tests. You look into a bowl-shaped machine and respond by pressing buttons when you see the object. Common types of static tests are the Humphrey and the Octopus.
How long does a visual field test take?
A test usually isn’t longer than about five to 10 minutes per eye.
What kind of visual field tests give more detailed information?
Computerized instruments are available to perform visual field tests and calculate results. These instruments give more reproducible and accurate results because:
* Your head is always in the same place during the test.
* The instrument has a large central “target” for you to look at, so the center of the visual field stays steady.
* The instrument uses tiny spots of light to test vision. The provider can change the brightness and color of the light to measure the sensitivity of vision at each location.
* There are clear standards for “normal” results. The instrument can compare each new test to these standards.
Results and Follow-Up:
What do the results of the visual field test mean?
A “normal” visual field test means that you can see about as well as people without vision issues.
The visual field test shows the amount of vision loss and the affected areas. The instrument prints the results as patterns of dots or numbers. The patterns tell your provider how well your eyes and visual field system work. This helps your provider diagnose an underlying health condition and what treatment you need.
A test that shows visual field loss means that vision in some areas isn’t as keen as it should be. A test could show that you have a small area of lost vision, or all vision lost in large areas.
When should I know the results of the test?
Generally, your provider should be able to give you results right away.
What are the next steps if the results are abnormal?
Abnormal results may mean different things. These results can indicate different types of issues, including glaucoma, macular degeneration or stroke. The follow-up will vary.
Your eye care provider will discuss treatment options with you.
When should I call my provider?
You should always contact your eye care provider if you have any new vision loss or eye discomfort. If you have sudden vision loss or eye pain, go to an emergency room for immediate medical help.
Additional Information
A visual field test is a diagnostic procedure that measures a person's entire field of vision, including peripheral (side) and central vision. It evaluates how well you can see in different areas of your vision and is commonly used to detect, diagnose, and monitor various eye and neurological conditions. The test plays a crucial role in identifying issues that may not be apparent during a routine eye exam, especially problems affecting peripheral vision.
Visual field testing can help uncover conditions such as glaucoma, retinal disorders, optic nerve damage, and neurological diseases like strokes or brain tumors. By mapping out the areas where vision is diminished or absent, it provides valuable insights into the health of your eyes and the visual pathways in your brain.
Importance of Test Results Interpretation
Accurate interpretation of visual field test results is critical for effective diagnosis and treatment planning. The results are presented as a detailed map showing areas where vision is normal, reduced, or absent. Key aspects of result interpretation include:
* Detection of Blind Spots: Identifying areas where vision is missing, which may indicate damage to the retina or optic nerve.
* Symmetry Analysis: Comparing the visual fields of both eyes to detect asymmetrical vision loss, which can be a sign of neurological conditions.
* Severity and Progression: Monitoring changes over time to assess the progression of diseases like glaucoma.
Patients typically receive a detailed explanation of their test results from an eye care professional, including recommendations for treatment or follow-up testing if necessary.
Uses of a Visual Field Test
Visual field tests serve a variety of purposes in both ophthalmology and neurology. Common uses include:
* Glaucoma Diagnosis and Monitoring: Identifies early signs of vision loss associated with glaucoma and tracks progression.
* Assessment of Retinal Disorders: Detects damage caused by conditions like diabetic retinopathy or retinal detachment.
* Optic Nerve Evaluation: Evaluates the health of the optic nerve, often impacted by optic neuritis or optic neuropathy.
* Neurological Conditions: Identifies vision changes due to strokes, brain tumors, or other neurological disorders.
* Pre-Surgical Planning: Assists in determining the extent of vision impairment before eye surgeries.
* Evaluation of Medication Effects: Monitors vision changes in patients taking medications that may affect eye health.
How to Prepare for a Visual Field Test
Proper preparation ensures accurate results from a visual field test. Follow these steps to get ready:
* Inform Your Eye Doctor: Share your medical history, including any eye conditions, neurological issues, or medications you are taking.
* Rest Well: Ensure you are well-rested before the test to reduce fatigue, which can affect performance.
* Wear Glasses or Contacts if Needed: Bring any corrective eyewear to the appointment, as the test may require you to wear them.
* Avoid Driving Before the Test: The procedure may involve pupil dilation, temporarily affecting your ability to drive.
* Follow Specific Instructions: Your doctor may provide additional preparation guidelines based on your individual needs.
By following these steps, you can help ensure the test provides the most accurate representation of your visual field.
What to Expect During the Procedure
A visual field test is a painless and non-invasive procedure typically performed in an eye doctor's office. Here is what what you can expect:
* Positioning: You will sit in front of a specialized machine and place your chin on a rest to stabilize your head.
* Focus on a Target: You will be asked to focus on a central point while small lights or objects appear in different parts of your visual field.
* Responding to Stimuli: You'll press a button or verbally indicate when you see the lights.
* Eye-by-Eye Testing: Each eye is tested separately by covering the other eye.
* Duration: The test typically takes 15-30 minutes to complete.
Patients can resume normal activities immediately after the test unless they have had their pupils dilated, in which case temporary visual sensitivity may occur.
Normal Range for Visual Field Test Results
Normal results indicate that your visual field is intact and free of significant blind spots beyond the natural blind spot (caused by the optic nerve head). Specific findings in a normal test include:
* Symmetrical vision between both eyes.
* Full peripheral vision within the expected range for your age.
* No unexplained areas of vision loss or distortion.
* Abnormal results may require further investigation to determine the underlying cause and develop an appropriate treatment plan.
Benefits of a Visual Field Test
Visual field testing offers numerous benefits for maintaining eye and neurological health. These include:
* Early Detection: Identifies vision problems before noticeable symptoms develop.
* Comprehensive Assessment: Provides a detailed map of your visual capabilities.
* Monitoring Disease Progression: Tracks changes in vision over time for conditions like glaucoma.
* Guiding Treatment Decisions: Helps tailor treatments based on the specific pattern of vision loss.
* Preventing Vision Loss: Enables timely interventions to preserve remaining vision.
Limitations and Risks of a Visual Field Test
While visual field testing is highly beneficial, it has certain limitations and risks:
* False Positives or Negatives: Patient fatigue or inattention can lead to inaccurate results.
* Limited Scope: Does not provide detailed images of the eyes' internal structures.
* Temporary Discomfort: Prolonged focus during the test may cause mild eye strain.
* Not a Standalone Diagnostic Tool: Often combined with other tests for a complete evaluation.
Understanding these limitations can help set realistic expectations for the procedure.
Frequently Asked Questions (FAQs) About Visual Field Tests:
1. Why is a visual field test important?
A visual field test is essential for detecting early signs of eye and neurological conditions, including glaucoma and optic nerve damage. It provides a detailed map of your field of vision, allowing doctors to diagnose problems that may not be noticeable during routine eye exams. Early detection through this test helps prevent further vision loss by enabling timely treatment and monitoring.
2. How often should I get a visual field test?
The frequency of visual field testing depends on your age, medical history, and risk factors. People with glaucoma or other eye conditions may need regular testing every 6-12 months. For routine eye health, adults should have a visual field test every 1-2 years as part of a comprehensive eye exam. Consult your doctor for personalized recommendations.
3. Is the visual field test painful?
No, the visual field test is completely painless and non-invasive. It involves sitting comfortably and responding to visual stimuli. Some patients may find it slightly tiring to maintain focus during the test, but there is no physical discomfort involved.
4. What do abnormal visual field test results mean?
Abnormal results indicate areas of reduced or missing vision, which could be caused by glaucoma, retinal conditions, optic nerve damage, or neurological issues like strokes. Your doctor will interpret the results and may recommend additional tests to determine the cause and guide treatment.
5. Can children undergo a visual field test?
Yes, children can undergo visual field testing if recommended by their doctor. The procedure is modified to suit their age and ability to follow instructions. It is often used to diagnose conditions like optic nerve disorders or monitor vision changes caused by neurological issues in children.
6. What is the difference between central and peripheral vision testing?
Central vision testing evaluates the ability to see details in the center of your vision, while peripheral vision testing assesses your ability to detect objects and movement in the outer areas of your field of vision. Visual field tests often include both types to provide a complete assessment.
7. Can a visual field test detect brain tumors?
Yes, a visual field test can help detect vision changes caused by brain tumors. Tumors affecting the optic pathways or visual centers in the brain can cause specific patterns of vision loss, which are identifiable through this test. Further imaging tests may be required for confirmation.
8. How accurate is a visual field test?
Visual field tests are highly accurate when performed correctly and under optimal conditions. Factors like patient attentiveness and proper calibration of the equipment influence the reliability of the results. Repeat testing may be necessary to confirm findings.
9. Are there alternatives to a visual field test?
Alternatives include fundus photography, optical coherence tomography (OCT), and perimetry tests. Each method has unique applications, and your doctor will choose the most appropriate one based on your condition and diagnostic needs.
10. What should I do if I fail a visual field test?
Failing a visual field test doesn’t always mean permanent vision loss. It indicates areas requiring further investigation. Follow your doctor’s recommendations for additional testing or treatment. Early intervention can often prevent further deterioration and improve outcomes.
Conclusion
The visual field test is an invaluable diagnostic tool for assessing and preserving eye and neurological health. By identifying early signs of vision loss and guiding treatment decisions, it plays a vital role in managing conditions like glaucoma and neurological disorders. While the procedure has certain limitations, its benefits in early detection and monitoring far outweigh them. Regular visual field testing, combined with comprehensive eye care, can help maintain optimal vision and quality of life. Consult your eye doctor to learn more about this important test and how it fits into your overall health plan.

Arthritis
Gist
Arthritis is inflammation of one or more joints, causing pain, stiffness, and swelling. Common types include osteoarthritis (wear-and-tear) and rheumatoid arthritis (autoimmune). Key risk factors include age, genetics, and obesity. Treatments include medications, physical therapy, and lifestyle changes, aiming to manage symptoms and improve function.
Good approaches for arthritis include low-impact exercise, weight management, heat/cold therapy, and an anti-inflammatory diet rich in fruits, vegetables, fish, and whole grains, while avoiding processed foods and sugar, alongside potential medications, physical therapy, and sometimes surgery for severe cases, with lifestyle changes being key. Balancing activity and rest, maintaining good posture, and using assistive devices can also significantly ease symptoms.
Summary
Arthritis is a general medical term used to describe a disorder in which the smooth cartilagenous layer that lines a joint is lost, resulting in bone grinding on bone during joint movement. Symptoms generally include joint pain and stiffness. Other symptoms may include redness, warmth, swelling, and decreased range of motion of the affected joints. In certain types of arthritis, other organs, such as the skin, are also affected. Onset can be gradual or sudden.
There are several types of arthritis. The most common forms are osteoarthritis (most commonly seen in weightbearing joints) and rheumatoid arthritis. Osteoarthritis usually occurs as a person ages and often affects the hips, knees, shoulders, and fingers. Rheumatoid arthritis is an autoimmune disorder that often affects the hands and feet. Other types of arthritis include gout, lupus, and septic arthritis. These are inflammatory based types of rheumatic disease.
Early treatment for arthritis commonly includes resting the affected joint and conservative measures such as heating or icing. Weight loss and exercise may also be useful to reduce the force across a weightbearing joint. Medication intervention for symptoms depends on the form of arthritis. These may include anti-inflammatory medications such as ibuprofen and paracetamol (acetaminophen). With severe cases of arthritis, joint replacement surgery may be necessary.
Osteoarthritis is the most common form of arthritis affecting more than 3.8% of people, while rheumatoid arthritis is the second most common affecting about 0.24% of people. In Australia about 15% of people are affected by arthritis, while in the United States more than 20% have a type of arthritis. Overall arthritis becomes more common with age. Arthritis is a common reason people are unable to carry out their work and can result in decreased ability to complete activities of daily living. The term arthritis is derived from arthr- (meaning 'joint') and -itis (meaning 'inflammation').
Details:
Overview
Arthritis and other rheumatic diseases are common conditions that cause pain, swelling, and limited movement. They affect joints and connective tissues around the body. Millions of people in the U.S. have some form of arthritis.
Arthritis means redness and swelling (inflammation) of a joint. A joint is where 2 or more bones meet. There are more than 100 different arthritis diseases. Rheumatic diseases include any condition that causes pain, stiffness, and swelling in joints, muscles, tendons, ligaments, or bones. Arthritis is usually ongoing (chronic).
Arthritis and other rheumatic diseases are more common in women than men. These conditions are often found in older people. But people of all ages may be affected.
The 2 most common forms of arthritis are:
* Osteoarthritis. This is the most common type of arthritis. It is a chronic disease of the joints, especially the weight-bearing joints of the knee, hip, and spine. It destroys the padding on the ends of bones (cartilage) and narrows the joint space. It can also cause bone overgrowth, bone spurs, and reduced function. It occurs in most people as they age. It may also occur in young people from an injury or overuse.
* Rheumatoid arthritis. This is an autoimmune disease that causes inflammation in the joint linings. The inflammation may affect all the joints. It can also affect organs, such as the heart or lungs.
Other forms of arthritis or related disorders include:
* Gout. This condition causes uric acid crystals to build up in small joints, such as the big toe. It causes pain and inflammation.
* Lupus. This is a chronic autoimmune disorder. It causes periods of inflammation and damage in joints, tendons, and organs.
* Scleroderma. This autoimmune disease causes thickening and hardening of the skin and other connective tissue in the body.
* Ankylosing spondylitis. This form of arthritis causes inflammation of the spinal joints. It may lead to severe chronic pain and discomfort. In more advanced cases, sections of the bones fuse together in an immobile position. It can also cause inflammation in other parts of the body. Though it primarily affects the spine, it can also affect the shoulders, hips, ribs, and the small joints of the hands and feet.
* Juvenile idiopathic arthritis or juvenile rheumatoid arthritis. This is a form of arthritis in children under the age of 16 that causes inflammation and joint stiffness. Children may have symptoms that last a limited time, such as a few months or years or in some cases a lifetime. Getting diagnosed and treated early may help prevent joint damage.
What causes arthritis?
The cause depends on the type of arthritis. Osteoarthritis is caused by wear and tear of the joint over time or because of overuse. Rheumatoid arthritis, lupus, and scleroderma are caused by the body’s immune system attacking the body’s own tissues. Gout is caused by the buildup of crystals in the joints. Some forms of arthritis can be linked to genes. People with genetic marker HLA-B27 have a higher risk for ankylosing spondylitis. For some other forms of arthritis, the cause is not known.
Who is at risk for arthritis?
Some risk factors for arthritis that can’t be changed include:
* Age. The older you are, the more likely you are to have arthritis.
* Gender. Women are more likely to have arthritis than men.
* Heredity. Some types of arthritis are linked to certain genes.
Risk factors that may be changed include:
* Weight. Being overweight or obese can damage your knee joints. This can make them more likely to develop osteoarthritis.
* Injury. A joint that has been damaged by an injury is more likely to develop arthritis at some point.
* Infection. Reactive arthritis can affect joints after an infection.
* Your job. Work that involves repeated bending or squatting can lead to knee arthritis.
What are the symptoms of arthritis?
Each person’s symptoms may vary. The most common symptoms include:
* Pain in 1 or more joints that doesn’t go away, or comes back.
* Warmth and redness in 1 or more joints.
* Swelling in 1 or more joints.
* Stiffness in 1 or more joints.
* Trouble moving 1 or more joints in a normal way.
These symptoms can look like other health conditions. Always see your health care provider for a diagnosis.
How is arthritis diagnosed?
Your provider will take your medical history and give you a physical exam. Tests may also be done. These include blood tests, such as:
* Antinuclear antibody test. This checks antibody levels in the blood.
* Complete blood count. This checks if your white blood cell, red blood cell, and platelet levels are normal.
* Creatinine. This test checks for kidney disease.
* Sedimentation rate. This test can find inflammation.
* Hematocrit. This test measures the number of red blood cells.
* RF (rheumatoid factor) and CCP (cyclic citrullinated peptide) antibody tests. These can help diagnose rheumatoid arthritis.
* White blood cell count. This checks the level of white blood cells in your blood.
* Uric acid. This helps diagnose gout.
Other tests may be done, such as:
* Joint aspiration (arthrocentesis). A small sample of synovial fluid is taken from a joint. It's tested to see if crystals or bacteria are present.
* X-rays or other imaging tests. These can tell how damaged a joint is.
* Urine test. This checks for protein and different kinds of blood cells.
* HLA tissue typing. This looks for genetic markers of ankylosing spondylitis.
* Skin biopsy. Tiny tissue samples are removed and checked under a microscope. This test helps to diagnose a type of arthritis that involves the skin, such as lupus or psoriatic arthritis.
* Muscle biopsy. Tiny tissue samples are removed and checked under a microscope. This test helps to diagnose conditions that affect muscles.
How is arthritis treated?
Treatment will depend on your symptoms, your age, and your general health. It will also depend on what type of arthritis you have and how bad the condition is. A treatment plan is tailored to each person with their provider.
There is no known cure for arthritis. The goal of treatment is often to limit pain and inflammation and to help the joint work. Treatment plans often use both short-term and long-term methods.
Short-term treatments include:
* Medicines. Short-term relief for pain and inflammation may include pain relievers, such as acetaminophen, aspirin, ibuprofen, or other nonsteroidal anti-inflammatory medicines.
* Heat and cold. Pain may be eased by using moist heat (warm bath or shower) or dry heat (heating pad) on the joint. Pain and swelling may be eased with cold (ice pack wrapped in a thin towel) on the joint.
* Joint immobilization. Using a splint or brace can help a joint rest and protect it from more injury.
* Massage. Lightly massaging painful muscles may increase blood flow and bring warmth to the muscle.
* Transcutaneous electrical nerve stimulation (TENS). Pain may be eased with a TENS device. The device sends mild, electrical pulses to nerve endings in the painful area. This blocks pain signals to the brain and changes how you feel pain.
* Acupuncture. Thin needles are inserted at certain points in the body. It may help the release of natural pain-relieving chemicals made by the nervous system. The procedure is done by a licensed provider.
Long-term treatments include:
* Disease-modifying antirheumatic drugs. These prescription medicines may slow down the disease and treat any immune system problems linked to the disease. Examples of these medicines include methotrexate, hydroxychloroquine, sulfasalazine, and chlorambucil.
* Corticosteroids. Corticosteroids reduce inflammation and swelling. These medicines, such as prednisone, can be taken by mouth (orally) or as a shot.
* Hyaluronic acid therapy. This is a joint fluid that appears to break down in people with osteoarthritis. It can be injected into a joint, such as the knee to help ease symptoms.
* Surgery. There are many types of surgery, depending on which joints are affected. Surgery may include arthroscopy, fusion, or joint replacement. Full recovery after surgery may take up to 6 months. A rehabilitation program after surgery is an important part of the treatment.
Arthritis treatment can include a team of health care providers, such as:
* Orthopedist or orthopedic surgeon.
* Rheumatologist.
* Physiatrist.
* Primary care (family medicine or internal medicine).
* Rehabilitation nurse.
* Dietitian.
* Physical therapist.
* Occupational therapist.
* Social worker.
* Psychologist or psychiatrist.
* Recreational therapist.
* Vocational therapist.
What are possible complications of arthritis?
Because arthritis causes joints to get worse over time, it can cause disability. It can cause pain and movement problems. You may be less able to carry out normal daily activities and tasks.
Living with arthritis
There is no known cure for arthritis. But it’s important to help keep joints working by reducing pain and inflammation. Work on a treatment plan with your provider that includes medicine and therapy. Work on lifestyle changes that can improve your quality of life. Lifestyle changes include:
* Weight loss. Extra weight puts more stress on weight-bearing joints, such as the hips and knees.
* Exercise. Some exercises may help reduce joint pain and stiffness. These include swimming, walking, low-impact aerobic exercise, and range-of-motion exercises. Stretching exercises may also help keep the joints flexible.
* Activity and rest. To reduce stress on your joints, switch between activity and rest. This can help protect your joints and reduce your symptoms.
* Using assistive devices. Canes, crutches, and walkers can help keep stress off certain joints and improve balance. Make sure walkers, canes, and other mobility devices are adjusted to meet your height and posture.
* Using adaptive equipment. Resachers and grabbers let you extend your reach and reduce straining. Dressing aids help you get dressed more easily.
* Managing use of medicines. Long-term use of some anti-inflammatory medicines can lead to stomach bleeding and other possible side effects. Work with your to create a plan to reduce this risk and manage your pain.
When to contact your doctor
Contact your if you have questions about your medicines, your symptoms get worse, or you have new symptoms.
Additional Information
Arthritis is extremely common, especially in people older than 50. It causes joint pain, stiffness and inflammation. Your provider will help you understand which type of arthritis you have, what’s causing it and which treatments you’ll need. You may need a joint replacement if you have severe arthritis that you can’t manage with other treatments.
Overview:
What is arthritis?
Arthritis is a disease that causes damage in your joints. Joints are places in your body where two bones meet.
Some joints naturally wear down as you age. Lots of people develop arthritis after that normal, lifelong wear and tear. Some types of arthritis happen after injuries that damage a joint. Certain health conditions also cause arthritis.
Arthritis can affect any joint, but is most common in people’s:
* Hands and wrists.
* Knees.
* Hips.
* Feet and ankles.
* Shoulders.
* Lower back (lumbar spine).
A healthcare provider will help you find ways to manage symptoms like pain and stiffness. Some people with severe arthritis eventually need surgery to replace their affected joints.
Visit a healthcare provider if you’re experiencing joint pain that’s severe enough to affect your daily routine or if it feels like you can’t move or use your joints as well as usual.
Types of arthritis
There are more than 100 different types of arthritis. Some of the most common types include:
* Osteoarthritis: Wear and tear arthritis.
* Rheumatoid arthritis: Arthritis that happens when your immune system mistakenly damages your joints.
* Gout: Arthritis that causes sharp uric acid crystals to form in your joints.
* Ankylosing spondylitis: Arthritis that affects joints near your lower back.
* Psoriatic arthritis: Arthritis that affects people who have psoriasis.
* Juvenile arthritis: Arthritis in kids and teens younger than 16.
Depending on which type of arthritis you have, it can break down the natural tissue in your joint (degeneration) or cause inflammation (swelling). Some types cause inflammation that leads to degeneration.
How common is arthritis?
Arthritis is extremely common. Experts estimate that more than one-third of Americans have some degree of arthritis in their joints.
Osteoarthritis is the most common type. Studies have found that around half of all adults will develop osteoarthritis at some point.
Symptoms and Causes
There are more than 100 types of arthritis, but they share several common signs and symptoms.
The most common signs and symptoms of arthritis usually affect your joints and your ability to use them.
What are arthritis symptoms and signs?
The most common arthritis symptoms and signs include:
* Joint pain.
* Stiffness or reduced range of motion (how far you can move a joint).
* Swelling (inflammation).
* Skin discoloration.
* Tenderness or sensitivity to touch around a joint.
* A feeling of heat or warmth near your joints.
Where you experience symptoms depends on which type of arthritis you have, and which of your joints it affects.
Some types of arthritis cause symptoms in waves that come and go called flares or flare-ups. Others make your joints feel painful or stiff all the time, or after being physically active.
What is the main cause of arthritis?
What causes arthritis varies depending on which type you have:
* Osteoarthritis happens naturally as you age — a lifetime of using your joints can eventually wear down their cartilage cushioning.
* You may develop gout if you have too much uric acid in your blood (hyperuricemia).
* Your immune system can cause arthritis (including rheumatoid arthritis) when it damages your joints by mistake.
* Certain viral infections (including COVID-19) can trigger viral arthritis.
* Sometimes, arthritis happens with no cause or trigger. Providers call this idiopathic arthritis.
What are the risk factors?
Anyone can develop arthritis, but some factors may make you more likely to, including:
* Tobacco use: Smoking and using other tobacco products increases your risk.
* Family history: People whose biological family members have arthritis are more likely to develop it.
* Activity level: You might be more likely to have arthritis if you aren’t physically active regularly.
* Other health conditions: Having autoimmune diseases, obesity or any condition that affects your joints increases the chances you’ll develop arthritis.
Some people have a higher arthritis risk, including:
* People older than 50.
* Females.
* Athletes, especially those who play contact sports.
* People who have physically demanding jobs or do work that puts a lot of stress on their joints (standing, crouching, being on your hands and knees for a long time, etc.).
At what age does arthritis usually start?
Arthritis can develop at any age. When it starts depends on which type you have and what’s causing it.
In general, osteoarthritis affects adults older than 50. Rheumatoid arthritis usually develops in adults age 30 to 60.
Other types that have a more direct cause usually start closer to that specific trigger. For example, people with post-traumatic arthritis don’t develop it until after their joints are injured, and gout doesn’t develop until after you’ve had high uric acid levels for at least several months.
Talk to a healthcare provider about your unique arthritis risk, and when you should start watching for signs or changes in your joints.
Diagnosis and Tests:
How do healthcare providers diagnose arthritis?
A healthcare provider will diagnose arthritis with a physical exam. They’ll examine your affected joints and ask about your symptoms. Tell your provider when you first noticed symptoms like pain and stiffness, and if any activities or times of day make them worse.
Your provider will probably check your range of motion (how far you can move a joint). They may compare one joint’s range of motion to other, similar joints (your other knee, ankle or fingers, for example).
Arthritis tests
Your provider might use imaging tests to take pictures of your joints, including:
* X-ray.
* Ultrasound.
* Magnetic resonance imaging (MRI).
* A computed tomography (CT) scan.
These tests can help your provider see damage inside your joints. They can also help your provider rule out other injuries or issues that might cause similar symptoms, like bone fractures (broken bones).
Your provider may use blood tests to check your uric acid levels if they think you have gout. Blood tests can also show signs of infections or autoimmune diseases.
Management and Treatment:
What is arthritis treatment?
There’s no cure for arthritis, but your healthcare provider will help you find treatments that manage your symptoms. Which treatments you’ll need depend on what’s causing the arthritis, which type you have and which joints it affects.
The most common arthritis treatments include:
* Over-the-counter (OTC) anti-inflammatory medicine like NSAIDs or acetaminophen.
* Corticosteroids (prescription anti-inflammatory medicine, including cortisone shots).
* Disease-modifying antirheumatic drugs (DMARDs) if you have rheumatoid or psoriatic arthritis.
* Physical therapy or occupational therapy can help you improve your strength, range of motion and confidence while you’re moving.
* Surgery (usually only if nonsurgical treatments don’t relieve your symptoms).
Arthritis surgery
You may need surgery if you have severe arthritis and other treatments don’t work. The two most common types of arthritis surgery are joint fusion and joint replacement.
Joint fusion is exactly what it sounds like: surgically joining bones together. It’s most common for bones in your spine (spinal fusion) or your ankle (ankle fusion).
If your joints are damaged or you’ve experienced bone loss, you might need an arthroplasty (joint replacement). Your surgeon will remove your damaged natural joint and replace it with a prosthesis (artificial joint). You might need a partial or total joint replacement.
Your provider or surgeon will tell you which type of surgery you’ll need and what to expect.
Outlook / Prognosis:
What can I expect if I have arthritis?
You should expect to manage arthritis symptoms for a long time (probably the rest of your life). Your provider will help you find treatments that reduce how much (and how often) arthritis impacts your daily routine.
Some people with arthritis experience more severe symptoms as they age. Ask your provider how often you should have follow-up visits to check for changes in your joints.

Hi,
#10789. What does the term in Biology Ichthyology mean?
#10790. What does the term in Biology Immune response mean?
Hi,
#5995. What does the noun newscast mean?
#5996. What does the noun newsreel mean?
Hi,
#2594. What does the medical term Levator scapulae muscle mean?
Hi,
#9880.
Hi,
#6373.
Hi,
2734.
2520) Fathometer
Gist
A fathometer is a specialized sonar-based instrument used to measure the depth of water (bathymetry) by calculating the time it takes for a sound wave to travel from the surface to the bottom and return as an echo. It is primarily used for navigation, mapping the seafloor, identifying fish, and checking ice thickness.
What is a fathometer used for measuring?
A fathometer is a device that measures the depth of water by measuring the time it takes for a sound wave to travel from the surface to the bottom and for its echo to be returned.
Summary
“A fathometer is a device that measures the depth of water by measuring the time it takes for a sound wave to travel from the surface to the bottom and for its echo to be returned.”
“Echo sounding or depth sounding is the use of sonar for ranging, normally to determine the depth of water (bathymetry). It involves transmitting acoustic waves into water and recording the time interval between emission and return of a pulse; the resulting time of flight, along with knowledge of the speed of sound in water, allows determining the distance between sonar and target. This information is then typically used for navigation purposes or in order to obtain depths for charting purposes.”
Had the fathometer, or echo sounder, been available, it might have changed maritime warfare during the First World War. The Fathometer, first offered for sale in 1923, uses sound waves to quickly and accurately determine water depth and detect underwater objects, like submarines. Before echo sounders, sailors dropped and retrieved weighted hand-held lines (lead lines) to estimate water depth. After the sinking of Titanic in 1912, Canadian engineer Reginald A. Fessenden wanted to improve underwater object detection and communications. An expert in wireless radio technology, Fessenden built an electromagnetic device, called the oscillator, which produced underwater sound waves. These waves bounced off objects and, by measuring the echoes produced by the returning waves, users could estimate the position of the object. Tests in 1914 proved the oscillator could not only detect icebergs, but also determine water depth. Fessenden urged war planners to use the oscillator as a submarine detector, but with his prickly personality, he was rebuffed. After the war, the Submarine Signal Company of Boston incorporated Fessenden’s oscillator technology in the fathometer, opening a new era of navigation, subsurface detection, and marine mapping.
Herbert Grove Dorsey (April 24, 1876 – 1961) was an American engineer, inventor and physicist. He was principal engineer of the United States Coast and Geodetic Survey Radio sonic Laboratory in the 1930s. He invented the first practical fathometer, a water depth measuring instrument for ships.
The depth of the ocean is calculated by knowing how fast sound travels in the water (approximately 1,500 meters per second). This method of seafloor mapping is called echo sounding. … Water depth is typically measured by echo sounders that transmit sound at 12 kilohertz (kHz).
Depth finder, also called echo sounder, device used on ships to determine the depth of water by measuring the time it takes a sound (sonic pulse) produced just below the water surface to return, or echo, from the bottom of the body of water. … sonar devices are used to measure the depth of sea.
Details
A fathometer is a device that measures the depth of water by measuring the time it takes for a sound wave to travel from the surface to the bottom and for its echo to be returned.
A Fathometer is used in ocean sounding when the depth of water is considerably deep and keeps a continuous and precise record of the depth of water under the boat or ship on which it is placed. It is an echo-sounding device in which water depths are determined by measuring the time it takes for vibrations produced by sound waves to travel from a location near the top of the water to the bottom and back. Depending on the kind of water being utilised, it is calibrated to read depth in line with the velocity of sound in that water. A fathometer may either visually represent the depth of the water or graphically indicate the depth of the water on a roll that is continually rotating and can produce a virtual profile of that water body. The fathometer meaning is derived from the word fathom. Fathom is a unit of water depth. Herbert Grove Dorsey invented and patented the first practical fathometer in 1928.
Working of fathometer
Fathometer working is quite simple and accurate compared to older methods of distance measuring of water bodies like the lead lines. A fathometer contains the following parts:
1. Transmitting and receiving oscillators ( for sending and receiving the sound waves )
2. Recorder unit ( for the recording of data)
3. Transmitter / Power unit. (power supply)
The distance between the signal’s leaving pulse and its return is calculated by multiplying half of the time between the signal’s outgoing pulse and its return by the sound speed in the water, roughly 1.5 kilometres per second. When using echo-sounding for precise applications such as hydrography, it is necessary to measure the sound speed, which is commonly accomplished by submerging a sound velocity probe into the water. Echo sounding is a special-purpose use of sonar that is used to find the bottom of any water bodies.
The Fessenden Fathometer was one of the earliest commercial echo- machines, and it made use of the Fessenden oscillator to create sound waves. Submarine Signal Company fitted this for the first time aboard the M&M liner S.S. Berkshire in 1924.
The fathometer is more accurate because it obtains a sounding that is exactly vertical. The vessel’s speed causes it to diverge significantly from the vertical. The precision of 7.5 cm is possible in ports and harbours when the water is at normal levels and conditions. When there is a strong current, and the weather is not conducive to taking soundings with the lead line, a fathometer may be employed (an Old device to measure the depth of water). The fathometer has a higher sensitivity than the lead line.
Uses of fathometer
1. Fathometer echo sounding is a technique that is often employed in fishing. Variations in elevation are often associated with areas where fish gather. In addition, schools of fish will be recorded. Fishfinder is an echo-sounding instrument, similar to a fathometer, that is used by both recreational and commercial fishermen to locate fish and other marine life.
2. This fathometer is installed on almost all ships and submarines in order to get an idea of the depth of water bodies and the morphology of rocks and seabeds in the area surrounding the ships and submarines.
3. It may also be used to measure the rise and fall of the tides in areas where the water is shallow.
4. The lead line is also one of the techniques for measuring the depth of water bodies, but it takes time and cannot be used in bad weather, so it has limited applications. The fathometer, on the other hand, can be used in bad weather and provides a more accurate result in a shorter period of time, making it more useful.
5. The same technique as the fathometer is used to send out sonic pulses in order to identify underwater things.
6. When it comes to submarines, a fathometer is quite important. Along with shoal water protection, other peacetime applications include identifying fish, assessing the thickness of ice in Arctic areas, and mapping the ocean’s surface.
7. A fathometer, also known as a Sonic depth finder, may be used to create a profile of the ocean bottom by recording thousands of soundings each hour over a long period of time. The use of echo sounders in oceanography and survey work to locate underwater pinnacles and shoals is common practice among hydrographers.
Conclusion
A fathometer is a device that uses echo sounding to measure the depth of the ocean or any other water bodies. Fathom is the unit for measuring the depth of any water body. A fathometer is more useful compared to conventional instruments to measure the depth because it can be used in bad weather and is more accurate with only an error of ± 7 cm. A fathometer is also used to locate the iceberg below the sea surface and schools of fish.
Additional Information
A Fathometer is used in ocean sounding where the depth of water is too much, and to make a continuous and accurate record of the depth of water below the boat or ship at which it is installed. It is an echo-sounding instrument in which water depths are obtained be determining the time required for the sound waves to travel from a point near the surface of the water to the bottom and back. It is adjusted to read depth on accordance with the velocity of sound in the type of water in which it is being used. A fathometer may indicate the depth visually or indicate graphically on a roll which continuously goes on revolving and provide a
virtual profile of the lake or sea.
What are the components of echo sounding instrument?
The main parts of an echo-sounding apparatus are:
1. Transmitting and receiving oscillators.
2. Recorder unit.
3. Transmitter / Power unit.
It consists in recording the interval of time between the emission of a sound impulse direct to the bottom of the sea and the reception of the wave or echo, reflected from the bottom. If the speed of sound in that water is v and the time interval between the transmitter and receiver is t, the depth h is given by
h = ½ vt
…
Due to the small distance between the receiver and the transmitter, a slight correction is necessary in shallow waters. The error between the true depth and the recorded depth can be calculated very easily by simple geometry. If the error is plotted against the recorded depth, the true depth can be easily known. The recording of the sounding is produced by the action of a small current passing through chemically impregnated paper from a rotating stylus to an anode plate. The stylus is fixed at one end of a radial arm which revolves at constant speed. The stylus makes a record on the paper at the instants when the sound impulse is transmitted and when the echo returns to the receiver.
Advantage of echo-sounding
Echo-sounding has the following advantages over the older method of lead line and rod:
1. It is more accurate as a truly vertical sounding is obtained. The speed of the vessel does deviate it appreciably from the vertical. Under normal water conditions, in ports and harbors an accuracy of 7.5 cm may be obtained.
2. It can be used when a strong current is running and when the weather is unsuitable for the soundings to be taken with the lead line.
3. It is more sensitive than the lead line.
4. A record of the depth is plotted immediately and provides a continuous record of the bottom as the vessel moves forward.
5. The speed of sounding and plotting is increased.
6. The error due to estimation of water level in a choppy sea is reduced owing to the instability of the boat.
7. Rock underlying softer material is recorded and this valuable information is obtained more cheaply than would be the case where sub-marine borings are taken.
Making the soundings
If the depth is less than 25 m, the soundings can be taken when the boat is in motion. In the case of soundings with rod the leadsman stands in the bow and plunges the rod at a forward angle, depending on the speed o the boat, such that the rod is vertical when the boat reaches the point at which soundings is being recorded. The rod should be read very quickly. The nature of the bottom should also be recorded at intervals in the note-book.
If the sounding is taken with a lead, the leadsman stands in the bow of the boat and casts the lead forward at such a distances that the line will become vertical and will reach the bottom at a point where sounding is required. The lead is withdrawn from the water after the reading is taken. If the depth is great, the lead is not withdrawn from the water, but is lifted between the soundings.
