History of numerical control

The history of numerical control (NC) began when the automation of machine tools first incorporated concepts of abstractly programmable logic, and it continues today with the ongoing evolution of computer numerical control (CNC) technology.

The first NC machines were built in the 1940s and 1950s, based on existing tools that were modified with motors that moved the controls to follow points fed into the system on punched tape. These early servomechanisms were rapidly augmented with analog and digital computers, creating the modern CNC machine tools that have revolutionized the machining processes.

Earlier forms of automation

Cams

Siemens CNC panel.

The automation of machine tool control began in the 19th century with cams that "played" a machine tool in the way that cams had long been playing musical boxes or operating elaborate cuckoo clocks. Thomas Blanchard built his gun-stock-copying lathes (1820s–30s), and the work of people such as Christopher Miner Spencer developed the turret lathe into the screw machine (1870s). Cam-based automation had already reached a highly advanced state by World War I (1910s).

However, automation via cams is fundamentally different from numerical control because it cannot be abstractly programmed. Cams can encode information, but getting the information from the abstract level (engineering drawing, CAD model, or other design intent) into the cam is a manual process that requires machining or filing. In contrast, numerical control allows information to be transferred from design intent to machine control using abstractions such as numbers and programming languages.

Various forms of abstractly programmable control had existed during the 19th century: those of the Jacquard loom, player pianos, and mechanical computers pioneered by Charles Babbage and others. These developments had the potential for convergence with the automation of machine tool control starting in that century, but the convergence did not happen until many decades later.

Tracer control

The application of hydraulics to cam-based automation resulted in tracing machines that used a stylus to trace a template, such as the enormous Pratt & Whitney "Keller Machine", which could copy templates several feet across.[1] Another approach was "record and playback", pioneered at General Motors (GM) in the 1950s, which used a storage system to record the movements of a human machinist, and then play them back on demand. Analogous systems are common even today, notably the "teaching lathe" which gives new machinists a hands-on feel for the process. None of these were numerically programmable, however, and required an experienced machinist at some point in the process, because the "programming" was physical rather than numerical.

Servos and selsyns

One barrier to complete automation was the required tolerances of the machining process, which are routinely on the order of thousandths of an inch. Although connecting some sort of control to a storage device like punched cards was easy, ensuring that the controls were moved to the correct position with the required accuracy was another issue. The movement of the tool resulted in varying forces on the controls that would mean a linear input would not result in linear tool motion. In other words, a control such as that of the Jacquard loom could not work on machine tools because its movements were not strong enough; the metal being cut "fought back" against it with more force than the control could properly counteract.

The key development in this area was the introduction of the servomechanism, which produced powerful, controlled movement, with highly accurate measurement information. Attaching two servos together produced a selsyn, where a remote servo's motions were accurately matched by another. Using a variety of mechanical or electrical systems, the output of the selsyns could be read to ensure proper movement had occurred (in other words, forming a closed-loop control system).

The first serious suggestion that selsyns could be used for machining control was made by Ernst F. W. Alexanderson, a Swedish immigrant to the U.S. working at General Electric (GE). Alexanderson had worked on the problem of torque amplification that allowed the small output of a mechanical computer to drive very large motors, which GE used as part of a larger gun laying system for US Navy ships. Like machining, gun laying requires very high accuracy – fractions of a degree – and the forces during the motion of the gun turrets was non-linear. In November 1931 Alexanderson suggested to the Industrial Engineering Department that the same systems could be used to drive the inputs of machine tools, allowing it to follow the outline of a template without the strong physical contact needed by existing tools like the Keller Machine. He stated that it was a "matter of straight engineering development".[2] However, the concept was ahead of its time from a business development perspective, and GE did not take the matter seriously until years later, when others had pioneered the field.

Parsons and Sikorsky

The birth of NC is generally credited to John T. Parsons,[3] a machinist and salesman at his father's machining company, Parsons Corp.

In 1942 he was told that helicopters were going to be the "next big thing" by the former head of Ford Trimotor production, Bill Stout. He called Sikorsky Aircraft to inquire about possible work, and soon got a contract to build the wooden stringers in the rotor blades. At the time, rotor blades (rotary wings) were built in the same fashion that fixed wings were, consisting of a long tubular steel spar with stringers (or more accurately ribs) set on them to provide the aerodynamic shape that was then covered with a stressed skin. The stringers for the rotors were built from a design provided by Sikorsky, which was sent to Parsons as a series of 17 points defining the outline. Parsons then had to "fill in" the dots with a French curve to generate an outline. A wooden jig was built up to form the outside of the outline, and the pieces of wood forming the stringer were placed under pressure against the inside of the jig so they formed the proper curve. A series of trusswork members were then assembled inside this outline to provide strength.[4]

After setting up production at a disused furniture factory and ramping up production, one of the blades failed and it was traced to a problem in the spar. At least some of the problem appeared to stem from spot welding a metal collar on the stringer to the metal spar. The collar was built into the stringer during construction, then slid onto the spar and welded in the proper position. Parsons suggested a new method of attaching the stringers directly to the spar using adhesives, never before tried on an aircraft design.[4]

That development led Parsons to consider the possibility of using stamped metal stringers instead of wood. These would not only be much stronger, but far easier to make as well, as they would eliminate the complex layup and glue and screw fastening on the wood. Duplicating this in a metal punch would require the wooden jig to be replaced by a metal cutting tool made of tool steel. Such a device would not be easy to produce given the complex outline. Looking for ideas, Parsons visited Wright Field to see Frank Stulen, the head of the Propeller Lab Rotary Wing Branch. During their conversation, Stulen concluded that Parsons didn't really know what he was talking about. Parsons realized Stulen had reached this conclusion, and hired him on the spot. Stulen started work on 1 April 1946 and hired three new engineers to join him.[4]

Stulen's brother worked at Curtis Wright Propeller, and mentioned that they were using punched card calculators for engineering calculations. Stulen decided to adopt the idea to run stress calculations on the rotors, the first detailed automated calculations on helicopter rotors.[4] When Parsons saw what Stulen was doing with the punched card machines, he asked Stulen if they could be used to generate an outline with 200 points instead of the 17 they were given, and offset each point by the radius of a mill cutting tool. If you cut at each of those points, it would produce a relatively accurate cutout of the stringer. This could cut the tool steel and then easily be filed down to a smooth template for stamping metal stringers.[4]

Stullen had no problem making such a program, and used it to produce large tables of numbers that would be taken onto the machine floor. Here, one operator read the numbers off the charts to two other operators, one on each of the X- and Y- axes. For each pair of numbers the operators would move the cutting head to the indicated spot and then lower the tool to make the cut.[4] This was called the "by-the-numbers method", or more technically, "plunge-cutting positioning".[5] It was a labor-intensive prototype of today's 2.5 axis machining (two-and-a-half-axis machining).

Punch cards and first tries at NC

At that point Parsons conceived of a fully automated machine tool. With enough points on the outline, no manual working would be needed to clean it up. However, with manual operation the time saved by having the part more closely match the outline was offset by the time needed to move the controls. If the machine's inputs were attached directly to the card reader, this delay, and any associated manual errors, would be removed and the number of points could be dramatically increased. Such a machine could repeatedly punch out perfectly accurate templates on command. But at the time Parsons had no funds to develop his ideas.

When one of Parsons's salesmen was on a visit to Wright Field, he was told of the problems the newly formed US Air Force was having with new jet-powered designs. He asked if Parsons had anything to help them. Parsons showed Lockheed their idea of an automated mill, but they were uninterested. They decided to use 5-axis template copiers to produce the stringers, cutting from a metal template, and had already ordered the expensive cutting machine. But as Parsons noted:

Now just picture the situation for a minute. Lockheed had contracted to design a machine to make these wings. This machine had five axes of cutter movement, and each of these was tracer controlled using a template. Nobody was using my method of making templates, so just imagine what chance they were going to have of making an accurate airfoil shape with inaccurate templates.[4]

Parson's worries soon came true, and Lockheed's protests that they could fix the problem eventually rang hollow. In 1949 the Air Force arranged funding for Parsons to build his machines on his own.[4] Early work with Snyder Machine & Tool Corp proved the system of directly driving the controls from motors failed to give the accuracy needed to set the machine for a perfectly smooth cut. Since the mechanical controls did not respond in a linear fashion, you couldn't simply drive it with a given amount of power, because the differing forces meant the same amount of power would not always produce the same amount of motion in the controls. No matter how many points you included, the outline would still be rough. Parsons was confronted by the same problem that had prevented convergence of Jacquard-type controls with machining.

First commercial numerically controlled machine

In 1952, Arma Corporation which had done much defense work on rangefinders during the war, announced the first commercial numerically controlled lathe, developed by Dr. F. W. Cunningham. Arma's first automated lathe was made in 1948, and announced in 1950.[6][7][8]

Enter MIT

This was not an impossible problem to solve, but would require some sort of feedback system, like a selsyn, to directly measure how far the controls had actually turned. Faced with the daunting task of building such a system, in the spring of 1949 Parsons turned to Gordon S. Brown's Servomechanisms Laboratory at MIT, which was a world leader in mechanical computing and feedback systems.[9] During the war the Lab had built a number of complex motor-driven devices like the motorized gun turret systems for the Boeing B-29 Superfortress and the automatic tracking system for the SCR-584 radar. They were naturally suited to technological transfer into a prototype of Parsons's automated "by-the-numbers" machine.

The MIT team was led by William Pease assisted by James McDonough. They quickly concluded that Parsons's design could be greatly improved; if the machine did not simply cut at points A and B, but instead moved smoothly between the points, then not only would it make a perfectly smooth cut, but could do so with many fewer points – the mill could cut lines directly instead of having to define a large number of cutting points to "simulate" a line. A three-way agreement was arranged between Parsons, MIT, and the Air Force, and the project officially ran from July 1949 to June 1950.[10] The contract called for the construction of two "Card-a-matic Milling Machines", a prototype and a production system. Both to be handed to Parsons for attachment to one of their mills in order to develop a deliverable system for cutting stringers.

Instead, in 1950 MIT bought a surplus Cincinnati Milling Machine Company "Hydro-Tel" mill of their own and arranged a new contract directly with the Air Force that froze Parsons out of further development.[4] Parsons would later comment that he "never dreamed that anybody as reputable as MIT would deliberately go ahead and take over my project."[4] In spite of the development being handed to MIT, Parsons filed for a patent on "Motor Controlled Apparatus for Positioning Machine Tool" on 5 May 1952, sparking a filing by MIT for a "Numerical Control Servo-System" on 14 August 1952. Parsons received US Patent 2,820,187[11] on 14 January 1958, and the company sold an exclusive license to Bendix. IBM, Fujitsu and General Electric all took sub-licenses after having already started development of their own devices.

MIT's machine

MIT fitted gears to the various handwheel inputs and drove them with roller chains connected to motors, one for each of the machine's three axes (X, Y, and Z). The associated controller consisted of five refrigerator-sized cabinets that, together, were almost as large as the mill they were connected to. Three of the cabinets contained the motor controllers, one controller for each motor, the other two the digital reading system.[1]

Unlike Parsons's original punched card design, the MIT design used standard 7-track punch tape for input. Three of the tracks were used to control the different axes of the machine, while the other four encoded various control information.[1] The tape was read in a cabinet that also housed six relay-based hardware registers, two for each axis. With every read operation the previously read point was copied into the "starting point" register, and the newly read one into the "ending point" register.[1] The tape was read continually and the number in the registers incremented with each hole encountered in their control track until a "stop" instruction was encountered, four holes in a line.

The final cabinet held a clock that sent pulses through the registers, compared them, and generated output pulses that interpolated between the points. For instance, if the points were far apart the output would have pulses with every clock cycle, whereas closely spaced points would only generate pulses after multiple clock cycles. The pulses were sent into a summing register in the motor controllers, counting up by the number of pulses every time they were received. The summing registers were connected to a digital to analog converter that increased power to the motors as the count in the registers increased, making the controls move faster.[1]

The registers were decremented by encoders attached to the motors and the mill itself, which would reduce the count by one for every one degree of rotation. Once the second point was reached the counter would hold a zero, the pulses from the clock would stop, and the motors would stop turning. Each 1 degree rotation of the controls produced a 0.0005 inch movement of the cutting head. The programmer could control the speed of the cut by selecting points that were closer together for slow movements, or further apart for rapid ones.[1]

The system was publicly demonstrated in September 1952,[12] appearing in that month's Scientific American.[1] MIT's system was an outstanding success by any technical measure, quickly making any complex cut with extremely high accuracy that could not easily be duplicated by hand. However, the system was terribly complex, including 250 vacuum tubes, 175 relays and numerous moving parts, reducing its reliability in a production environment. It was also expensive; the total bill presented to the Air Force was $360,000.14 ($2,641,727.63 in 2005 dollars).[13] Between 1952 and 1956 the system was used to mill a number of one-off designs for various aviation firms, in order to study their potential economic impact.[14]

Proliferation of NC

The Air Force Numeric Control and Milling Machine projects formally concluded in 1953, but development continued at the Giddings and Lewis Machine Tool Co. and other locations. In 1955 many of the MIT team left to form Concord Controls, a commercial NC company with Giddings' backing, producing the Numericord controller.[14] Numericord was similar to the MIT design, but replaced the punch tape with a magnetic tape reader that General Electric was working on. The tape contained a number of signals of different phases, which directly encoded the angle of the various controls. The tape was played at a constant speed in the controller, which set its half of the selsyn to the encoded angles while the remote side was attached to the machine controls. Designs were still encoded on paper tape, but the tapes were transferred to a reader/writer that converted them into magnetic form. The magtapes could then be used on any of the machines on the floor, where the controllers were greatly reduced in complexity. Developed to produce highly accurate dies for an aircraft skinning press, the Numericord "NC5" went into operation at G&L's plant at Fond du Lac, WI in 1955.[15]

Monarch Machine Tool also developed a numerical controlled lathe, starting in 1952. They demonstrated their machine at the 1955 Chicago Machine Tool Show (predecessor of today's IMTS), along with a number of other vendors with punched card or paper tape machines that were either fully developed or in prototype form. These included Kearney & Trecker’s Milwaukee-Matic II that could change its cutting tool under numerical control,[15] a common feature on modern machines.

A Boeing report noted that "numerical control has proved it can reduce costs, reduce lead times, improve quality, reduce tooling and increase productivity.”[15] In spite of these developments, and glowing reviews from the few users, uptake of NC was relatively slow. As Parsons later noted:

The NC concept was so strange to manufacturers, and so slow to catch on, that the US Army itself finally had to build 120 NC machines and lease them to various manufacturers to begin popularizing its use.[4]

In 1958 MIT published its report on the economics of NC. They concluded that the tools were competitive with human operators, but simply moved the time from the machining to the creation of the tapes. In Forces of Production, Noble[16] claims that this was the whole point as far as the Air Force was concerned; moving the process off of the highly unionized factory floor and into the non-unionized white collar design office. The cultural context of the early 1950s, a second Red Scare with a widespread fear of a bomber gap and of domestic subversion, sheds light on this interpretation. It was strongly feared that the West would lose the defense production race to the Communists, and that syndicalist power was a path toward losing, either by "getting too soft" (less output, greater unit expense) or even by Communist sympathy and subversion within unions (arising from their common theme of empowering the working class).

Aside from what ever economic inefficiencies the first attempts at NC displayed, the time and effort required in the creation of the tapes also introduced possibilities for production errors. This would be a motivation for Air Force contracts ongoing in 1958 like the Automatically Programmed Tool project and the report, then later project, Computer-Aided Design: A Statement of Objectives 1960 of Douglas T. Ross.

CNC arrives

Many of the commands for the experimental parts were programmed "by hand" to produce the punch tapes that were used as input. During the development of Whirlwind, MIT's real-time computer, John Runyon coded a number of subroutines to produce these tapes under computer control. Users could enter a list of points and speeds, and the program would calculate the points needed and automatically generate the punch tape. In one instance, this process reduced the time required to produce the instruction list and mill the part from 8 hours to 15 minutes. This led to a proposal to the Air Force to produce a generalized "programming" language for numerical control, which was accepted in June 1956.[14] Doug Ross was given leadership of the project and was made head of another newly created MIT research department. He chose to name the unit the Computer Applications Group feeling the word "application" fit with the vision that general purpose machines could be "programmed" to fill many roles.[17]

Starting in September, Ross and Pople outlined a language for machine control that was based on points and lines, developing this over several years into the APT programming language.[18] In 1957 the Aircraft Industries Association (AIA) and Air Material Command at Wright-Patterson Air Force Base joined with MIT to standardize this work and produce a fully computer-controlled NC system. On 25 February 1959 the combined team held a press conference showing the results, including a 3D machined aluminum ash tray that was handed out in the press kit.[14][19][20]

Meanwhile, Patrick Hanratty was making similar developments at GE as part of their partnership with G&L on the Numericord. His language, PRONTO, beat APT into commercial use when it was released in 1958.[21] Hanratty then went on to develop MICR magnetic ink characters that were used in cheque processing, before moving to General Motors to work on the groundbreaking DAC-1 CAD system.


APT was soon extended to include "real" curves in 2D-APT-II. With its release into the Public Domain, MIT reduced its focus on NC as it moved into CAD experiments. APT development was picked up with the AIA in San Diego, and in 1962, by Illinois Institute of Technology Research. Work on making APT an international standard started in 1963 under USASI X3.4.7, but any manufacturers of NC machines were free to add their own one-off additions (like PRONTO), so standardization was not completed until 1968, when there were 25 optional add-ins to the basic system.[14]

Just as APT was being released in the early 1960s, a second generation of lower-cost transistorized computers was hitting the market that were able to process much larger volumes of information in production settings. This reduced the cost of programming for NC machines and by the mid-1960s, APT runs accounted for a third of all computer time at large aviation firms.

CADCAM meets CNC

CAD CNC example.

While the Servomechanisms Lab was in the process of developing their first mill, in 1953, MIT's Mechanical Engineering Department dropped the requirement that undergraduates take courses in drawing. The instructors formerly teaching these programs were merged into the Design Division, where an informal discussion of computerized design started. Meanwhile, the Electronic Systems Laboratory, the newly rechristened Servomechanisms Laboratory, had been discussing whether or not design would ever start with paper diagrams in the future.[22]

In January 1959, an informal meeting was held involving individuals from both the Electronic Systems Laboratory and the Mechanical Engineering Department's Design Division. Formal meetings followed in April and May, which resulted in the "Computer-Aided Design Project".[23] In December 1959, the Air Force issued a one-year contract to ESL for $223,000 to fund the project, including $20,800 earmarked for 104 hours of computer time at $200 per hour.[24] This proved to be far too little for the ambitious program they had in mind[24] In 1959 that was a lot of money. Newly graduated engineers were making perhaps $500 to $600 per month at the time. To augment the Air Force's commitment Ross replayed the success of the APT development model. The AED Cooperative Program which ultimately ran for a five-year period had outside corporate staff, deeply experienced design manpower on loan from companies . Some relocating to MIT for half a year to 14 or 18 months at a time. Ross later estimated this value at almost six million dollars in support of AED development work, systems research, compilers. AED was a machine independent software engineering job and an extension of ALGOL 60 the standard for the publication of algorithms by research computer scientists. Development started out in parallel on the IBM 709 and the TX-0 which later enabled projects to run at various sites. The engineering calculation and systems development system, AED, was released to the Public Domain in March 1965.

In 1959, General Motors started an experimental project to digitize, store and print the many design sketches being generated in the various GM design departments. When the basic concept demonstrated that it could work, they started the DAC-1 – Design Augmented by Computer – project with IBM to develop a production version. One part of the DAC project was the direct conversion of paper diagrams into 3D models, which were then converted into APT commands and cut on milling machines. In November 1963 a design for the lid of a trunk moved from 2D paper sketch to 3D clay prototype for the first time.[25] With the exception of the initial sketch, the design-to-production loop had been closed.

Meanwhile, MIT's offsite Lincoln Labs was building computers to test new transistorized designs. The ultimate goal was essentially a transistorized Whirlwind known as TX-2, but in order to test various circuit designs a smaller version known as TX-0 was built first. When construction of TX-2 started, time in TX-0 freed up and this led to a number of experiments involving interactive input and use of the machine's CRT display for graphics. Further development of these concepts led to Ivan Sutherland's groundbreaking Sketchpad program on the TX-2.

Sutherland moved to the University of Utah after his Sketchpad work, but it inspired other MIT graduates to attempt the first true CAD system. It was Electronic Drafting Machine (EDM), sold to Control Data and known as "Digigraphics", which Lockheed used to build production parts for the C-5 Galaxy, the first example of an end-to-end CAD/CNC production system.

By 1970 there were a wide variety of CAD firms including Intergraph, Applicon, Computervision, Auto-trol Technology, UGS Corp. and others, as well as large vendors like CDC and IBM.

Proliferation of CNC

Paper tape reader on a computer numerical control (CNC) machine.

The price of computer cycles fell drastically during the 1960s with the widespread introduction of useful minicomputers. Eventually it became less expensive to handle the motor control and feedback with a computer program than it was with dedicated servo systems. Small computers were dedicated to a single mill, placing the entire process in a small box. PDP-8's and Data General Nova computers were common in these roles. The introduction of the microprocessor in the 1970s further reduced the cost of implementation, and today almost all CNC machines use some form of microprocessor to handle all operations.

The introduction of lower-cost CNC machines radically changed the manufacturing industry. Curves are as easy to cut as straight lines, complex 3-D structures are relatively easy to produce, and the number of machining steps that required human action have been dramatically reduced. With the increased automation of manufacturing processes with CNC machining, considerable improvements in consistency and quality have been achieved with no strain on the operator. CNC automation reduced the frequency of errors and provided CNC operators with time to perform additional tasks. CNC automation also allows for more flexibility in the way parts are held in the manufacturing process and the time required changing the machine to produce different components.

During the early 1970s the Western economies were mired in slow economic growth and rising employment costs, and NC machines started to become more attractive. The major U.S. vendors were slow to respond to the demand for machines suitable for lower-cost NC systems, and into this void stepped the Germans. In 1979, sales of German machines surpassed the U.S. designs for the first time. This cycle quickly repeated itself, and by 1980 Japan had taken a leadership position, U.S. sales dropping all the time. Once sitting in the #1 position in terms of sales on a top-ten chart consisting entirely of U.S. companies in 1971, by 1987 Cincinnati Milacron was in 8th place on a chart heavily dominated by Japanese firms.[26]

Many researchers have commented that the U.S. focus on high-end applications left them in an uncompetitive situation when the economic downturn in the early 1970s led to greatly increased demand for low-cost NC systems. Unlike the U.S. companies, who had focused on the highly profitable aerospace market, German and Japanese manufacturers targeted lower-profit segments from the start and were able to enter the low-cost markets much more easily. Additionally large Japanese companies established their own subsidiaries or strengthened their machine divisions to produce the machines they needed. This was seen as a national effort and largely encouraged by MITI, the Japanese Ministry for International Trade and Industry. In the early years of the development, MITI provided focused resources for the transfer of technological know-how.[26][27] National efforts in the US were focused on integrated manufacturing from the historical perspective the defence sector maintained. This evolved in the later 1980`s, as the so-called machine tool crisis was recognized, into a number of programs that sought to broaden transfer of know how to domestic tool makers. The Air Force sponsored Next Generation Controller Program 1989 as an example. This process continued through the 1990`s to the present day from DARPA incubators and myriad research grants.

As computing and networking evolved, so did direct numerical control (DNC). Its long-term coexistence with less networked variants of NC and CNC is explained by the fact that individual firms tend to stick with whatever is profitable, and their time and money for trying out alternatives is limited. This explains why machine tool models and tape storage media persist in grandfathered fashion even as the state of the art advances.

DIY, hobby, and personal CNC

Recent developments in small scale CNC have been enabled, in large part, by the Enhanced Machine Controller project in 1989 from the National Institute of Standards and Technology (NIST), an agency of the US Government's Department of Commerce. EMC [LinuxCNC] is a public domain program operating under the Linux operating system and working on PC based hardware. After the NIST project ended, development continued, leading to LinuxCNC[28] which is licensed under the GNU General Public License and Lesser GNU General Public License (GPL and LGPL). Derivations of the original EMC software have also led to several proprietary low cost PC based programs notably TurboCNC, and Mach3, as well as embedded systems based on proprietary hardware. The availability of these PC based control programs has led to the development of DIY CNC, allowing hobbyists to build their own[29][30] using open source hardware designs. The same basic architecture has allowed manufacturers, such as Sherline and Taig, to produce turnkey lightweight desktop milling machines for hobbyists.

The easy availability of PC based software and support information of Mach3, written by Art Fenerty, lets anyone with some time and technical expertise make complex parts for home and prototype use. Fenerty is considered a principal founder of Windows-based PC CNC machining.[31]

Eventually, the homebrew architecture was fully commercialized and used to create larger machinery suitable for commercial and industrial applications. This class of equipment has been referred to as Personal CNC. Parallel to the evolution of personal computers, Personal CNC has its roots in EMC and PC based control, but has evolved to the point where it can replace larger conventional equipment in many instances. As with the Personal Computer, Personal CNC is characterized by equipment whose size, capabilities, and original sales price make it useful for individuals, and which is intended to be operated directly by an end user, often without professional training in CNC technology.

Today

Tape readers may still be found on current CNC facilities, since machine tools have a long operating life. Other methods of transferring CNC programs to machine tools, such as diskettes or direct connection of a portable computer, are also used. Punched mylar tapes are more robust. Floppy disks, USB flash drives and local area networking have replaced the tapes to some degree, especially in larger environments that are highly integrated.

The proliferation of CNC led to the need for new CNC standards that were not encumbered by licensing or particular design concepts, like proprietary extensions to APT.[18] A number of different "standards" proliferated for a time, often based around vector graphics markup languages supported by plotters. One such standard has since become very common, the "G-code" that was originally used on Gerber Scientific plotters and then adapted for CNC use. The file format became so widely used that it has been embodied in an EIA standard. In turn, while G-code is the predominant language used by CNC machines today, there is a push to supplant it with STEP-NC, a system that was deliberately designed for CNC, rather than grown from an existing plotter standard.

While G-code is the most common method of programming, some machine-tool/control manufacturers also have invented their own proprietary "conversational" methods of programming, trying to make it easier to program simple parts and make set-up and modifications at the machine easier (such as Mazak's Mazatrol, Okuma's IGF, and Hurco). These have met with varying success.

A more recent advancement in CNC interpreters is support of logical commands, known as parametric programming (also known as macro programming). Parametric programs include both device commands as well as a control language similar to BASIC. The programmer can make if/then/else statements, loops, subprogram calls, perform various arithmetic, and manipulate variables to create a large degree of freedom within one program. An entire product line of different sizes can be programmed using logic and simple math to create and scale an entire range of parts, or create a stock part that can be scaled to any size a customer demands.

Since about 2006, the idea has been suggested and pursued to foster the convergence with CNC and DNC of several trends elsewhere in the world of information technology that have not yet much affected CNC and DNC. One of these trends is the combination of greater data collection (more sensors), greater and more automated data exchange (via building new, open industry-standard XML schemas), and data mining to yield a new level of business intelligence and workflow automation in manufacturing. Another of these trends is the emergence of widely published APIs together with the aforementioned open data standards to encourage an ecosystem of user-generated apps and mashups, which can be both open and commercial – in other words, taking the new IT culture of app marketplaces that began in web development and smartphone app development and spreading it to CNC, DNC, and the other factory automation systems that are networked with the CNC/DNC. MTConnect is a leading effort to bring these ideas into successful implementation.

See also

References

  1. 1 2 3 4 5 6 7 Pease, William (1952), "An automatic machine tool", Scientific American, 187 (3): 101–115, ISSN 0036-8733, doi:10.1038/scientificamerican0952-101.
  2. Brittain 1992, pp. 210–211.
  3. The International Biographical Dictionary of Computer Pioneers refers to Parsons as "the father of computerized milling machines", and the Society of Manufacturing Engineers awarded him a citation for "conceptualization of numerical control marked the beginning of the second industrial revolution."
  4. 1 2 3 4 5 6 7 8 9 10 11 "The Father of the Second Industrial Revolution", Manufacturing Engineering, 127 (2), August 2001
  5. "Numerically Controlled Milling Machine"
  6. Fertigungsautomatisierung: Automatisierungsmittel, Gestaltung und Funktion Stefan Hesse; Springer-Verlag, 2013 ; page 54
  7. Electricity in the American Economy: Agent of Technological Progress Sam H. Schurr; 1990; page 66
  8. NEW DEVICE SHOWN FOR MASS OUTPUT; Arma Corp., Bosch Subsidiary, Demonstrates 'Piano Roll' Metal Fabricating Tool PATTERN FROM BLUEPRINT Provides Precision Production Control on Standard Lathe With Little Human Aid New York Times; HARTLEY W. BARCLAY; June 30, 195 0
  9. Reintjes 1991, p. 16.
  10. Wildes & Lindgren 1985, p. 220.
  11. Link to google patents http://www.google.com/patents?id=rRpqAAAAEBAJ&dq=2820187
  12. Susskind, Alfred K.; McDonough, James O. (March 1953). "Numerically Controlled Milling Machine" (PDF). Review of Input and Output Equipment Used in Computing Systems. International Workshop on Managing Requirements Knowledge. New York City: American Institute of Electrical Engineers. pp. 133–137. LCCN 53-7874. Retrieved 2015-02-24.
  13. New Technology, pg. 47
  14. 1 2 3 4 5 Ross, Douglas T. (August 1978), "Origins of the APT language for automatically programmed tools" (PDF), ACM SIGPLAN Notices, 13 (8): 61–99, doi:10.1145/960118.808374, archived from the original (PDF) on 2010-03-09.
  15. 1 2 3 Makely, William (August 2005), "Numbers Take Control: NC Machines" (PDF), Cutting Tool Engineering, 57 (8): 4–5, archived from the original (PDF) on 2010-03-09.
  16. Noble 1984.
  17. ROSS O'Neill CBI Oral History http://conservancy.umn.edu/bitstream/107611/1/oh178dtr.pdf
  18. 1 2 Aptos free open source APT http://aptos.sourceforge.net/
  19. Milled Ashtray "Archived copy". Archived from the original on 2013-11-13. Retrieved 2013-05-11.
  20. https://archive.org/details/bitsavers_mitwhirlwie1GeneralDescriptionoftheAPTSystem1959_5339230
  21. "The CAD/CAM Hall of Fame: Patrick J. Hanratty", American Machinist
  22. Weisberg, pp. 3–9.
  23. Computer-Aided Design Project http://images.designworldonline.com.s3.amazonaws.com/CADhistory/8436-TM-4.pdf
  24. 1 2 Weisberg, pp. 3–10.
  25. Krull, F.N. (September 1994), "The origin of computer graphics within General Motors", IEEE Annals of the History of Computing, 16 (3): 40–56, ISSN 1058-6180, doi:10.1109/MAHC.1994.298419.
  26. 1 2 Arnold, Heinrich Martin (November 2001), "The recent history of the machine tool industry and the effects of technological change", LMU, CiteSeerX 10.1.1.119.2125Freely accessible
  27. Holland 1989.
  28. linuxcnc.org LinuxCNC
  29. Home Made CNC Machine. Hacked Gadgets – DIY Tech Blog.
  30. Desktop Manufacturing. Make (magazine) Vol 21, Feb, 2010.
  31. CNCzone discussion of Fenerty

Cited sources

Further reading

This article is issued from Wikipedia. The text is licensed under Creative Commons - Attribution - Sharealike. Additional terms may apply for the media files.