Ewa Wisnerska, paraglider, got caught in one of Australia’s notorious thunderstorm updrafts, spiraled up to 32,000 ft, was pelted with hail, encased in ice, lost consciousness from the low oxygen, and returned safely to Earth fifty miles from where she started. Her trip was logged by her GPS device. She intends to complete the competition she’s attending.
From the Air Force Materiel Command News Service:
Engineers here are testing a new kind of transparent armor – stronger and lighter than traditional materials – that could stop armor-piercing weapons from penetrating vehicle windows. The Air Force Research Laboratory’s materials and manufacturing directorate is testing aluminum oxynitride – ALONtm – as a replacement for the traditional multi-layered glass transparencies now used in existing ground and air armored vehicles.
…ALONtm is a ceramic compound with a high compressive strength and durability… “The substance itself is light years ahead of glass,” said 1st Lt. Joseph La Monica, adding that it offers “higher performance and lighter weight.” ..ALONtm is virtually scratch resistant, offers substantial impact resistance, and provides better durability and protection against armor piercing threats, at roughly half the weight and half the thickness of traditional glass transparent armor, said the lieutenant.
…”With glass, to get the protection against higher threats, you have to keep building layers upon layers. But with ALONtm, the material only needs to be increased a few millimeters.” …”Eventually, with a conventional glass surface, degradation takes place and results in a loss of transparency,” Ron Hoffman, an investigator at University of Dayton Research Institute, said. “Things such as sand have little or no impact on ALONtm, and it probably has a life expectancy many times that of glass.”
…”Traditional transparent armor costs a little over $3 per square inch. The ALONtm Transparent Armor cost is $10 to $15 per square inch,” Lieutenant La Monica said. “The difficulties arise with heating and polishing processes, which lead to higher costs. But we are looking at more cost effective alternatives.”
Make her teach! The Valley News has the story of Dartmouth Professor Mara Sabinson who “alleges that after she refused to quit, college administrators punished her by assigning her to teach freshman writing seminars.” When I toured Dartmouth, their big selling point was that the professors are really engaged in teaching, not “I’ll teach the postdocs and have grad students teach the undergrads,” so this is exactly the wrong kind of press for the Marketing Department.
The attitude that teaching freshmen is punishment is hardly hidden by these professors, and the students aren’t so dumb as to not understand what’s going on. Those who would take a professorship and not love teaching should go out and get a job in industry. Those who would eschew the industry route in favor of the safety of the tenure track while still despising teaching have no place in higher education, and are a threat to the competitiveness of the U.S. .
This weekend I attended a seminar at Dartmouth College commemorating the fifthieth anniversary of the first Dartmouth Summer Symposium on Artificial Intelligence, where the term was coined. It was a great seminar and it really made me think.
The first speaker was George Cybenko, Professor of Engineering at Thayer. His work is in the current application of Artificial Intelligence, and he had lots of great slides and demos of the state of AI, the autonomous military vehicle project, and the trend towards computing density. According to current rates and Moore’s Law, within 30 years a microchip will have the connection complexity of the human min. Today’s Itanium 2 chip is as complex as a honeybee’s brain.
There has been some theory in the literature about how animal brains may operate on the quantum rather than digital level, with microtubules within cells storing quantum superpositions of information. I asked Professor Cybenko about the current state of this research and it’s still an open question. If it’s true, simply making an Itanium 10 in 2030 won’t approach the complexity of the human brain - we’ll need quantum computing to do that. He gave the great example of an art expert being able to look at a vase and say, instantly, “that’s clearly a fake, but I’m not sure why,” which then takes weeks of experts’ time to prove empirically. To me, that bolsters the argument for our brain storing quantum superpositions.
The most surprising speaker for me was James Moor, Chair of the Philosophy Department. I naively had low expectations for the talk, what with trouble getting PowerPoint up and all, but that was a misleading book cover. His talk covered many aspects of how we should deal with an AI, but the most striking part of his talk was AI Ethics. Right now we’re at the point where we’d like to think about how we should teach ethics to AI’s, if they’re ever to become autonomous. So, the next thirty years may be just that kind of mundane part of how we code ethics into a machine-parseable framework. Where things get interesting is what happens after we achieve it.
I asked the question during the QA session, “Is Ethics Static?” to which Professor Moor gave a good response that ethics by its nature has to evolve. So, I asked the followup question, “so once AI’s learn ethics, and in thirty years they may be as smart as we are, isn’t it natural that they’d decide to continue to evolve the ethics we taught them, perhaps not applying the same values as we would?” This led to a discussion of the three models of the AI future - AI’s as slaves, AI’s conquering humans, and humans merging with AI’s. “What if learning Polish was as simple as plugging in a new ‘chip’? - What would that mean for Dartmouth 300 years from now?” the professor asked.
Discussion continued after the talk and I spoke with Professor Moor for a few minutes about the cyborg outcome. I haven’t given it much credibility in the past, but after this talk I’m changing my mind (rimshot). There exist three certainties, given a flexible enough timeframe:
- we will invent technologies more complex than our brains
- we will teach these AI’s ethics
- these AI’s will evolve themselves to be smarter and more capable than we are
Insofar as we cannot know the mind of God, we don’t know what these AI’s will think of us or how their ethics will evolve. We do know the human race is resource-intensive and damaging to the environment to a degree that an artificial being would not be, so they may view us as pests, or if we’re lucky they may simply view us as pets or respected ancestors.
Either way, Artificial Intelligence isn’t the right term anymore, that’s just one aspect of the project. We’re about to (on most timescales) create a new form of life here on Earth and if we intend to compete for any of the same niches as Engineered Life, well, evolution does not tend to be favorable to the less capable. So, we may be the species that gets to decide if it wants to evolve or not. In this case, the evolution would be to compete with the Engineered Lifeforms, and that makes the Cyborg outcome the most likely for the humans who wish to compete. The degree to which this will be necessary to maintain the species is decidedly unclear and perhaps disquieting for those of us who will be around then and whose children will be smack in the middle of it.
One thing’s for certain - it won’t happen suddenly. I gave the example of treating a patient with Alzheimer’s Disease - what if we had a nano-treatment that could prop up the patient’s failed neurons with engineered replacement neurons to return quality of life to those suffering from the Disease? Hardly anybody would argue that was unethical or question the essence of the person. Now, suppose the person’s brain starts to fail in other ways simply from ‘old age’? Would it be wrong to continue the treatment to prevent the person from dying? It would be hard to argue for letting the person simply die on principle. So, over the next dozen years, perhaps all the natural brain has died and been replaced with engineered structures. But the outward appearance of the person is unchanged, because the process was gradual and the neurons are an exact replacement for the person’s original brain. “Mommy, is Grandma a Computer?”
Some other interesting notes from the Conference:
EXAMPLES OF LINGUISTIC AMBIGUITY (how it’s hard to teach an AI)
flying over Zurich.
Fred saw the mountains flying over Zurich.
The police arrested the demonstrators because they feared violence.
The police arrested the demonstrators because they advocated violence.
These are hard for humans to parse correctly, let alone an AI.
A software program using neural networks such that given a board layout and dice rules, but no game rules, after 1.5 million games simulated in the neural network, the resulting AI was able to beat a champion-level human player in Backgammon. Again, he wasn’t taught the rules. Oh, look, I’ve just gone and anthropomorphized an AI.
When last I wrote about the Outlawing of Fair Use of video, I wrote that HR 4569 aims to require analog transmissions to maintain DRM “somehow”.
It turns out the “somehow” has a name, but we don’t know the how. The name is VEIL, and its workings are secret. For $10K and an NDA you can get a look at the decoder specs, but nobody can learn how the encoder works. Passage of this bill would create a secret law, where the citizenry must comply with a technology but may not know how it works or be able to debate its merits in the legislative process.
Again, write your Representative.
Bob Cringely’s column this week offers the first sensible explaination I’ve heard on why the NSA is skipping the FISA court for its ‘wiretapping’ activities. The hypothesis fits the data, at least, so it’s a resonable start to understand what’s going on.
New Scientist has an accessible article on Heim Theory – an 8 dimensional theory developed to unify quantum physics and relativity. The theory is not well documented or published but it does manage to derive the masses of elemental particles with astonishing accuracy and provides a mechanism/alterative for dark energy/matter. Should it be true it may be possible to build a space propulsion system using high energy magnetism that exceeds our current hopes, which has attracted interest from military and national labs.
I still maintain that constants are limitations in our knowledge and a proper theory will allow for the derivation of these what we now consider magic numbers.
The ACLU has a cute Flash demo showing a possible future where there are no consumer data privacy regulations. Combined with very tight integration between query software, databases, high speed internet, nation ID numbers, and corporate partnerships, this particular future looks bleak. It encourages one to think about the root causes of such a future and what might be the right thing to do in the present.
Joe Born, CEO of Neuros has posted his letter to Reps. Sensenbrenner and Conyers over their new bill, HR 4569, the Digital Transition Content Security Act. Neuros makes recording devices for people to time/place-shift their media viewing.
This bill aims to outlaw any electronic devices which convert analog video signals to any digital form without “somehow” preserving DRM.
It thereby also outlaws Fair Use of digital media since the DMCA made the “analog hole” the sole remedy for fair use of digital media. To my readers from Wisconsin or Minnesota: remember how these guys think come Election Day. Also ask yourself if they’re doing this for their citizens’ benefit or if they’re bought and paid for.
Please write your Congressman to express your thoughts on this issue. You don’t even need a stamp.
Here’s an article at Wood & Wood Products Magazine describing a new type of glue made from soy flour that’s replacing the urea formaldehyde glues typically used in plywoods.
The urea formaldehyde glues are especially troublesome as they outgas formaldehyde into the living space. The new type of glue was made by Kaichang Li at Oregon State University who studied the adhesive mussels use to attach themselves to rocks. Li figured out how the adhesive works and set to developing a process that can modify soy proteins to have similar properties. The method is simple enough that he can do it in his kitchen mixer.
According to this month’s Journal of Light Construction (not yet online) the new plywood can be boiled for 20 hours, dried, and boiled again without delaminating. The standard test for outdoor plywoods is a 4-hour boil and today’s plywoods cannot survive the 20-hour test.
Columbia Forest Products has already switched its hardwood veneer plywoods over to the new adhesive and is investigating switching over its other plywoods. As urea formaldehyde is derived from fossil fuels, its price is expected to increase with oil.