protocol

Libre Launches Revolutionary “MAVID”

By IoT – Internet of Things

Libre Wireless Technologies, Inc. recently announced “MAVID,” an entirely new embedded voice enabled multi-protocol wireless connectivity and IoT platform. This breakthrough platform features ultra low power and incredibly small size and incorporates a complete hardware and software solution for embedded voice, IoT, audio and wireless connectivity, and features Amazon Alexa Voice Service support. In addition […]

The post Libre Launches Revolutionary “MAVID” appeared first on IoT – Internet of Things.

Read more here:: iot.do/feed

Libre Launches Revolutionary “MAVID”

By News Aggregator

By IoT – Internet of Things

Libre Wireless Technologies, Inc. recently announced “MAVID,” an entirely new embedded voice enabled multi-protocol wireless connectivity and IoT platform. This breakthrough platform features ultra low power and incredibly small size and incorporates a complete hardware and software solution for embedded voice, IoT, audio and wireless connectivity, and features Amazon Alexa Voice Service support. In addition […]

The post Libre Launches Revolutionary “MAVID” appeared first on IoT – Internet of Things.

Read more here:: iot.do/feed

The post Libre Launches Revolutionary “MAVID” appeared on IPv6.net.

Read more here:: IPv6 News Aggregator

Voices in AI – Episode 17: A Conversation with James Barrat

By News Aggregator

By Byron Reese

Today’s leading minds talk AI with host Byron Reese

In this episode, Byron and James talk about jobs, human vs. artificial intelligence, and more.



0:00


0:00


0:00

Voices in AI

Visit VoicesInAI.com to access the podcast, or subscribe now:

Byron Reese: Hello, this is Voices in AI, brought to you by Gigaom. I am Byron Reese. Today I am so excited that our guest is James Barrat. He wrote a book called Our Final Invention, subtitled Artificial Intelligence and the End of the Human Era. James Barratt is also a renowned documentary filmmaker, as well as an author. Welcome to the show, James.

James Barrat: Hello.

So, let’s start off with, what is artificial intelligence?

Very good question. Basically, artificial intelligence is when machines perform tasks that are normally ascribed to human intelligence. I have a very simple definition of intelligence that I like. Because ‘artificial intelligence’—the definition just throws the ideas back to humans, and [to] human intelligence, which is the intelligence we know the most about.

The definition I like is: intelligence is the ability to achieve goals in a variety of novel environments, and to learn. And that’s a simple definition, but a lot is packed into it. Your intelligence has to achieve goals, it has to do something—whether that’s play Go, or drive a car, or solve proofs, or navigate, or identify objects. And if it doesn’t have some goal that it achieves, it’s not very useful intelligence.

If it can achieve goals in a variety of environments, if it can do object recognition and do navigation and do car-driving like our intelligence can, then it’s better intelligence. So, it’s goal-achieving in a bunch of novel environments, and then it learns. And that’s probably the most important part. Intelligence learns and it builds on its learning.

And you wrote a widely well-received book, Artificial Intelligence: Our Final Invention. Can you explain to the audience just your overall thesis, and the main ideas of the book?

Sure. Our Final Invention is basically making the argument that AI is a dual-use technology. A dual-use technology is one that can be used for great good, or great harm. Right now we’re in a real honeymoon phase of AI, where we’re seeing a lot of nifty tools come out of it, and a lot more are on the horizon. AI, right now, can find cancer clusters in x-rays better than humans. It can do business analytics better than humans. AI is doing what first year legal associates do, it’s doing legal discovery.

So we are finding a lot of really useful applications. It’s going to make us all better drivers, because we won’t be driving anymore. But it’s a dual-use technology because, for one thing, it’s going to be taking a lot of jobs. You know, there are five million professional drivers in the United States, seven million back-office accountants—those jobs are going to go away. And a lot of others.

So the thesis of my book is that we need to look under the hood of AI, look at its applications, look who’s controlling it, and then in a longer term, look at whether or not we can control it at all.

Let’s start with that point and work backwards. That’s an ominous statement. Can we record it at all? What are you thinking there?

Can we control it at all.

I’m sorry, yes. Control it at all.

Well, let me start, I prefer to start the other way. Stephen Hawking said that the trouble with AI is, in the short term, who controls it, and in the long term, can we control it at all? And in the short term, we’ve already suffered some from AI. You know, the NSA recently was accessing your phone data and mine, and getting your phone book and mine. And it was, basically, seizing our phone records, and that used to be illegal.

Used to be that if I wanted to seize, to get your phone records, I needed to go to a court, and get a court order. And that was to avoid abridging the Fourth Amendment, which prevents illegal search and seizure of property. Your phone messages are your property. The NSA went around that, and grabbed our phone messages and our phone data, and they are able to sift through this ocean of data because of AI, because of advanced data mining software.

One other example—and there are many—one other example of, in the short term, who controls the AI, is, right now there are a lot of countries developing battlefield robots and drones that will be autonomous. And these are robots and drones that kill people without a human in the loop. And these are AI issues. There are fifty-six nations developing battlefield robots.

The most sought after will be autonomous battlefield robots. There was an article just a couple of days ago about how the Marines have a robot that shoots a machinegun on a battlefield. They control it with a tablet, but their goal, as stated there, is to make it autonomous, to work on its own.

In the longer-term we, I’ll put it in the way that Arthur C. Clark put it to me, when I interviewed him. Arthur C. Clark was a mathematician and a physicist before he was a science fiction writer. And he created the HAL 9000 from 2001: A Space Odyssey, probably the most famous homicidal AI. And he said, when I asked him about the control problem of artificial intelligence, he said something like this: He said, “We humans steer the future not because we are the fastest or the strongest creatures, but because we are the most intelligent. And when we share the planet with something that’s more intelligent than we are, it will steer the future.”

So the problem we’re facing, the problem we’re on the cusp of, I can simplify it with a concept called ‘the intelligence explosion’. The intelligence explosion was an idea created by a statistician named I. J. Good in the 1960s. He said, “Once we create machines that do everything as well or better than humans, one of the things they’ll do is create smart machines.”

And we’ve seen artificial intelligence systems slowly begin to do things better than we do, and it’s not a stretch to think about a time to come, when artificial intelligence systems do advanced AI research and development better that humans. And I. J. Good said, “Then, when that happens, we humans will no longer set the pace of intelligence advancement, it will be machines that will set the pace of advancement.”

The trouble of that is, we know nothing about how to control a machine, or a cognitive architecture, that’s a thousand or million times more intelligent than we are. We have no experience with anything like that. We can look around us for analogies in the animal world.

How do we treat things that we’re a thousand times more intelligent than? Well, we treat all animals in a very negligent way. And the smart ones are either endangered, or they’re in zoos, or we eat them. That’s a very human-centric analogy, but I think it’s probably appropriate.

Let’s push on this just a little bit. So do you…

Sure.

Do you believe… Some people say ‘AI’ is kind of this specter of a term now, that, it isn’t really anything different than any other computer programs we’ve ever run, right? It’s better and faster and all of that, but it isn’t qualitatively anything different than what we’ve had for decades.

And so why do you think that? And when you say that AIs are going to be smarter than us, a million times smarter than us, ‘smarter’ is also a really nebulous term.

I mean, they may be able to do some incredibly narrow thing better than us. I may not be able to drive a car as well as an AI, but that doesn’t mean that same AI is going to beat me at Parcheesi. So what do you think is different? Why isn’t this just incrementally… Because so far, we haven’t had any trouble.

What do you think is going to be the catalyst, or what is qualitatively different about what we are dealing with now?

Sure. Well, there’s a lot of interesting questions packed into what you just said. And one thing you said—which I think is important to draw out—is that there are many kinds of intelligence. There’s emotional intelligence, there’s rational intelligence, there’s instinctive and animal intelligence.

And so, when I say something will be much more intelligent than we are, I’m using a shorthand for: It will be better at our definition of intelligence, it will be better at solving problems in a variety of novel environments, it will be better at learning.

And to put what you asked in another way, you’re saying that there is an irreducible promise and peril to every technology, including computers. All technologies, back to fire, have some good points and some bad points. AI I find qualitatively different. And I’ll argue by analogy, for a second. AI to me is like nuclear fission. Nuclear fission is a dual-use technology capable of great good and great harm.

Nuclear fission is the power behind atom bombs and behind nuclear reactors. When we were developing it in the ‘20s and ‘30s, we thought that nuclear fission was a way to get free energy by splitting the atom. Then it was quickly weaponized. And then we used it to incinerate cities. And then we as a species held a gun at our own heads for fifty years with the arms race. We threatened to make ourselves extinct. And that almost succeeded a number of times, and that struggle isn’t over.

To me, AI is a lot more like that. You said it hasn’t been used for nefarious reasons, and I totally disagree. I gave you an example with the NSA. A couple of weeks ago, Facebook was caught up because they were targeting emotionally-challenged and despairing children for advertising.

To me, that’s extremely exploitative. It’s a rather soulless and exploitative commercial application of artificial intelligence. So I think these pitfalls are around us. They’re already taking place. So I think the qualitative difference with artificial intelligence is that intelligence is our superpower, the human superpower.

It’s the ability to be creative, the ability to invent technology. That was one thing Stephen Hawking brought up when he was asked about, “What are the pitfalls of artificial intelligence?”

He said, “Well, for one thing, they’ll be able to develop weapons we don’t even understand.” So, I think the qualitative difference is that AI is the invention that creates inventions. And we’re on the cusp, this is happening now, and we’re on the cusp of an AI revolution, it’s going to bring us great profit and also great vulnerability.

You’re no doubt familiar with Searle’s “Chinese Room” kind of question, but all of the readers, all of the listeners might not be… So let me set that up, and then get your thought on it. It goes like this:

There’s a person in a room, a giant room full of very special books. And he doesn’t—we’ll call him the librarian—and the librarian doesn’t speak a word of Chinese. He’s absolutely unfamiliar with the language.

And people slide him questions under the door which are written in Chinese, and what he does—what he’s learned to do—is to look at the first character in that message, and he finds the book, of the tens of thousands that he has, that has that on the spine. And in that book he looks up the second character. And the book then says, “Okay, go pull this book.”

And in that book he looks up the third, and the fourth, and the fifth, all the way until he gets to the end. And when he gets to the end, it says “Copy this down.” And so he copies these characters again that he doesn’t understand, doesn’t have any clue whatsoever what they are.

He copies them down very carefully, very faithfully, slides it back under the door… Somebody’s outside who picks it up, a Chinese speaker. They read it, and it’s just brilliant! It’s just absolutely brilliant! It rhymes, it’s Haiku, I mean it’s just awesome!

Now, the question, the kind of ta-da question at the end is: Does the man, does the librarian understand Chinese? Does he understand Chinese?

Now, many people in the computer world would say yes. I mean, Alan Turing would say yes, right? The Chinese room passes the Turing Test. The Chinese speakers outside, as far as they know, they are conversing with a Chinese speaker.

So do you think the man understands Chinese? And do you think… And if he doesn’t understand Chinese… Because obviously, the analogy of it is: that’s all that computer does. A computer doesn’t understand anything. It doesn’t know if it’s talking about cholera or coffee beans or anything whatsoever. It runs this program, and it has no idea what it’s doing.

And therefore it has no volition, and therefore it has no consciousness; therefore it has nothing that even remotely looks like human intelligence. So what would you just say to that?

The Chinese Room problem is fascinating, and you could write books about it, because it’s about the nature of consciousness. And what we don’t know about consciousness, you could fill many books with. And I used to think I wanted to explore consciousness, but it made exploring AI look easy.

I don’t know if it matters that the machine thinks as we do or not. I think the point is that it will be able to solve problems. We don’t know about the volition question. Let me give you another analogy. When Ferrucci, [when] he was the head of Team Watson, he was asked a very provocative question: “Was Watson thinking when it beat all those masters at Jeopardy?” And his answer was, “Does a submarine swim?”

And what he meant was—and this is the twist on on the Chinese Room problem—he meant [that] when they created submarines, they learned principles of swimming from fish. But then they created something that swims farther and faster and carries a huge payload, so it’s really much more powerful than fish.

It doesn’t reproduce and it doesn’t do some of the miraculous things fish do, but as far as swimming, it does it. Does an airplane fly? Well, the aviation pioneers used principles of flight from birds, but quickly went beyond that, to create things that fly farther and faster and carry a huge payload.

I don’t think it matters. So, two answers to your question. One is, I don’t think it matters. And I don’t think it’s possible that a machine will think qualitatively as we do. So, I think it will think farther and faster and carry a huge payload. I think it’s possible for a machine to be generally intelligent in a variety of domains.

We can see intelligence growing in a bunch of domains. If you think of them as rippling pools, ripples in a pool, like different circles of expertise ultimately joining, you can see how general intelligence is sort of demonstrably on its way.

Whether or not it thinks like a human, I think it won’t. And I think that’s a danger, because I think it won’t have our mammalian sense of empathy. It’ll also be good, because it won’t have a lot of sentimentality, and a lot of cognitive biases that our brains are labored with. But you said it won’t have volition. And I don’t think we can bet on that.

In my book, Our Final Invention, I interviewed at length Steve Omohundro, who’s taken upon himself—he’s an AI maker and physicist—and he’d taken it upon himself to create more or less a science for understanding super intelligent machines. Or machines that are more intelligent than we are.

And among the things that he argues for, using rational-age and economic theory—and I won’t go into that whole thing—but it’s in Our Final Invention, it’s also in Steve Omohundro’s many websites. Machines that are self-aware and are self-programming, he thinks, will develop basic drives that are not unlike our own.

And they include things like self-protection, creativity, efficiency with resources,and other drives that will make them very challenging to control—unless we get ahead of the game and create this science for understanding them, as he’s doing.

Right now, computers are not generally intelligent, they are not conscious. All the limitations of the Chinese Room, they have. But I think it’s unrealistic to think that we are frozen in development. I think it’s very realistic to think that we’ll create machines whose cognitive abilities match and then outstrip our own.

But, just kind of going a little deeper on the question. So we have this idea of intelligence, which there is no consensus definition on it. Then within that, you have human intelligence—which, again, is something we certainly don’t understand. Human intelligence comes from our brain, which is—people say—‘the most complicated object in the galaxy’.

We don’t understand how it works. We don’t know how thoughts are encoded. We know incredibly little, in the grand scheme of things, about how the brain works. But we do know that humans have these amazing abilities, like consciousness, and the ability to generalize intelligence very effortlessly. We have something that certainly feels like free will, we certainly have something that feels like… and all of that.

Then on the other hand, you think back to a clockwork, right? You wind up a clock back in the olden days and it just ran a bunch of gears. And while it may be true that the computers of the day add more gears and have more things, all we’re doing is winding it up and letting it go.

And, isn’t it, like… not only a stretch, not only a supposition, not only just sensationalistic, to say, “Oh no, no. Someday we’ll add enough gears that, you wind that thing up, and it’s actually going to be a lot smarter than you.”

Isn’t that, I mean at least it’s fair to say there’s absolutely nothing we understand about human intelligence, and human consciousness, and human will… that even remotely implies that something that’s a hundred percent mechanical, a hundred percent deterministic, a hundred percent… Just wind it and it doesn’t do anything. But…

Well, you’re wrong about being a hundred percent deterministic, and it’s not really a hundred percent mechanical. When you talk about things like will, will is such an anthropomorphic term, I’m not sure if we can really, if we can attribute it to computers.

Well, I’m specifically saying we have something that feels and seems like will, that we don’t understand.

If you look, if you look at artificial neural nets, there’s a great deal about them we don’t understand. We know what the inputs are, and we know what the outputs are; and when we want to make better output—like a better translation—we know how to adjust the inputs. But we don’t know what’s going on in a multilayered neural net system. We don’t know what’s going on in a high resolution way. And that’s why they’re called black box systems, and evolutionary algorithms.

In evolutionary algorithms, we have a sense of how they work. We have a sense of how they combine pieces of algorithms, how we introduce mutations. But often, we don’t understand the output, and we certainly don’t understand how it got there, so that’s not completely deterministic. There’s a bunch of stuff we can’t really determine in there.

And I think we’ve got a lot of unexplained behavior in computers that’s, at this stage, we simply attribute to our lack of understanding. But I think in the longer term, we’ll see that computers are doing things on their own. I’m talking about a lot of the algorithms on Wall Street, a lot of the flash crashes we’ve seen, a lot of the cognitive architectures. There’s not one person who can describe the whole system… the ‘quants’, they call them, or the guys that are programming Wall Street’s algorithms.

They’ve already gone, in complexity, beyond any individual’s ability to really strip them down.

So, we’re surrounded by systems of immense power. Gartner and company think that in the AI space—because of the exponential nature of the investment… I think it started out, and it’s doubled every year since 2009—Gartner estimates that by 2025, that space will be worth twenty-five trillion dollars of value. So to me, that’s a couple of things.

That anticipates enormous growth, and enormous growth in power in what these systems will do. We’re in an era now that’s different from other eras. But it is like other Industrial Revolutions. We’re in an era now where everything that’s electrified—to paraphrase Kevin Kelly, the futurist—everything that’s electrified is being cognitized.

We can’t pretend that it will always be like a clock. Even now it’s not like a clock. A clock you can take apart, and you can understand every piece of it.

The cognitive architectures we’re creating now… When Ferrucci was watching Watson play, and he said, “Why did he answer like that?” There’s nobody on his team that knew the answer. When it made mistakes… It did really, really well; it beat the humans. But comparing [that] to a clock, I think that’s the wrong metaphor.

Well, let’s just poke at it just one more minute, and then we can move on to something else. Is that really fair to say, that because humans don’t understand how it works, it must be somehow working differently than other machines?

Put another way, it is fair to say, because we’ve added enough gears now, that nobody could kind of keep them all straight. I mean nobody understands why the Google algorithm—even at Google—turns up what it does when you search. But nobody’s suggesting anything nondeterministic, nothing emergent, anything like that is happening.

I mean, our computers are completely deterministic, are they not?

I don’t think that they are. I think if they were completely deterministic, then enough brains put together could figure out a multi-tiered neural net, and I don’t think there’s any evidence that we can right now.

Well, that’s exciting.

I’m not saying that it’s coming up with brilliant new ideas… But a system that’s so sophisticated that it defeats Go, and teaches grandmasters new ideas about Go—which is what the grandmaster who it defeated three out of four times said—[he] said, “I have new insights about this game,” that nobody could explain what it was doing, but it was thinking creatively in a way that we don’t understand.

Go is not like chess. On a chess board, I don’t know how many possible positions there are, but it’s calculable. On a Go board, it’s incalculable. There are more—I’ve heard it said, and I don’t really understand it very well—I heard it said there are more possible positions on a Go board than there are atoms in the universe.

So when it’s beating Go masters… Therefore, playing the game requires a great deal of intuition. It’s not just pattern-matching. Like, I’ve played a million games of Go—and that’s sort of what chess is [pattern-matching].

You know, the grandmasters are people who have seen every board you could possibly come up with. They’ve probably seen it before, and they know what to do. Go’s not like that. It requires a lot more undefinable intuition.

And so we’re moving rapidly into that territory. The program that beat the Go masters is called AlphaGo. It comes out of DeepMind. DeepMind was bought four years ago by Google. Going deep into reinforcement learning and artificial neural nets, I think your argument would be apt if we were talking about some of the old languages—Fortran, Basic, Pascal—where you could look at every line of code and figure out what was going on.

That’s no longer possible, and you’ve got Go grandmasters saying “I learned new insights.” So we’re in a brave new world here.

So you had a great part of the book, where you do a really smart kind of roll-up of when we may have an AGI. Where you went into different ideas behind it. And the question I’m really curious about is this: On the one hand, you have Elon Musk saying we can have it much sooner than you think. You have Stephen Hawking, who you quoted. You have Bill Gates saying he’s worried about it.

So you have all of these people who say it’s soon, it’s real, and it’s potentially scary. We need to watch what we do. Then on the other camp, you have people who are equally immersed in the technology, equally smart, equally, equally, equally all these other things… like Andrew Ng, who up until recently headed up AI at Baidu, who says worrying about AGI is like worrying about overpopulation on Mars. You have other people saying the soonest it could possibly happen is five hundred years from now.

So I’m curious about this. Why do you think, among these big brains, super smart people, why do they have… What is it that they believe or know or think, or whatever, that gives them such radically different views about this technology? How do you get your head around why they differ?

Excellent question. I first heard that Mars analogy from, I think it was Sebastian Thrun, who said we don’t know how to get to Mars. We don’t know how to live on Mars. But we know how to get a rocket to the moon, and gradually and slowly, little by little—No, it was Peter Norvig, who wrote the sort of standard text on artificial intelligence, called AI: A Modern Approach.

He said, you know, “We can’t live on Mars yet, but we’re putting the rockets together. Some companies are putting in some money. We’re eventually going to get to Mars, and there’ll be people living on Mars, and then people will be setting another horizon.” We haven’t left our solar system yet.

It’s a very interesting question, and very timely, about when will we achieve human-level intelligence in a machine, if ever. I did a poll about it. It was kind of a biased poll; it was of people who were at a conference about AGI, about artificial general intelligence. And then I’ve seen a lot of polls, and there’s two points to this.

One is the polls go all over the place. Some people said… Ray Kurzweil says 2029. Ray Kurzweil’s been very good at anticipating the progress of technology, he says 2029. Ray Kurzweil’s working for Google right now—this is parenthetically—he said he wants to create a machine that makes three hundred trillion calculations per second, and to share that with a billion people online. So what’s that? That’s basically reverse engineering of a brain.

Making three hundred trillion calculations per second, which is sort of a rough estimate of what a brain does. And then sharing it with a billion people online, which is making superintelligence a service, which would be incredibly useful. You could do pharmacological research. You could do really advanced weather modeling, and climate modeling. You could do weapons research, you could develop incredible weapons. He says 2029.

Some people said one hundred years from now. The mean date that I got was about 2045 for human-level intelligence in a machine. And then my book, Our Final Invention, got reviewed by Gary Marcus in the New Yorker, and he said something that stuck with me. He said whether or not it’s ten years or one hundred years, the more important question is: What happens next?

Will it be integrated into our lives? Or will it suddenly appear? How are we positioned for our own safety and security when it appears, whether it’s in fifty years or one hundred? So I think about it as… Nobody thought Go was going to be beaten for another ten years.

And here’s another way… So those are the two ways to think about it: one is, there’s a lot of guesses; and two, does it really matter what happens next? But the third part of that is this, and I write about it in Our Final Invention: If we don’t achieve it in one hundred years, do you think we’re just going to stop? Or do you think we’re going to keep beating at this problem until we solve it?

And as I said before, I don’t think we’re going to create exactly human-like intelligence in a machine. I think we’re going to create something extremely smart and extremely useful, to some extent, but something we, in a very deep way, don’t understand. So I don’t think it’ll be like human intelligence… it will be like an alien intelligence.

So that’s kind of where I am on that. I think it could happen in a variety of timelines. It doesn’t really matter when, and we’re not going to stop until we get there. So ultimately, we’re going to be confronted with machines that are a thousand or a million times more intelligent than we are.

And what are we going to do?

Well, I guess the underlying assumption is… it speaks to the credibility of the forecast, right? Like, if there’s a lab, and they’re working on inventing the lightbulb, like: “We’re trying to build the incandescent light bulb.” And you go in there and you say, “When will you have the incandescent light bulb?” and they say “Three or four weeks, five weeks. Five weeks tops, we’re going to have it.”

Or if they say, “Uh, a hundred years. It may be five hundred, I don’t know.” I mean in those things you take a completely different view of, do we understand the problem? Do we know what we’re building? Do we know how to build an AGI? Do we even have a clue?

Do you believe… or here, let me ask it this way: Do you think an AGI is just an evolutionary… Like, we have AlphaGo, we have Watson, and we’re making them better every day. And eventually, that kind of becomes—gradually—this AGI. Or do you think there’s some “A-ha” thing we don’t know how to do, and at some point we’re like “Oh, here’s how you do it! And this is how you get a synapse to work.”

So, do you think we are nineteen revolutionary breakthroughs away, or “No, no, no, we’re on the path. We’re going to be there in three to five years.”?

Ben Goertzel, who is definitely in the race to make AGI—I interviewed him in my book—said we need some sort of breakthrough. And then we got to artificial neural nets and deep learning, and deep learning combined with reinforcement learning, which is an older technique, and that was kind of a breakthrough. And then people started to beat—IBM’s Deep Blue—to beat chess, it really was just looking up tables of positions.

But to beat Go, as we’ve discussed, was something different.

I think we’ve just had a big breakthrough. I don’t know how many revolutions we are away from a breakthrough that makes intelligence general. But let me give you this… the way I think about it.

There’s long been talk in the AI community about an algorithm… I don’t know exactly what they call it. But it’s basically an open-domain problem-solver that asks something simple like, what’s the next best move? What’s the next best thing to do? Best being based on some goals that you’ve got. What’s the next best thing to do?

Well, that’s sort of how DeepMind took on all the Atari games. They could drop the algorithm into a game, and it didn’t even know the rules. It just noticed when it was scoring or not scoring, and so it was figuring out what’s the next best thing to do.

Well if you can drop it into every Atari game, and then you drop it into something that’s many orders of magnitude above it, like Go, then why are we so far from dropping that into a robot and setting it out into the environment, and having it learn the environment and learn common sense about the environment—like, “Things go under, and things go over; and I can’t jump into the tree; I can climb the tree.”

It seems to me that general intelligence might be as simple as a program that says “What’s the next best thing to do?” And then it learns the environment, and then it solves problems in the environment.

So some people are going about that by training algorithms, artificial neural net systems and defeating games. Some people are really trying to reverse-engineer a brain, one neuron at a time. That’s sort of, in a nutshell—to vastly overgeneralize—that’s called the bottom-up, and the top-down approach for creating AGI.

So are we a certain number of revolutions away, or are we going to be surprised? I’m surprised a little too frequently for my own comfort about how fast things are moving. Faster than when I was writing the book. I’m wondering what the next milestone is. I think the Turing Test has not been achieved, or even close. I think that’s a good milestone.

It wouldn’t surprise me if IBM, which is great at issuing itself grand challenges and then beating them… But what’s great about IBM is, they’re upfront. They take on a big challenge… You know, they were beaten—Deep Blue was beaten several times before it won. When they took on Jeopardy, they weren’t sure they were going to win, but they had the chutzpah to get out there and say, “We’re gonna try.” And then they won.

I bet IBM will say, “You know what, in 2020, we’re going to take on the Turing Test. And we’re going to have a machine that you can’t tell that it’s a machine. You can’t tell the difference between a machine and a human.”

So, I’m surprised all the time. I don’t know how far or how close we are, but I’d say I come at it from a position of caution. So I would say, the window in which we have to create safe AI is closing.

Yes, no… I’m with you; I was just taking that in. I’ll insert some ominous “Dun, dun, dun…” Take that a little further.

Everybody has a role to play in this conversation, and mine happens to be canary in a coal mine. Despite the title of my book, I really like AI. I like its potential. Medical potential. I don’t like its war potential… If we see autonomous battlefield robots on the battlefield, you know what’s going to happen. Like every other piece of used military equipment, it’s going to come home.

Well, the thing is, about the military… and the thing about technology is…If you told my dad that he would invite into his home a representative of Google, and that representative would sit in a chair in a corner of the house, and he would take down everything we said, and would sell that data to our insurance company, so our insurance rates might go up… and it would sell that data to mortgage bankers, so they might cut off our ability to get a mortgage… because dad talks about going bankrupt, or dad talks about his heart condition… and he can’t get insurance anymore.

But if we hire a corporate guy, and we pay for it, and put him in our living room… Well, that’s exactly what we’re doing with Amazon Echo, with all the digital assistants. All this data is being gathered all the time, and it’s being sold… Buying and selling data is a four billion dollar-a-year industry. So we’re doing really foolish things with this technology. Things that are bad for our own interests.

So let me ask you an open-ended question… prognostication over shorter time frames is always easier. Tell me what you think is in store for the world, I don’t know, between now and 2030, the next thirteen years. Talk to me about unemployment, talk to me about economics, all of that. Tell me the next thirteen years.

Well, brace yourself for some futurism, which is a giant gamble and often wrong. To paraphrase Kevin Kelly again, everything that’s electrical will be cognitized. Our economy will be dramatically shaped by the ubiquity of artificial intelligence. With the Internet of Things, with the intelligence of everything around us—our phones, our cars…

I can already talk to my car. I’m inside my car, I can ask for directions, I can do some other basic stuff. That’s just going to get smarter, until my car drives itself. A lot of people… MIT did a study, that was quoting a Cambridge study, that said: “Forty-five percent of our jobs will be able to be replaced within twenty years.” I think they downgraded that to like ten years.

Not that they will be replaced, but they will be able to be replaced. But when AI is a twenty-five trillion dollar—when it’s worth twenty-five trillion dollars in 2025—everybody will be able to do anything, will be able to replace any employee that’s doing anything that’s remotely repetitive, and this includes doctors and lawyers… We’ll be able to replace them with the AI.

And this cuts deep into the middle class. This isn’t just people working in factories or driving cars. This is all accountants, this is a lot of the doctors, this is a lot of the lawyers. So we’re going to see giant dislocation, or giant disruption, in the economy. And giant money being made by fewer and fewer people.

And the trouble with that is, that we’ve got to figure out a way to keep a huge part of our population from starving, from not making a wage. People have proposed a basic minimum income, but to do that we would need tax revenue. And the big companies, Amazon, Google, Facebook, they pay taxes in places like Ireland, where there’s very low corporate tax. They don’t pay taxes where they get their wealth. So they don’t contribute to your roads.

Google is not contributing to your road system. Amazon is not contributing to your water supply, or to making your country safe. So there’s a giant inequity there. So we have to confront that inequity and, unfortunately, that is going to require political solutions, and our politicians are about the most technologically-backward people in our culture.

So, what I see is, a lot of unemployment. I see a lot of nifty things coming out of AI, and I am willing to be surprised by job creation in AI, and robotics, and automation. And I’d like to be surprised by that. But the general trend is… When you replace the biggest contract manufacturer in the world… Foxconn just replaced thirty-thousand people in Asia with thirty-thousand robots.

And all those people can’t be retrained, because if you’re doing something that’s that repetitive, and that mechanical… what can you be retrained to do? Well, maybe one out of every hundred could be a floor manager in a robot factory, but what about all the others? Disruption is going to come from all the people that don’t have jobs, and there’s nothing to be retrained to.

Because our robots are made in factories where robots make the robots. Our cars are made in factories where robots make the cars.

Isn’t that the same argument they used during the Industrial Revolution, when they said, “You got ninety percent of people out there who are farmers, and we’re going to lose all these farm jobs… And you don’t expect those farmers are going to, like, come work in a factory, where they have to learn completely new things.”

Well, what really happened in the different technology revolutions, back from the cotton gin onward is, a small sector… The Industrial Revolution didn’t suddenly put farms out of business. A hundred years ago, ninety percent of people worked on farms, now it’s ten percent.

But what happened with the Industrial Revolution is, sector by sector, it took away jobs, but then those people could retrain, and could go to other sectors, because there were still giant sectors that weren’t replaced by industrialization. There was still a lot of manual labor to do. And some of them could be trained upwards, into management and other things.

This, as the author Ford wrote in The Rise of Robots—and there’s also a great book called The Fourth Industrial Age. As they both argue, what’s different about this revolution is that AI works in every industry. So it’s not like the old revolutions, where one sector was replaced at a time, and there was time to absorb that change, time to reabsorb those workers and retrain them in some fashion.

But everybody is going to be… My point is, all sectors of the economy are going to be hit at once. The ubiquity of AI is going to impact a lot of the economy, all at the same time, and there is going to be a giant dislocation all at the same time. And it’s very unclear, unlike in the old days, how those people can be retrained and retargeted for jobs. So, I think it’s very different from other Industrial Revolutions, or rather technology revolutions.

Other than the adoption of coal—it went from generating five percent to eighty percent of all of our power in twenty years—the electrification of industry happened incredibly fast. Mechanization, replacement of animal power with mechanical power, happened incredibly fast. And yet, unemployment remains between four and nine percent in this country.

Other than the Depression, without ever even hiccupping—like, no matter what disruption, no matter what speed you threw at it—the economy never couldn’t just use that technology to create more jobs. And isn’t that maybe a lack of imagination that says “Well, no, now we’re out. And no more jobs to create. Or not ones that these people who’ve been displaced can do.”

I mean, isn’t that what people would’ve said for two hundred years?

Yes, that’s a somewhat persuasive argument. I think you’ve got a point that the economy was able to absorb those jobs, and the unemployment remained steady. I do think this is different. I think it’s a kind of a puzzle, and we’ll have to see what happens. But I can’t imagine… Where do professional drivers… they’re not unskilled, but they’re right next to it. And it’s the job of choice for people who don’t have a lot of education.

What do you retrain professional drivers to do once their jobs are taken? It’s not going to be factory work, it’s not going to be simple accounting. It’s not going to be anything repetitive, because that’s going to be the job of automation and AI.

So I anticipate problems, but I’d love to be pleasantly surprised. If it worked like the old days, then all those people that were cut off the farm would go to work in the factories, and make Ford automobiles, and make enough money to buy one. I don’t see all those driverless people going off to factories to make cars, or to manufacture anything.

A case in point of what’s happening is… Rethink Robotics, which is Rodney Brooks’ company, just built something called Baxter; and now Baxter is a generation old, and I can’t think of what replaced it. But it costs about twenty-two thousand dollars to get one of these robots. These robots cost basically what a minimum wage worker makes in a year. But they work 24/7, so they really replace three shifts, so they really are replacing three people.

Where do those people go? Do they go to shops that make Baxter? Or maybe you’re right, maybe it’s a failure of imagination to not be able to anticipate the jobs that would be created by Baxter and by autonomous cars. Right now, it’s failing a lot of people’s imagination. And there are not ready answers.

I mean, if it were 1995 and the Internet was, you’re just hearing about it, just getting online, just hearing it… And somebody said, “You know what? There’s going to be a lot of companies that just come out and make hundreds of billions of dollars, one after the other, all because we’ve learned how to connect computers and use this hypertext protocol to communicate.” I mean, that would not have seemed like a reasonable surmise.

No, and that’s a great example. If you were told that trillions of dollars of value are going to come out of this invention, who would’ve thought? And maybe I personally, just can’t imagine the next wave that is going to create that much value. I can see how AI and automation will create a lot of value, I only see it going into a few pockets though. I don’t see it being distributed in any way that the Silicon Valley startups, at least initially, were.

So let’s talk about you for a moment. Your background is in documentary filmmaking. Do you see yourself returning to that world? What are you working on, another book? What kind of thing is keeping you busy by day right now?

Well, I like making documentary films. I just had one on PBS last year… If you Google “Spillover” and “PBS” you can see it is streaming online. It was about spillover diseases—Ebola, Zika and others—and it was about the Ebola crisis, and how viruses spread. And then now I’m working on a film about paleontology, about a recent discovery that’s kind of secret, that I can’t talk about… from sixty-six million years ago.

And I am starting to work on another book that I can’t talk about. So I am keeping an eye on AI, because this issue is… Despite everything I talk about, I really like the technology; I think it’s pretty amazing.

Well, let’s close with, give me a scenario that you think is plausible, that things work out. That we have something that looks like full employment, and…

Good, Byron. That’s a great way to go out. I see people getting individually educated about the promise and peril of AI, so that we as a culture are ready for the revolution that’s coming. And that forces businesses to be responsible, and politicians to be savvy, about developments in artificial intelligence. Then they invest some money to make artificial intelligence advancement transparent and safe.

And therefore, when we get to machines that are as smart as humans, that [they] are actually our allies, and never our competitors. And that somehow on top of this giant wedding cake I’m imagining, we also manage to keep full employment, or nearly-full employment. Because we’re aware, and because we’re working all the time to make sure that the future is kind to humans.

Alright, well, that is a great place to leave it. I am going to thank you very much.

Well, thank you. Great questions. I really enjoyed the back-and-forth.

Byron explores issues around artificial intelligence and conscious computers in his upcoming book The Fourth Age, to be published in April by Atria, an imprint of Simon & Schuster. Pre-order a copy here.

Voices in AI

Visit VoicesInAI.com to access the podcast, or subscribe now:

Read more here:: gigaom.com/feed/

The post Voices in AI – Episode 17: A Conversation with James Barrat appeared on IPv6.net.

Read more here:: IPv6 News Aggregator

LoRa technology is driving IoT adoption, changing lives

By Zenobia Hedge

As decades-old technologies, Wi-Fi, Bluetooth and cellular are well established, but they were not designed to allow long-range communications at a low bit rate among things such as sensors powered by a battery. These conditions require Low-Power Wide-Area Network (LPWAN) which aims to connect billions of battery-powered sensors to the Cloud, and this is where LoRa® devices and wireless radio frequency technology (LoRa Technology) and the open LoRaWAN™ protocol both are driving and enabling new applications and business models, says Mike Wong, vice president of Semtech.

The most pressing challenges for Internet of Things (IoT) are interoperability of various networks, security for billions of sensors and the data they produce, providing carrier-grade quality, and reliability at consumer price points. These challenges are tied together because adoption will slow down if IoT options are not available at accessible prices, and the devices will not be economically feasible if there is little adoption.

LoRa technology, originally developed by Semtech, is one of the founders of the LoRa Alliance™ which launched in 2015 and now boasts more than 500 companies. The organisation is dedicated to developing devices, technologies and applications under the same set of guidelines, with the same purpose of making IoT possible and driving the adoption of the LoRaWAN open protocol.

It is fast emerging as the best option for many markets and verticals including M2M applications, supply chain and logistics applications, smart cities, smart metering applications, and agriculture. The platform is seeing success with these applications because it is able to improve quality of life and helps solve real-world problems such as water conservation, environmental monitoring for pollutants, flood-level monitors in cities, and improving food safety and quality. We are all affected by these issues although it may not be obvious to most people that LoRa technology enables these improvements.

Smart metering, for example, uses LoRa-enabled devices to accurately measure the flow of gas or water, and takes away a lot of guesswork by delivering precise intelligence on critical points like when to provide maintenance. Three key features – low-power, low-cost and an open interoperable standard – make LoRa desirable for myriad industrial, commercial and domestic applications where IoT solutions are the future.

LoRaWAN deployments typically come in the form of a star network topology where the concentrators serve as a central coordinator for network traffic. Due to limitations of radio technology, one of the limitations of alternatives such as mesh networks is their inability to extend the range. With star networks and LoRa technology, range is a problem that has been solved.

If you look at the architecture of a LoRaWAN network, a lot of focus has been on keeping end node and concentrators simple, low-power and low-cost because this is what is needed to scale up to billions of nodes. Most of the processing happens in the cloud in a LoRaWAN network, eliminating the need for complex software that, for example, makes it necessary for individual nodes to maintain redundant information to switch roles which drives up the deployment and hardware costs. It’s a […]

The post LoRa technology is driving IoT adoption, changing lives appeared first on IoT Now – How to run an IoT enabled business.

Read more here:: www.m2mnow.biz/feed/

LoRa technology is driving IoT adoption, changing lives

By News Aggregator

By Zenobia Hedge

As decades-old technologies, Wi-Fi, Bluetooth and cellular are well established, but they were not designed to allow long-range communications at a low bit rate among things such as sensors powered by a battery. These conditions require Low-Power Wide-Area Network (LPWAN) which aims to connect billions of battery-powered sensors to the Cloud, and this is where LoRa® devices and wireless radio frequency technology (LoRa Technology) and the open LoRaWAN™ protocol both are driving and enabling new applications and business models, says Mike Wong, vice president of Semtech.

The most pressing challenges for Internet of Things (IoT) are interoperability of various networks, security for billions of sensors and the data they produce, providing carrier-grade quality, and reliability at consumer price points. These challenges are tied together because adoption will slow down if IoT options are not available at accessible prices, and the devices will not be economically feasible if there is little adoption.

LoRa technology, originally developed by Semtech, is one of the founders of the LoRa Alliance™ which launched in 2015 and now boasts more than 500 companies. The organisation is dedicated to developing devices, technologies and applications under the same set of guidelines, with the same purpose of making IoT possible and driving the adoption of the LoRaWAN open protocol.

It is fast emerging as the best option for many markets and verticals including M2M applications, supply chain and logistics applications, smart cities, smart metering applications, and agriculture. The platform is seeing success with these applications because it is able to improve quality of life and helps solve real-world problems such as water conservation, environmental monitoring for pollutants, flood-level monitors in cities, and improving food safety and quality. We are all affected by these issues although it may not be obvious to most people that LoRa technology enables these improvements.

Smart metering, for example, uses LoRa-enabled devices to accurately measure the flow of gas or water, and takes away a lot of guesswork by delivering precise intelligence on critical points like when to provide maintenance. Three key features – low-power, low-cost and an open interoperable standard – make LoRa desirable for myriad industrial, commercial and domestic applications where IoT solutions are the future.

LoRaWAN deployments typically come in the form of a star network topology where the concentrators serve as a central coordinator for network traffic. Due to limitations of radio technology, one of the limitations of alternatives such as mesh networks is their inability to extend the range. With star networks and LoRa technology, range is a problem that has been solved.

If you look at the architecture of a LoRaWAN network, a lot of focus has been on keeping end node and concentrators simple, low-power and low-cost because this is what is needed to scale up to billions of nodes. Most of the processing happens in the cloud in a LoRaWAN network, eliminating the need for complex software that, for example, makes it necessary for individual nodes to maintain redundant information to switch roles which drives up the deployment and hardware costs. It’s a […]

The post LoRa technology is driving IoT adoption, changing lives appeared first on IoT Now – How to run an IoT enabled business.

Read more here:: www.m2mnow.biz/feed/

The post LoRa technology is driving IoT adoption, changing lives appeared on IPv6.net.

Read more here:: IPv6 News Aggregator

Huawei launches All-Intelligent Network to unleash the power of ultra-broadband

By Sheetal Kumbhar

At Ultra-Broadband Forum 2017, David Wang, president of products & solutions, Huawei, launched the All-Intelligent Network. By building intelligent, simple, and superfast ultra-broadband networks, Huawei aims to help carriers address future challenges brought by business uncertainties, and achieve new growth in the B2B, B2C, and B2H segments.

The forthcoming intelligent world will present new opportunities for the development of ultra-broadband. The number of video users has increased significantly, and millions of companies are migrating to the cloud. Leading countries have released timetables for 5G commercial deployment. However, many carrier networks still lack agile operation systems. They are hard to maintain, and rely on only a small number of security features for protection, which makes it difficult to support the future development of business where uncertainties are rife.

“In the physical world, intelligent transportation can address uncertainties in the transport network, such as traffic jams and other inefficiencies. Likewise, the uncertainties of the digital world also need to be addressed through intelligence. Intelligence arises from simplicity, and simplicity accelerates the integration of intelligence into networks,” said David Wang.

To this end, Huawei has launched the All-Intelligent Network, to help carriers address future business uncertainties. The All-Intelligent Network has three key features: intelligent, simple, and superfast.

Intelligence is the intelligent business operations and network O&M. There are three steps to build intelligent networks: from automatic to adaptive, and ultimately to autonomous. In this context, ‘automatic’ means the ability to automatically provision services and automate network deployment and maintenance. By integrating network managers and controllers, the All-Intelligent Network connects new SDN networks and traditional networks. This will enable one-click services and end-to-end automatic service provisioning.

‘Adaptive’ is built on intelligent analysers on top of automation, to collect real-time network data and perceive network status. The All-Intelligent Network is able to automatically generate and optimise policies based on service and network SLAs, enabling networks to evolve from open-loop configuration to closed-loop optimisation. ‘Autonomous’ is the enhancement of analysers by introducing AI and machine learning algorithms to enable network self-learning. Networks will evolve, from following programmed, static policies to learning dynamic policies on their own, thus achieving network autonomy.

The second feature is simplicity. The All-Intelligent Network optimises traditionally siloed architecture by leveraging IT concepts and technologies. It decouples networks into two layers – an elastic and reliable transport layer, and a flexible and agile service layer. The architecture gets simplified, where network resources are adjusted fast and flexibly.

David Wang, President of Products & Solutions at Huawei, delivers a keynote speech titled ‘Unleash the Power of UBB with All Intelligence’

By using a unified protocol, networking, which used to run on multiple IP protocols, becomes more standardised, and a standard service configuration model will be created to reduce the difficulty of O&M and configuration. The number of layers from OTN to CO is also reduced to enable one-hop access to any application, helping carriers reduce network latency, improve service experience, and flatten network topology.

The third feature is superfast. The definition of ultra-broadband is renewed to encompass ultra-low latency, ultra-high bandwidth, and […]

The post Huawei launches All-Intelligent Network to unleash the power of ultra-broadband appeared first on IoT Now – How to run an IoT enabled business.

Read more here:: www.m2mnow.biz/feed/

Huawei launches All-Intelligent Network to unleash the power of ultra-broadband

By News Aggregator

By Sheetal Kumbhar

At Ultra-Broadband Forum 2017, David Wang, president of products & solutions, Huawei, launched the All-Intelligent Network. By building intelligent, simple, and superfast ultra-broadband networks, Huawei aims to help carriers address future challenges brought by business uncertainties, and achieve new growth in the B2B, B2C, and B2H segments.

The forthcoming intelligent world will present new opportunities for the development of ultra-broadband. The number of video users has increased significantly, and millions of companies are migrating to the cloud. Leading countries have released timetables for 5G commercial deployment. However, many carrier networks still lack agile operation systems. They are hard to maintain, and rely on only a small number of security features for protection, which makes it difficult to support the future development of business where uncertainties are rife.

“In the physical world, intelligent transportation can address uncertainties in the transport network, such as traffic jams and other inefficiencies. Likewise, the uncertainties of the digital world also need to be addressed through intelligence. Intelligence arises from simplicity, and simplicity accelerates the integration of intelligence into networks,” said David Wang.

To this end, Huawei has launched the All-Intelligent Network, to help carriers address future business uncertainties. The All-Intelligent Network has three key features: intelligent, simple, and superfast.

Intelligence is the intelligent business operations and network O&M. There are three steps to build intelligent networks: from automatic to adaptive, and ultimately to autonomous. In this context, ‘automatic’ means the ability to automatically provision services and automate network deployment and maintenance. By integrating network managers and controllers, the All-Intelligent Network connects new SDN networks and traditional networks. This will enable one-click services and end-to-end automatic service provisioning.

‘Adaptive’ is built on intelligent analysers on top of automation, to collect real-time network data and perceive network status. The All-Intelligent Network is able to automatically generate and optimise policies based on service and network SLAs, enabling networks to evolve from open-loop configuration to closed-loop optimisation. ‘Autonomous’ is the enhancement of analysers by introducing AI and machine learning algorithms to enable network self-learning. Networks will evolve, from following programmed, static policies to learning dynamic policies on their own, thus achieving network autonomy.

The second feature is simplicity. The All-Intelligent Network optimises traditionally siloed architecture by leveraging IT concepts and technologies. It decouples networks into two layers – an elastic and reliable transport layer, and a flexible and agile service layer. The architecture gets simplified, where network resources are adjusted fast and flexibly.

David Wang, President of Products & Solutions at Huawei, delivers a keynote speech titled ‘Unleash the Power of UBB with All Intelligence’

By using a unified protocol, networking, which used to run on multiple IP protocols, becomes more standardised, and a standard service configuration model will be created to reduce the difficulty of O&M and configuration. The number of layers from OTN to CO is also reduced to enable one-hop access to any application, helping carriers reduce network latency, improve service experience, and flatten network topology.

The third feature is superfast. The definition of ultra-broadband is renewed to encompass ultra-low latency, ultra-high bandwidth, and […]

The post Huawei launches All-Intelligent Network to unleash the power of ultra-broadband appeared first on IoT Now – How to run an IoT enabled business.

Read more here:: www.m2mnow.biz/feed/

The post Huawei launches All-Intelligent Network to unleash the power of ultra-broadband appeared on IPv6.net.

Read more here:: IPv6 News Aggregator

IPv6 is Not Optional

By Mattias Lindgren

IPv6 Case Study from University of Colorado Denver in Denver, Colorado

The University of Colorado in Denver has a presence across two main campuses – the Anschutz Medical Campus in Aurora, Colorado and the Auraria Campus in Downtown Denver. Over the past year, I spearheaded the effort by the network team to complete a project of enabling IPv6 across both campuses. This involved extensive planning phases, going through project creation and approval, and finally the roll-out to our approximately 30,000 users.

Planning

We began by obtaining IPv6 space from ARIN. Lining up supporting documentation as well as approval of the contract amendment by the University purchasing team was a time-consuming portion of the planning phase. We knew we wanted to get ample space for the future, so we requested and got approval for a /40.

Make it Official

In discussion with my manager, we decided that the best way to get buy-in from the whole IT department was to make it an official project. This is key as it sets expectations for all the project members and stakeholders, and holds each team member accountable for completing their assigned tasks. We regularly hold a Project Review Board where members of the IT staff can present project frameworks. At one such meeting, I presented the benefits of implementing IPv6. Representing the network team, I argued that migrating to IPv6 is not optional, and adoption of it over the legacy IPv4 protocol is happening at other major universities and will soon become the dominant protocol on the internet. The earlier you adopt it, the easier implementation will be as networks only increase in complexity over time. We also discussed efficiency gains as complex NAT schemes would no longer be needed for pure IPv6 traffic. The project was approved by our CIO and it became an official mandate across central IT to deploy IPv6.

Project Rollout

We felt that an important part of the rollout was to educate the technical staff outside of the central IT department. To facilitate this, we brought in an outside vendor to give a half day presentation on IPv6 and what it meant. We had a good turnout of about 150 people.

Over the past couple of years, we implemented MPLS across both of our campuses. This made implementation of IPv6 easier as we could target well-defined chunks of our network for rollout. We saw a brand new IPv6 network as an opportunity to atone for legacy protocol sins, and designed the new dual-stack network with redundancy and scalability in mind from day one. Having already rolled out MPLS also presented its own challenges, as we found that some of the features present in the IPv4 world hadn’t yet become available on the IPv6 side. It has become a running joke that IPv6, at 20-some years old, is such a young protocol that we can’t expect vendors to support it just yet.

Since our legacy network routing table was sprawling with thousands of routes, we took the opportunity to summarize our IPv6 routing table as much as possible and it is currently less than 50 routes.

We decided on DHCPv6 for our addressing scheme, and began handing out addresses from our centralized DHCP and DNS appliance. This negated the need for disabling privacy-extensions and also made the client addresses a little more manageable.

One issue we’ve had for a long time is the presence of 6to4 adapters on our Windows computers. Since we still have workstations with public IPv4 addresses, they automatically create a 6to4 tunnel. While this was intended as a transition technology, we’ve found that it can cause issues by generating thousands of incorrect DNS entries and clients will mistakenly use these 6to4 adapters which can lead to connectivity problems. This added challenges to user buy-in because they believed this to be an IPv6 problem. As part of our implementation plan, we disable these adapters through Windows Group Policy.

Final Words

It is a great advantage to start as early as possible in your IPv6 deployment. This way you’re not forced to react to future requirements and you can do a controlled roll-out. From a budgeting perspective, it did not require any capital expenditures to implement IPv6, however with our MPLS project we had refreshed quite a bit of our hardware already. We did find some caveats with older supervisor modules that did not have all the IPv6 support we needed, so we had to withhold rollouts to certain segments until equipment could be replaced.

I also feel buy-in from the entire department is crucial as it creates clear expectations for the entire team and produces a cohesive message from help-desk and workstation support to system administrators and network engineers.

After the success of our IPv6 rollout, I’ve had conversations with network engineers from our sister institutions expressing interest in our IPv6 implementation. As collaboration between organizations who often have overlapping private IPv4 space increases, I see IPv6 becoming even more important.

Overall, we’ve seen no complaints from our end-users as we’ve rolled out IPv6 across both of our campuses. To be honest, very few people are even aware they have IPv6 access, which is exactly as it should be.

The post IPv6 is Not Optional appeared first on Team ARIN.

Read more here:: teamarin.net/feed/

IPv6 is Not Optional

By News Aggregator

By Mattias Lindgren

IPv6 Case Study from University of Colorado Denver in Denver, Colorado

The University of Colorado in Denver has a presence across two main campuses – the Anschutz Medical Campus in Aurora, Colorado and the Auraria Campus in Downtown Denver. Over the past year, I spearheaded the effort by the network team to complete a project of enabling IPv6 across both campuses. This involved extensive planning phases, going through project creation and approval, and finally the roll-out to our approximately 30,000 users.

Planning

We began by obtaining IPv6 space from ARIN. Lining up supporting documentation as well as approval of the contract amendment by the University purchasing team was a time-consuming portion of the planning phase. We knew we wanted to get ample space for the future, so we requested and got approval for a /40.

Make it Official

In discussion with my manager, we decided that the best way to get buy-in from the whole IT department was to make it an official project. This is key as it sets expectations for all the project members and stakeholders, and holds each team member accountable for completing their assigned tasks. We regularly hold a Project Review Board where members of the IT staff can present project frameworks. At one such meeting, I presented the benefits of implementing IPv6. Representing the network team, I argued that migrating to IPv6 is not optional, and adoption of it over the legacy IPv4 protocol is happening at other major universities and will soon become the dominant protocol on the internet. The earlier you adopt it, the easier implementation will be as networks only increase in complexity over time. We also discussed efficiency gains as complex NAT schemes would no longer be needed for pure IPv6 traffic. The project was approved by our CIO and it became an official mandate across central IT to deploy IPv6.

Project Rollout

We felt that an important part of the rollout was to educate the technical staff outside of the central IT department. To facilitate this, we brought in an outside vendor to give a half day presentation on IPv6 and what it meant. We had a good turnout of about 150 people.

Over the past couple of years, we implemented MPLS across both of our campuses. This made implementation of IPv6 easier as we could target well-defined chunks of our network for rollout. We saw a brand new IPv6 network as an opportunity to atone for legacy protocol sins, and designed the new dual-stack network with redundancy and scalability in mind from day one. Having already rolled out MPLS also presented its own challenges, as we found that some of the features present in the IPv4 world hadn’t yet become available on the IPv6 side. It has become a running joke that IPv6, at 20-some years old, is such a young protocol that we can’t expect vendors to support it just yet.

Since our legacy network routing table was sprawling with thousands of routes, we took the opportunity to summarize our IPv6 routing table as much as possible and it is currently less than 50 routes.

We decided on DHCPv6 for our addressing scheme, and began handing out addresses from our centralized DHCP and DNS appliance. This negated the need for disabling privacy-extensions and also made the client addresses a little more manageable.

One issue we’ve had for a long time is the presence of 6to4 adapters on our Windows computers. Since we still have workstations with public IPv4 addresses, they automatically create a 6to4 tunnel. While this was intended as a transition technology, we’ve found that it can cause issues by generating thousands of incorrect DNS entries and clients will mistakenly use these 6to4 adapters which can lead to connectivity problems. This added challenges to user buy-in because they believed this to be an IPv6 problem. As part of our implementation plan, we disable these adapters through Windows Group Policy.

Final Words

It is a great advantage to start as early as possible in your IPv6 deployment. This way you’re not forced to react to future requirements and you can do a controlled roll-out. From a budgeting perspective, it did not require any capital expenditures to implement IPv6, however with our MPLS project we had refreshed quite a bit of our hardware already. We did find some caveats with older supervisor modules that did not have all the IPv6 support we needed, so we had to withhold rollouts to certain segments until equipment could be replaced.

I also feel buy-in from the entire department is crucial as it creates clear expectations for the entire team and produces a cohesive message from help-desk and workstation support to system administrators and network engineers.

After the success of our IPv6 rollout, I’ve had conversations with network engineers from our sister institutions expressing interest in our IPv6 implementation. As collaboration between organizations who often have overlapping private IPv4 space increases, I see IPv6 becoming even more important.

Overall, we’ve seen no complaints from our end-users as we’ve rolled out IPv6 across both of our campuses. To be honest, very few people are even aware they have IPv6 access, which is exactly as it should be.

The post IPv6 is Not Optional appeared first on Team ARIN.

Read more here:: teamarin.net/feed/

The post IPv6 is Not Optional appeared on IPv6.net.

Read more here:: IPv6 News Aggregator

[session] Protocol Technology Impacting IoT Adoption | @ThingsExpo @Sigma_Designs #IoT #M2M #SmartCities

As popularity of the smart home is growing and continues to go mainstream, technological factors play a greater role. The IoT protocol houses the interoperability battery consumption, security, and configuration of a smart home device, and it can be difficult for companies to choose the right kind for their product. For both DIY and professionally installed smart homes, developers need to consider each of these elements for their product to be successful in the market and current smart homes.

read more

Read more here:: iot.sys-con.com/index.rss