Mar 6, 2019
Craig is with Ken and Phil, with Phil Harriman covering for Matt. They talked about smart assistants' ethical code, China mapping pigs' faces, and can you trust online reviews?
These and more tech tips, news, and updates visit - CraigPeterson.com
---
Related Articles:
Can You Trust Online Reviews? Here’s How To Find The Fakes
China’s Tech Firms
Are Mapping Pig Faces
---
Transcript:
Below is a rush transcript of this segment, it might contain errors.
Airing date: 03/06/2019
Moral Artificial Intelligence - China Mapping Pig Faces - Trusting Online Reviews
Craig Peterson 0:00
Hey, good morning, everybody. Craig Peterson here. I had fun this
morning, I was on with Ken Altshuler, and he had a guest host in
with him Phil Harriman, he's a former senator there in the state of
Maine. And we talked about a couple of things. And, and I managed
to rip him a bit. He's a big time supporter of left wing causes. I
don't want to call him a Democrat because he's not really a
Democrat. But you know, I got to rip him this morning, we talked
about smart assistance and whether or not they should have a moral
artificial intelligence and we got into the Chinese tech firms
mapping pig faces this morning. It was really kind of fun. You
gotta listen anyhow, hope you're having a great day as well. And we
did talk a little bit about online reviews, but not as much as
yesterday with Jim so here we go.
Ken Altshuler 0:53
Always on a Wednesday morning at 7:38. We welcome in our tech guru
Craig Peterson. Good morning to you Craig.
Craig 1:02
Hey good morning. Yeah I'm a little clogged up myself this morning
I was just coughing and hacking. Oh my. That time of the year.
Ken 1:10
New England March isn't the weather like 20 degrees below what it's
supposed to be?
Craig 1:15
I thought it was global warming. Is it not Ken?
Ken 1:20
It is. No, it's climate change. Climate change.
Craig 1:24
Oh so when global warming doesn't work that it becomes climate
Ken 1:28
Correct
Phil Harriman 1:28
No no no no. Its extreme weather.
Ken 1:32
Yes exactly. Extreme weather. Highs and lows and bad hurricanes,
bad tornadoes.
Craig 1:37
Here in Maine, I'm a firm believer in climate change. Absolutely.
We have four seasons at least every year. That's four times a
climate changes.
Ken 1:46
In Maine we have two seasons Craig, winter and Fourth of
July.
Craig 1:51
That's road repair and winter.
Ken 1:55
Exactly. So let's talk about some tests. So I have as you know,
Craig, Alexa. And let me, for all you people listening out
there, Alexa, pay attention. So we do something with these
smart assistants that kind of have a moral code to
follow?
Craig 2:05
Yeah, this is real interesting stuff. And I'm not sure exactly
which side of this I come down on. We've heard a lot about
artificial intelligence, right? It's been kind of the bane of
science fiction writers for ever. Now the carbon project in the
early 60s, a big movie that was out all the way through today. And
of course, even before that, but artificial intelligence. The whole
idea is that our computers are going to be smart enough to learn by
themselves make decisions by themselves. And today we have really
no artificial intelligence per se, but we do have what's called
machine learning. Where are our computers etc. can learn your
iPhone your newer iPhones have a machine learning chip built into
them. And that helps them learn a little bit about you. And every
time you use your Alexa or use your Google Home device, it's
learning a little bit more about you and what you do and how you do
it. So it's coming. It's kind of inevitable, we think and nothing's
past what's called the Turing test yet for those of us that are a
little more geeky about this but when we're talking about our Alexa
there's been a proposal that comes out obviously it makes sense to
have your Alexa be able to call the police, it can call your
friends anybody that's in your phone book. My granddaughter used an
Alexa we got for them just the other day. She was stuck outside the
house she locked herself out. And we had an Amazon Alexa in the
house that we've given them and she went ahead and yelled through
the window Alexa, call mom and of course it did. And Mom said call
me back. This is a bad connection. No, no, no, like outside the
house until mom knew. And she managed to get home with the key and
was able to to let my granddaughter in. Those are all really good
things.
But we've got some scientists now over in Norway at the University of Bergen spore just speaking last week at a conference out in Hawaii and they were saying that what should happen well they're saying all of our smart devices should listen to what's going on in a home and then use artificial intelligence to determine if maybe they should report the goings on in the home to the police, child Family Services etc etc. And this is at a conference of ethics and society in Hawaii. You know Ken when you you're talking about Orwellian I can't think of anything more Orwellian than this. Where we have been putting these devices into our homes into our cars now pretty much everywhere in our phones and they're talking about turning them into spies for who, right? How far can this go? I don't know Phil and Ken. This really scares me.
Phil 5:36
Well just listening to you describe the power in a device in your
home could have is frightening and in I would assume this has got
a
an onboarding or in boarding opportunity as well meaning that they
can put things into your system that gets communicated into the
house or control things that are in your house that you know that
you have no influence over?
Craig 5:54
Yeah, absolutely. Yeah, they could and they can upgrade them. For
instance, Google just got nailed because they've been selling this
home security device, and no one knew that Google had hidden inside
of it a microphone that Google just activated this month. So it's,
it's really, really scary. And we're talking about an ethical
conflict here between people in the family, you know, between mom
and dad in the kid, you know what happens when the kid is reaching
up to grab a hot pot of water off of the stove, and mom screams at
that the kid this is legitimately and now that little device in
your home makes an ethical moral decision and report you to the
police along with all of the recordings that devices made of
anytime that device thinks, well, this was marginal, but it's not
enough to call the police. And now all of a sudden the police have
completely out of context, all of this stuff that makes you look
really, really bad, you know, and then the ethical conflicts
between not only the members of family, but the manufacturer, the
shareholders, the manufacturers, programmers, the police
department, and, and having these universities say yes, indeed we
should have them recording and having other people jump on board
and agree with them. I think we were in for some rough road.
But take that to our cars, right. And in in Maine, we've got some of these artificial intelligence companies doing some development along with a mapping technology, much of which originated right here in Maine. Right? Who, if you're in an accident, whose fault is it? Is it yours? Because you own the car? Is it the people who wrote the software? Is it the car manufacturer who hired the people who hired the company that hired the people that wrote the software, you know, Ken, man maybe you should get out of the family lawn into some of the ethical law because for the next 20 to 50 years things are going to be just crazy in that side of the business.
Ken 8:07
We're talking to Craig Peterson, our tech guru.
Phil 8:08
Big Brother.
Ken 8:09
What are you talking about?
Phil 8:10
It was just frightening what he's saying.
Craig 8:14
BOC's on board. I'm sure so it'll be okay.
Ken 8:19
Is there a reason why you guys are ganging up on me today?
Phil 8:24
Yeah. Because I like you.
Ken 8:26
I think she does a very nice dance on rooftops. That's all I want
to say. We're joined by Craig Peterson our tech guru who joins us
Wednesdays at 7:38. Online reviews. Now, I went the other day to
buy a wrist brace for my right wrist. I have something when I
played piqua. And, you know, I went to read the reviews because how
else do I know which of these products have good? I mean, I assume
that those reviews were accurate? Not necessarily, huh?
Craig 8:53
Yeah, the legitimacy of these things comes into question and it can
be a very, very big deal because we're using them all the time I
use them on Yelp. I I go into a new town, I'm I'm traveling
somewhere. And I just sought my Yelp app and say, you know, where's
the nice restaurants in the area? And how about you guys? What do
you do? Do I tend to not trust reviews? When it's like, one or two?
or five? Or 10, right?
Ken 9:25
If it's like, if it's 400 or 500, I tend to pay attention.
Craig 9:27
Yeah, exactly. And that makes sense. And there's reviews
everywhere, right? Well, there are some tips of what to do, what
not to do when you're looking at these reviews order to judge the
veracity. And there was even a study done on this, like there's a
study on everything nowadays, I think, frankly. But Amazon, Yelp,
Facebook, Google, and it is easy for businesses and others to
purchase hundreds of reviews within days. So there's your 400
number can you know that they can all be false. And then the other
side of this is businesses will sometimes post negative reviews for
their competitors which is another big problem. So when you get
right down to it and looking at some of these studies and some 30%
according to the study of online reviews are fake reviews. NBC News
created a gardening business on Facebook they paid 168 bucks to
some online websites that promise suppose positive reviews and for
168 bucks, they got 1000 the likes and a few days after that they
got more than 600 5-star reviews.
So I think Ken we've got to be careful about this. I personally look at the reviews I look at what's been written and how it's phrased and stuff and I try and evaluate it from that but one of the easiest things you can do to figure out if reviews or false is took at the language that's being used. Because when you do purchase the reviews, you give a sample of what you're looking for these reviewers, these fake reviewers to say. And oftentimes they'll repeat it. Phil, you ready? They'll they'll repeat it. Just like democrats repeat the morning news bites, as opposed to talking about all day long.
Phil 11:27
Do you like us know, Kenny? I can't. I can't let this go any
further. Without shifting gears to the China's tech firms are
mapping. Wait for it. Wait for it, folks. They're mapping pig
faces. Alright. Take it away. Take it away Craig.
Craig 11:44
We'll make this one really quick because I know we're out of time
here. But yes, there is a very big problem in China right now. And
it isn't the pig farts. What we're talking about is disease. And of
course there's a lot of trans-species disease stuff that can
happen. Certainly with birds. Most of our flus come from birds. But
in this case, here's what's happening. China, China has been using
facial recognition technology for a long time to spy on its own
people. They've been doing that a lot in London, as well as other
places in the UK. But right now what they're trying to do is track
the pig farmers because many of the small pig farms are polluting
the environment. Yes, indeed, that is the truth. An AOC thing was a
bit of a joke there, but polluting the environment. So they want to
keep track of the pigs where they came from, what diseases they
might have, and they're doing it with facial recognition technology
of the pigs as well as they're listening to the pigs conversations
in order to determine if a pig might be sick because apparently
pigs talk differently when they're not feeling well.
Phil 12:57
So Craig, as we move on from this point forward when we refer
to AOC we can also refer to KPA. Kenneth P. Altshuler.
Ken 13:09
Craig Peterson joining us. He joins us, Wednesdays at 7:38. Craig,
thanks for joining us. We'll talk to you next Wednesday.
Craig 13:16
Hey Ken, thanks for being a good sport. It was kind of fun.
Ken 13:20
I don't mind all. Thank you guys. We're gonna take a quick
break.
Craig 13:26
Hey, I released module three yesterday we had a great coaching
call, live coaching call yesterday as well for everyone in the
course. So shout out to you guys. You should have gotten Module
Three, let me know if you did not. And then three, of course, we're
delving into network security stuff, what you can do and how to do
it. Take care guys, I'll be back tomorrow. I'm going to do a couple
of security things this week on. One on Thursday, one on Friday, so
keep an eye out for those as well in the podcasts. Thanks again.
Bye bye.
---
More stories and tech updates at:
Don't miss an episode from Craig. Subscribe and give us a rating:
Follow me on Twitter for the latest in tech at:
For questions, call or text:
855-385-5553