Thứ Ba, 1 tháng 5, 2018

Youtube daily to make May 1 2018

For more infomation >> HOW TO MAKE ROYAL ICING COOKIES | Amazing Cookies Art Decorating Tutorial Ideas | Cookies And Cake - Duration: 10:48.

-------------------------------------------

How to make a presentation? Effective presentation - Duration: 12:16.

For more infomation >> How to make a presentation? Effective presentation - Duration: 12:16.

-------------------------------------------

Trump Clears The White House In Staggering Move To Make Way For Special Guests He Wants To Pray With - Duration: 4:09.

For more infomation >> Trump Clears The White House In Staggering Move To Make Way For Special Guests He Wants To Pray With - Duration: 4:09.

-------------------------------------------

How to Make Dosa | Popular Indian street Food | BD daily - Duration: 2:02.

How to Make Dosa | Popular Indian street Food | BD daily

For more infomation >> How to Make Dosa | Popular Indian street Food | BD daily - Duration: 2:02.

-------------------------------------------

Trump Clears The White House In Staggering Move To Make Way For Special Guests He Wants To Pray With - Duration: 4:12.

Trump Clears The White House In Staggering Move To Make Way For Special Guests He Wants

To Pray With Sutherland Springs, Texas is a community that

will forever be changed and the members of First Baptist Church while battered and bruised,

they are far from broken.

The White House has invited members of the Texas church where a two dozen people perished

last year to Washington for the National Day of Prayer.

The prayer service at the White House comes just two days before the six-month anniversary

of the church shooting that took the lives of 26 members of the First Baptist Church

of Sutherland Springs.

The surviving members of the heartbreaking incident were children with one just a few

short weeks from birth.

Another 20 members were injured.

Multiple members of the same family perished and are still very much healing and grieving.

The members who will be traveling to the nation's capital for the prayer service include Pastor

Frank Pomeroy and his wife, Sherri.

Pomeroy states both he and his wife will be speaking at the White House the day before

and then praying at Thursday's service.

Sherri describes the invitation as an honor.

The Pomeroys' own 14-year-old daughter Annabelle was among the 14-year-old victims on that

fateful November 5th day.

Pomeroy also spoke of the church's plans to hold a memorial on this coming Saturday

to honor victims and survivors.

A groundbreaking ceremony will immediately follow for the new Sutherland Springs church

building to be built next to the old church.

The old church will currently and forever sit as a memorial for those lost loved ones.

The church will be built of stone, symbolizing the strength and grit of a community that

has been forever changed.While it was hard to comprehend the incident, Pastor Pomeroy,

who lost his own daughter in the shooting, reminded his congregation to put their faith

in the Lord.

"You lean in to what you don't understand, you lean into the Lord," Frank Pomeroy told

reporters during a press conference.

"I don't understand but I know my God does."

Also speaking along with shooting survivor David Colbath, his son Morgan Colbath, and

Sherri Pomeroy's sister Sylvia Timmons will participate in a prayer service with the president

for the National Day of Prayer, the newspaper reports.According to the Dallas News –

"We did not expect a personal invitation to the White House from the President,"

Sherri Pomeroy told the Express-News via text message.

"When I received that email, I confirmed the legitimacy of the invitation and accepted,

of course!"

The five Sutherland Springs visitors will be attending the National Day of Prayer's

evening services at the Capitol's historic Statuary Hall.

The Pomeroys will be interviewed by Pastor Ronnie Floyd, president of the National Day

of Prayer Task Force, and Frank Pomeroy will lead a short prayer — all live-streamed

across the country.

They'll also attend a private banquet the night before hosted by the Task Force.We are

very honored that we are invited to the White House and asked to pray for our nation in

this historic place," Sherri Pomeroy said.

"While this invitation is bittersweet because of the events that brought us to this place,

may we never turn down a forum to share the gospel of Jesus Christ wherever it takes us."

The National Day of Prayer will be held in Washington DC's historic Statuary Hall,

and is an annual observance held on the first Thursday of May, inviting people of all faiths

to pray for the nation.

It was originally created in the year 1952 by a joint resolution of the United States

Congress.

It was signed into law by President Harry S. Truman.

For more infomation >> Trump Clears The White House In Staggering Move To Make Way For Special Guests He Wants To Pray With - Duration: 4:12.

-------------------------------------------

Autopilot Money Making System 2018 - How to Make Money Online - Duration: 3:41.

For more infomation >> Autopilot Money Making System 2018 - How to Make Money Online - Duration: 3:41.

-------------------------------------------

4 Ways to make a man feel special - Duration: 1:19.

For more infomation >> 4 Ways to make a man feel special - Duration: 1:19.

-------------------------------------------

Bees struggle to make it through tough spring - Duration: 1:16.

For more infomation >> Bees struggle to make it through tough spring - Duration: 1:16.

-------------------------------------------

How to Increase Retail Sales: Three ways to make your Facebook ads more effective - Duration: 3:52.

Hi Guys, it's Greg here from Rain Retail and in this video

we're going to show you how to increase retail sales with three quick tips to make your Facebook ads more effective.

According to Adespresso, "since 2015

worldwide social ad spent has increased by several billions of dollars, with Facebook ranking in the vast majority of that number."

Furthermore, as

indicated by LinkedIn,

"Marketers are continuing to increase their spending on Facebook ads as the platform becomes more and more effective." All this new advertising

translates to stiffer competition when it comes to attracting user attention. In general

it's said that you only have three seconds to get a users attention before they scroll past your ad in the news feed.

So here are three quick tips from Cox Media Group to get users to stop, read, and click your ad.

Number One:

Select photos that tell a story.

Since Facebook announced its preference for visual content, a change based on existing user behavior,

photo ads have increased in popularity. To stand out from the competition look for pictures that tell a story.

Nike excels at this by creating ads with images of athletes in mid action like the one here of LeBron James in midair

during a monstrous dunk. When users see these powerful images

they're more likely to stop and linger

because the photos stand out and help to establish a real connection with them. Another thing to consider is that there is a

particular group of cells in our brains that fire only when we see a face. According to Adespresso, "it is a phenomenon

that's deeply ingrained in our brains, a vestige of our primal beings-- so use it in your facebook ads!"

Number Two: Take advantage of video. Video is growing in popularity as an advertising medium because of its effectiveness.

According to Forrester, "One minute of video is equivalent to 1.8 million words of text in the message

communicated." In addition to conveying a lot of content in a short amount of time,

consumers tend to remember the videos they watch. According to the Online Publishers Association,

"80% of users recalled watching a video in the past month and

46% of them followed up the message with looking for more information or visiting the advertisers website."

Which means that video is clearly engaging consumers and

forging connections with the brands that are taking advantage of this popular advertising media. When creating videos for your audience

pay special attention to format,

storytelling, and

platform. Your brand or product should be included in the first few frames

and you shouldn't make people click to get the message or call to action. Lastly, make sure your video

demonstrates how your product or service solves the customers problem.

Number Three: Prevent ad fatigue.

Facebook ad fatigue occurs when your ads target audience is shown your ad too many times and

your click-through rate drops as your frequency rate, how many times

it's seen,

increases. To make sure your ads don't fall into the ad fatigue rut you need to refresh your ad

creative every month. Use Facebook Ads manager to monitor frequency actions and CTR so you know which ads are performing at the level

you'd like them to and which ones need to be tweaked or

Rotated out. And remember the most effective Facebook ads are those that are able to make a connection between a brand and its intended audience.

Using these three steps will fine-tune your Facebook ads and help you to keep your audience's focus right where you want it.

Guys thanks for watching this video. To see more tips and strategies aimed at helping you succeed in retail, just click the link below. And

if you found this video helpful, then please give us a big thumbs up and be sure to hit that subscribe button.

I'll see you next time.

For more infomation >> How to Increase Retail Sales: Three ways to make your Facebook ads more effective - Duration: 3:52.

-------------------------------------------

Underground Root Cellar Easy Plans-Unique Root Cellar-How To Build A Root Cellar In Your Basement - Duration: 2:44.

Have you ever thought about living without electricity, internet or mobiles?

We can guarantee that many people can never imagine this kind of scenarios.

However, there are chances that this type of conditions arises in your life due to flooding,

tornadoes, draught or even war.

How could you survive in this type of dangerous condition?

We believe you should stay prepared by learning the essential skills needed to deal with these

disasters.

Easy Cellar has a secret that assisted our grandfathers or forefathers to stay protected

against the economic crises, famines, and other similar conditions.

Easy Cellar is an incredible program that will prepare you to survive in dangerous conditions

with a minimum supply of the resources and help you protect your family.

The author of this guide is Tom Griffith has great experience and skills to deal with these

types of scenarios.

Author Tom has tested this program for making sure the readers can survive in the worst

conditions.

Moreover, your children can follow this program and find out the ways to make their own resources

without any need to search in dustbins.

Furthermore, it is boosted by a 60 day money back guarantee that make sure you will get

enough time for reading this program and know its authenticity.

Click the link in the description below this video for details.

After reading this guide, you will learn about the ways to preserve food and water all year

round to stay prepared for the worst conditions.

In short, we can say that the quality of your life will improve to a significant level.

The author has shown the drastic differences between the lifestyle of the modern man and

their ancestors.

As we all know, the people in the past lived their life in a practical way without the

technological facilities in the present world.

The main motive of Easy Cellar is to tell you the secrets that can make to easily withstand

emergency conditions and protect yourself and your family.

Easy Cellar can be easily defined as one of the best and comprehensive life-saving guides

available in the market.

It is based on the principle to help you stay prepared for dealing with all types of situations.

You will be able to discover the secrets to creating cellars in your backyard where you

can keep food and water.

They are created by using a different technique so that your resources stay unspoiled for

many months.

These bunkers can be built by using a minimal amount of material and effort.

There is also a step by step video guide included in this guide to create the Easy Cellar.

It also explained 15 natural remedies for radiation that will be beneficial in many

different manners.

Your body will get the essential nutrients needed for working properly in the emergency

conditions.

The guide will make sure that you are never short of eating and drinking resources even

in the worst scenarios like a flood, tornadoes, wars, and so on.

Click the link in the description below this video for details.

For more infomation >> Underground Root Cellar Easy Plans-Unique Root Cellar-How To Build A Root Cellar In Your Basement - Duration: 2:44.

-------------------------------------------

Malloy set to make major jobs announcement - Duration: 0:14.

For more infomation >> Malloy set to make major jobs announcement - Duration: 0:14.

-------------------------------------------

How to make Paper Hat for Girls (Origami Hat) - Adorable Paper Crafts - Duration: 5:01.

Dear Paper Crafts Lover, Welcome to Paper Hat making instructions

You will need 2 pieces 20cm x 10cm paper

Please follow the instruction step by step for making a best origami hat

Make two unites same

Now you will need 0.7cm x 25cm paper and some decorative pearl.

For more infomation >> How to make Paper Hat for Girls (Origami Hat) - Adorable Paper Crafts - Duration: 5:01.

-------------------------------------------

[WancoBeads] How to make Mini Pig - Duration: 25:35.

Beads abacus 4mm yellow 67 Beads abacus 4mm pink 4 Beads round 4mm black 2 Nylon string 0.23mm(dia.) 120cm 1

At first, Add 5 beads(pink, pink, pink, yellow, pink) to red nylon string

It crossed in the last addition beads

The number of beads that are in a ring is 5

Pick up 1 bead(pink) in red nylon string

Add 2 beads(yellow) to blue nylon string, It crossed in the last addition beads

The number of beads that are in a ring is 4

Add 1 beads(yellow) to blue nylon string, Add 3 beads(black, yellow, yellow) to red nylon string, It crossed in the last addition beads

The number of beads that are in a ring is 5

Add 2 to blue, cross

The number of beads that are in a ring is 3

Pick up 1 in blue

Add 3(yellow, black, yellow) to red, cross

ring (5)

Pick up 2(pink) in red

Add 1, cross

ring (4)

Add 3 to red, cross

ring (4)

Pick up 1 in red

Add 2 to blue, cross

ring (4)

Pick up 1 in blue

Add 2 to red, cross

ring (4)

Pick up 1(black) in red

Add 2 to blue, cross

ring (4)

Pick up 1 in blue

Add 3 to blue, put the left ear

Add 3 to red, cross

ring (5)

Pick up 1 in red

Add 3 to blue, cross

ring (5)

Pick up 1 in blue

Add 3 to blue, put the right ear

Add 3 to red, cross

ring (5)

Pick up 2(black, yellow) in red

Add 1, cross

ring (4)

Pick up 1 in blue

Add 3 to red, cross

ring (5)

Pick up 1 in red

Add 2 to blue, cross

ring (4)

Pick up 2 in blue

Add 2 to red, cross

ring (5)

Pick up 1 in red

Add 2 to red, cross

ring (4)

Pick up 1 in red

Add 3 to red, cross

ring (5)

Add 3 to blue, cross

ring (4)

Pick up 3 in blue

Add 1, cross

ring (5)

Pick up 1 in blue

Add 2 to red, cross

ring (4)

Add 3 to red, cross

ring (4)

Pick up 3 in red

Add 1, cross

ring (5)

Pick up 2 in blue

Add 1, cross

ring (4)

Pick up 1 in blue

Pick up 1 in red

Add 2 to blue, cross

ring (5)

Pick up 2 in blue

Add 1, cross

ring (4)

Pick up 4 in blue

Pick up 1 in red, cross

Pass the red nylon string to the position of the tail

Add 1 to red, put the tail

Pass the red nylon string to the position of the right hind leg

Add 1 to blue, put the left hind leg

Pass the blue nylon string to the position of the left fore leg

Add 1 to blue, put the left fore leg

Add 1 to red, put the right hind leg

Pass the red nylon string to the position of the right fore leg

Add 1 to red, put the right fore leg

Pass the red nylon string to the position of stomach

Tie the nylon string

After connecting, through the nylon string to 2 or 3 beads

Cut a little pull state the nylon string

It was completed

Use 3mm of acrylic beads

For more infomation >> [WancoBeads] How to make Mini Pig - Duration: 25:35.

-------------------------------------------

Marble Cake Recipe - Duration: 13:01.

In this video I'm sharing with you how to make an outstanding marbled cake

using my homemade chocolate buttermilk cake and my vanilla or yellow cake

recipe. Yiu'll see one way to fix a lopsided cake and fill the layers with a

simple whipped cream with sliced peaches.

I'm going to show you today how I make a marble cake. First make sure that you

prepare your pans ahead. Now I do have a video on preparing cake pans, but I will

just tell you what I do, I put Crisco shortening around the entire pan, and you

can use a spray if you want, and then I put a little dusting of flour in there,

and I get all the excess out, and then I cut a piece of wax paper to fit the

bottom. And that way your cake is guaranteed not to stick. Now I made two

batters I made a yellow cake and I made a chocolate. Now the chocolate is on my

channel. It's my chocolate buttermilk cake. The yellow cake is not yet on my

channel. it's on my website weddingcakesforyou.com and you can also use my one

two three four cake which is also called vanilla cake, and it's also called my

wedding cake recipe. So the first thing you want to do is is fill your pans a

little bit with the yellow or the white cake whatever one you decide to use.

And you're going to do that for each cake.

And I do get a lot of questions about marble cake, so hopefully this will help.

Then I'm going to take the chocolate then I'm just going to put a little bit

in there.

And I'm gonna take a knife,you can use a spoon or a knife, I'll use a spoon. And

just run the spoon through and swirl the chocolate.

Don't overdo it though, because you don't want it to be too muddy. That's good.

This pan was a little bit less full so I'm going to use the rest of the white in

here.

So that's what it looks like, and I'm gonna pop it in the oven. In this case

I'm gonna do I used 325 most of the time with my oven

for my cakes and it's a convection oven. And bake those until they are somewhat

firm to the touch. So as you can see when I push down on it it bounces back a

little bit. And you can also use the toothpick method which is you put a

toothpick in the center and if it comes out clean that means your cake is done.

I'm going to add a little more chocolate to this one.

Okay here we go. I'm gonna pop them in the oven. Cakes are

done. Here's what they look like when they come out. Beautiful. Now if it sinks

a little bit like this one here is sunk in a little bit, it's fine. If it was

sunken deep then I would worry but it's fine. These are going to be delicious.

For the marble cake I'm using a whipped cream and a peach filling. I'm

using some peaches that I got at the grocery store in a jar and I'm gonna go

ahead and strain out the juice. The peaches are not in season right now so I

went with the nice,, I went with a really good-quality cling peach that's already

sliced and stored in syrup. I'm gonna set that aside. For your whipped cream

simply just put your whip on there. This is for a ten inch cake.

I'm going to go ahead and do the whole two cups it's gonna be probably more

than I need.

About a half, I have 1/3 cup of sugar here confectionery sugar. I'm gonna put

like half of that in there and a little bit of vanilla. I'm just gonna put a

couple of dollops in there and whip that up. I tasted it and it needs more

sugar so I'm putting the full 1/3. I just cut my peaches and what I did, sorry the

camera wasn't on and I thought it was. I just cut them in, here's what the

original thickness was, and I just cut them in thirds basically, lengthwise. and

then I just cut them in half. So here's the size and the thickness that I want

for these peaches. I cut this one out of a bigger board because

I didn't have any in the house. Got my cakes here. So the cakes

were frozen, there's they're not solid frozen they're just very very chilled.

They're a little bit frozen but not completely. They've been in there

overnight. I'm gonna leave them as a thicker layer and just do one thick

filling. And the sticky side is gonna go down. This is the sticky side that's on

top and it's now at the bottom. You can always go back in trim if you need to

trim the cardboard which I will probably do.

I'm just gonna get this one ready. Just gonna see how this might best fit. My

cakes came out a little uneven. So that looks pretty good right there. And what I do

a little trick, this is kind of a little tip, it's still not perfect. Let me see if

I can get that better. It's definitely lopsided. In a case like this I would

either trim the cake or I would add a little more butter cream on the side

that is off, which is what I'm gonna do, So just to mark it I'm just gonna put a

little line here with the knife.

Little notch. this is where I want it to be. Where I want to match it up. Okay so I

have my... You want to put buttercream into a pastry bag. I'm using a I think it's a

4B tip so any kind of a big tip you don't have to it doesn't have to be a 4B.

It could be just a round tip and you're just going to put a border all the way

around. Now I'm going to make it I'm going to come back around make it even

thicker on this side because I have a little bit of a lopsided cake here. And

then I'm going to fill. So you want to fill this area here with whipped cream.

Alright I'm gonna make this a little higher because I want to put more

whipped cream in there. And this dam will help keep the whipped cream in place.

So I've got my notch I'm going to line them up.

And look at that came out nice and straight.

So here's the finished cake. I'm working on creating a tutorial on how to do

these textured buttercream paintings on the side of a cake like this one and

this beautiful soft pastel version. Thank you so much for being here I really

appreciate it and if you haven't done so already subscribe to my channel and

click the little notification bell so you'll know when I upload all my videos.

And I also want to remind you I do have a book that's called Wedding Cakes with

Lorelie Step by Step and it has all my best recipes as well as how to make a

wedding cake. So thanks again for being here and I'll see you soon bye bye

For more infomation >> Marble Cake Recipe - Duration: 13:01.

-------------------------------------------

How To Make An Eye Dropper Water Purification Tips for Emergencies - Duration: 1:20.

how to make an eye dropper water purification tips for preppers emergencies how to purify water with bleach

hi it's AlaskaGranny you know how we're told a few drops of bleach added to

water can help purify it well what if you're in prepper an emergency situation and you

don't have an eyedropper how can you make one to purify water in emergencies with bleach it can be as simple as a ball

of cotton string and some kind of little lid cut off a piece of the string wind it

around into the lid and let a little piece of it hang over now fill your lid

with bleach and just be patient you can see if you will just be patient the bleach

will run down the string and drip one drop at a time right off the end of your

cord 2-4 drops of bleach per quart of water to purify wait 30 minutes

make sure when you're prepping for emergencies stocking your bug out bag emergency kit prepper stockpile along with your bleach have

an eyedropper or some cotton string or even jute twine to purify water one of the great ideas

about being a prepper is using the things you have to make the things that you need

learn more at alaskagranny.com please subscribe to the AlaskaGranny channel

For more infomation >> How To Make An Eye Dropper Water Purification Tips for Emergencies - Duration: 1:20.

-------------------------------------------

Тайны Коко-Как резать бумагу по-мексикански/Субтитры/Coco-How To Make Papel Picado - Duration: 2:20.

For more infomation >> Тайны Коко-Как резать бумагу по-мексикански/Субтитры/Coco-How To Make Papel Picado - Duration: 2:20.

-------------------------------------------

District attorney said he never asked to make changes to police report - Duration: 1:30.

For more infomation >> District attorney said he never asked to make changes to police report - Duration: 1:30.

-------------------------------------------

Watch how to make the Pastel Flower Bead Necklace by Fusion Beads. - Duration: 7:02.

- [Cody] Hi, welcome to Fusion Beads.

Today, I'm gonna show you how to make

this cute little pastel flower necklace.

So, let's get started.

(upbeat music)

OK, this necklace features

five millimeter Swarovski flower crystal beads

in violet, aquamarine, peridot, jonquil, and light rose.

We have some size 11 silver plated metal seed beads.

Then we have this .8 millimeter silver beading chain.

I've got a 12 millimeter silver lobster claw clasp.

And then I have these little silver,

thin fold-over cord ends.

I've got six millimeter eight gauge silver closed jump ring.

And then I've got two five millimeter

19 gauge silver open jump rings.

For my tools today, I will just be using

two chain nose pliers.

You can find links to all these beads and supplies

in the description below.

So, let's get started.

OK, so to begin,

I go ahead and I take my chain,

and the length of chain that you will have

is already cut to 18 inches, so you don't need to cut it.

So to begin, I'm just gonna pick up one

of the silver seed beads and slide it on the chain.

So tiny, sometimes it's hard to get them started.

There we go.

I'm just gonna kind of bring it to the middle.

I haven't put my cord in on the other side,

so I wanna be careful not to pick it up,

so that it slides off.

Then I'm going to pick up one of the violet flowers,

and then I'm gonna pick up another seed bead.

'Cause I'm gonna use these little seed beads

as a little spacer between each of my flowers.

And then I'm gonna pick up the aquamarine bead

and just slide it on here.

And then another spacer.

And then the peridot.

And then another spacer.

Sometimes these little guys are really tough to get

because they're so tiny.

There we go.

And then jonquil.

Little flower bead.

And then one more spacer.

And then the last flower bead, which is the light rose.

And then you end with a spacer.

Alright, so now I've got them on there.

And it looks pretty cute and sweet.

So now I'm just ready to do my cord ends.

And to do those, they're just the little fold-over cord end.

And so, it's kinda hard to see.

But um, you'll put in your chain

and then you'll just fold the sides over

and it will secure your chain end there with no glue.

So I'm going to grab my chain.

And a little trick that I do sometimes

just because they are kind of hard to hold on to,

is that I'll slide my chain in there and then

I put just a little bit of it under.

Like above where the folds are at

and I grab it with my thumbnail there

to kind of hold on to it.

And then I'll take my chain nose pliers

and gently on one side only,

I just start to fold over.

Kind of start towards the back and move to the front there.

So you get one side folded down.

And then you go over to the other side

and then you fold it over.

OK.

Just wanna make sure it's really in there.

So I'm just gonna give it a good smash

from this top side here.

So now it's in there.

There is a little tiny bit that's like hanging out the side.

So I'm just gonna very carefully use my wire cutters

and trim that off, but do not cut off your loop.

There.

So I've got one side done.

Now, I'm gonna just repeat that process

on the other side here.

You don't have to have that little extra bit to hold on to,

it's just easier for me to do.

You don't have to do that, you can just put it in there

and then fold it over.

Again, in the same fashion just fold over one side.

And then, you fold over the other side.

There.

And this has just a little tiny bit of extra chain there.

So I'm just gonna trim it very, very carefully.

Not to cut the loop, just to cut the little bit

of chain that I had extra there.

Alright, so I've got my cord ends on.

Now I'm ready to add my jump rings here

and my lobster claw clasp.

So, I'm gonna go ahead.

I'm using my two chain nose.

I'm gonna open the jump ring

by pulling one side towards me and one side away from me.

Putting it on the cord end here.

And then attaching the lobster claw clasp.

And then, close it.

And then, just in the same fashion,

I'm gonna add the closed jump ring on the other end.

So string on the closed jump ring,

and then attach the jump ring to the other cord end.

And then, close it.

There.

You can find all the products and tools

for this inspiration design at FusionBeads.com.

(upbeat music)

For more infomation >> Watch how to make the Pastel Flower Bead Necklace by Fusion Beads. - Duration: 7:02.

-------------------------------------------

CCSU to make changes to sexual assault reporting - Duration: 2:42.

For more infomation >> CCSU to make changes to sexual assault reporting - Duration: 2:42.

-------------------------------------------

Acquiring and Aggregating Information in Societal Contexts - Duration: 58:33.

>> I'm very happy to have Bo Waggoner here today.

He's currently a postdoc at Penn.

Before that, he got his PhD at Harvard.

He went to undergrad at Duke.

He's actually been interned at MSR in New England,

and interned twice at Google.

He is also a really serious runner.

He even won ACC award for

Athlete of the Year.

So, let's hear what he wants to talk to you about.

>> Yeah. Thanks very much. Thanks for

coming to the talk and putting this together,

especially thanks to Alex for making it happen.

Yeah, I'm really excited to get to talk to you guys.

So, the title of the talk is,

Acquiring and Aggregating

Information in Societal Context.

So, I want to think about a paradigm

that's very familiar to many of us,

especially in Machine Learning.

There's a bunch of information,

we feed it into our algorithms and they

output predictions or decisions.

So, it might be data about the weather,

and the output might be some classifier that

classifies weather as indicating sunny,

or rainy in the next day or whatever.

But while we traditionally like to

analyze our algorithms with this perspective,

actually, there's this recognition that

they live in a context that's bigger than this.

So, context of people.

In particular, the data that's coming

in is data that comes from people.

The data maybe about people in some way,

and the outputs, the decisions,

or predictions, also are affecting people.

So, this raises some really interesting questions.

I think both from a societal standpoint,

how should this be designed when we're

taking into account this context?

There are often really nice technical questions,

which I really like, coming

from more of a theory background.

So first, I just want to tell you briefly

about the kinds of questions that come up,

and then which ones I'll be

talking about today in more detail.

So, I like to

break the challenges down into three rough categories.

So, one is strategic behavior of agents.

So, when you take something like

machine learning type system,

and put it in the context with people,

they may be behaving strategically.

So, this input data you're getting,

for example, might not be the real or true data.

As one example of how they might behave strategically,

they could just misreport data.

They might do that because they care about

the outcomes, for example.

Another big concern is privacy of this data.

So, learning algorithms traditionally

weren't designed with any sort of privacy in mind,

but we often want to now redesign them,

or add this in because people care

about this data may be sensitive or private in some way.

This is a whole field of differential privacy,

and I'll talk a bit about that.

A final aspect, but one I won't get to today,

is this idea of fairness for decisions.

So, the decisions that

the algorithms make affect people in certain ways,

and the traditional approach of just trying to get

high accuracy may have ethical problems,

when they're put in that context of people.

So, it's a really interesting area.

There's this whole idea of fairness in Machine Learning,

but I won't get to that today.

So, instead I'll be talking about

projects from these first two categories.

Okay. So, the outline for the talk.

I'm going to tell you about a project dealing with

strategic behavior in Machine Learning.

So, how to do some learning in a situation where

agents are manipulating the data.

We'll spend more time on that,

and we'll spend a little time talking about

challenges that come up when we try to do

private Machine Learning and a project about sort of

translating theory

of private Machine Learning into practice,

and some challenges that arise,

then some stuff on future directions.

So, that's the outline. Please definitely stop me,

questions, discussion, would all be great.

I'd rather have some really interesting discussions here,

and have a little less time here,

if there are interesting questions to talk about.

Okay, good.

So, here's my overall picture of learning algorithms,

or algorithms in general in societal contexts.

For this first project, I want to focus

on the strategic behavior aspect.

So, this question of how do we do learning when

these people are being strategic about

the kind of data they give to the algorithms.

So, we're going to look at a particular setting,

and I want to motivate it with an application,

an example of detecting credit card fraud.

So, a high level example,

but imagine that over a series of rounds,

agents make a transaction,

and the system sees some data about this transaction,

and it wants to know if that's a fraudulent transaction,

or an honest one.

You can think of maybe

some feature vector is describing the transaction.

So, maybe it is a vector of real numbers or something.

So, the system is going to classify

it as being honest or fraudulent,

and has some classifier,

I'll be using beta for this.

Eventually, the system will learn what

the true status of that transaction is, let's say.

In fact, let's just

simplify things down and assume it learns very quickly,

like immediately, if it was an honest transaction,

or if it was fraud, because it gets

feedback from the account holder.

Then it wants to

update and learn better

classifiers based on this feedback.

So, what's the big challenge?

So, the challenges is that the fraudulent agents in

these systems are going to be

responding to the classifier that you pick.

So, that's why I put this hat on top of XT,

to denote that this future vector we're seeing,

or this transaction we're seeing,

may not be the truth that

these fraudulent agents would

like to have purchased, for example.

So, if there weren't some spam feature,

they just purchased whatever their favorite thing is.

But they're going to be- Instead of spam,

I should say fraud detection, right?

If there weren't fraud detection,

they just purchased whatever they want,

but they're going to be responding.

So, depending on what

the algorithms classifier looks like,

they're going to be buying things in such a way

to try to evade the classifier.

So, they're going to choose

these transactions strategically

depending on what classifier

you're deploying in the first place.

So, here's a picture of what that might look like.

So, if you picture each of

these transactions as a point in space,

we're modeling it as a point in R or d,

and you may have

all these blue points representing honest transactions,

and these red points,

I want you to think of these as the transactions

the fraudster would make if there were no consequences.

So, they'd like to buy, let's

say like chocolate ice cream,

or they'd like to buy a boat, or whatever,

but they're not going to

necessarily get away with that

because there's a fraud detection scheme.

In particular, if we had all this data,

we could deploy some classifier.

It might not be perfect,

but it would classify honest transactions from fraud,

and it would catch this fraud.

Yeah, there are obviously some mistakes.

So, I had a fun incident,

where my credit card was declined making a purchase,

and the location was the city I was living in,

and the purchase was running shoes.

So, I really don't know- if you know me,

that's like most of what I spend my income on.

So, I don't know what was-

These things aren't perfect in practice.

But life is even harder

because if the fraudsters

know that this is the classifier you're deploying,

and let's say in the worst case they

know or they have some good idea,

they're going to be responding

and modifying their transactions.

Okay. So, what you're going to see,

is going to end up looking more like that.

So, maybe instead of chocolate ice cream,

you have to buy strawberry,

maybe instead of a boat, you have to buy

a kayak, or whatever.

But you modify you transaction,

some way they try to increase

your chance of getting past this filter.

So, the filter has to be designed,

this detection classifier has

to be designed with these reactions in mind.

So, this will lead to some,

in game theory, the idea of the equilibrium.

But even knowing that fraudsters are best responding,

you want your classifier to be the optimal thing.

So, this is the perspective people have been taking.

So, there are a couple of papers on this problem.

Let me tell you a bit about the approach that they take.

So, at a high level they'll say,

suppose we have a bunch of data

drawn maybe IID from these populations,

some of it is honest, some of it is

fraud, it's all labeled.

Usually, these papers will suppose they

know the preferences of the fraudsters.

So, for example, one assumption might be

that there's a vector or direction,

and the fraudsters all want to

modify that direction or something like this.

So, using that information,

they can compute this classifier that does

well even given that they're best responding,

the fraudsters are best responding.

So, given that, learn some hypothesis,

some classifier, and it should do well.

So, the work I'd like to tell you about,

is going to try to weaken some of these assumptions.

So, one thing we want to think about

is an idea of agents arriving online,

and without having necessarily a distribution.

So, if you're familiar in Machine Learning,

often there are these two paradigms.

There's a IID data,

and then predicts well on that distribution.

But we instead want to think of

this worst case possibilities.

So, some transaction just arrives.

It's not necessarily from a distribution,

but in the long run we'd like to perform well.

So, after you sum notion of

regret, which I'll tell you about.

Another one is, we

really want to think about these agents,

especially the fraudulent agents.

It's heterogeneous and different.

So, they don't necessarily all want the same thing.

One might want ice cream,

one might want to sail across the river, or whatever.

A final important point is

that we didn't want to assume that

you can see the preferences of these fraudulent agents.

So, we don't necessarily know what they want.

We just see the transactions they tried to make.

So, we maybe saw them buying strawberry ice cream.

We'd like to try to infer from that,

did they actually want chocolate,

but they've modified it because of our classifier?

Or like do they actually like strawberry?

How can we use that to design better classifiers?

I hope that point makes sense.

Okay. So, that's the high level goal.

So, the paper is called strategic classification from

revealed preferences with some wonderful coauthors

at Penn or formerly of Penn.

Okay, and there are

three aspects to it that I'd like to discuss with you.

So,the first is how do we

model these aspects that I just talked about.

So, we have this idea of online arrivals and

limited feedback or limited knowledge about the agents.

And then what should we be shooting for?

And I think there's an interesting question there,

how do we measure how well we're doing?

And I'll explain why it's not super clear.

And then finally, how do we actually

try to solve the problem?

And for this talk with you today,

I actually want to,

I just want to spend a lot of time on

these because I think they're really interesting

and just a little bit of time on our solution

relative to maybe a normal outline.

So, it'll turn out that we can reduce

to known algorithms or

at least leverage them very heavily.

But we put a lot of

thinking and effort into this modeling questions and

I think that's also really

interesting for where to take future directions.

So, question so far?

I'll define the model next. Okay, awesome.

Okay, so in the fraud example I gave you before,

you already saw basically the model.

We'll make it a little bit more precise.

So, each agent that arrives,

they're either a fraudster or

an honest agent and they have

some what I'm saying true features,

which is the transaction they really want to make.

That's what we're thinking of it as.

Okay? And they have some true label.

Which is let's say one if they're

honest and minus one if they're not.

And finally, we assume they have some utility function,

which is measuring what's their utility

for facing this classifier beta or

this system described by beta when

their true purchase that they like to make is

x and they make a purchase x hat.

There are features that they present

to the classifier actually x hat,

so they've modified it in some way.

Okay.

So, what is this trying to capture?

Well, the utility intuitively has some tradeoff,

where if they're getting classified as

fraud or spam for certain,

then they're not going to get much utility.

But on the other hand, if they make

some transaction that they didn't

want in the first place and it goes through,

that's not all that useful to them either,

like strawberry ice cream, right?

So, apologies to anyone who actually likes,

I think I'm safe.

Okay. Good. So, we're

going to assume for this work that

honest agents don't act strategically in this way.

They just make the transaction they want to make.

So, they always said this x hat t.

This transaction actually ends up

happening just equal to the one they wanted.

They're not thinking about fraud detection schemes.

But we're assuming these fraudsters

do act strategically. Yes.

>> So are the fraudsters are being myopic or not [inaudible]

>> Yes a great question.

So, we assume they act myopically.

Yes.

So, we're going to assume that they

best respond with this utility function.

Yes.

>> Are we going to assume with their IIDs?

>> Yes.

So we're not going to assume they're IID for this.

Yes.

I mean I guess it would cover

that case and you could do some result for that.

Yes. Good yes. So we do assume

that the fraudsters will choose

this x hat t myopically

to maximize this utility function.

Yes.

Good. Okay. So we have this online model.

So there's a sequence of rounds.

The system has a current classifier any time,

the agent at that time

arrives and we assume the system observes okay

this transaction that they made or tried to

make and then it will learn this true label.

So for example if it was a fraudulent transaction,

even if they let it through,

the person who owned the credit card

will let them know that

it's being used wrong and they'll flag it as fraud.

Okay, so the system gets to learn

the true label for all the transactions.

Is that clear? It's okay.

Okay. And then we're going to use

some loss function to

measure how well the system performed on this data point.

Okay. So it's a loss as a function of

this classifier that you

were deploying or this hypothesis.

And then the transaction that was made and the label.

Okay. So this is just some function measuring performance

and natural one you might think of is that

this classifier is taking in X hat and just

saying one or minus one honest or fraud.

Okay. And then the loss could

be a loss of zero if it got it correct.

Right. So if it labelled it correctly according to

IT and a loss of one if it got it incorrect.

Okay so that's a natural loss function.

We won't end up working with it because

it's not computationally very

tractable but it's a good one

to keep in mind for the model.

And by the way so this model that I'm

presenting to you just as a heads up is very general.

We're going to end up solving some special case with

some strong assumptions but I think

it's fun to think about this general case so that in

future work people can try to solve more aspects of it.

Okay good. So the data points arrived,

the system had some loss,

the updates to a new classifier,

the next data point arrives.

Okay. Any questions about the model?

Yes.

>> [inaudible] with credit

cards [inaudible] particular person that

can be used in purchasing and the like [inaudible]

all the non- fraudulent?

>> Yes. That seems fair. One-

>> [inaudible].

>> Yes. So the formalism

that at this level I'm talking about can almost

even capture that because this xt

could include the cardholder's history,

somehow it can include their location or other things.

So it could probably include all that in this formalism.

Yes. In practice you might suppose you want to

just separate these problems like for each person.

>> It's like you give up the data points including like

features of the person

and what they're trying to purchase.

>> Yes, because we know

this credit card is associated

to this person or something.

So the other application we're thinking about as

motivation which you may have heard from

my verbiage is spam detection of emails.

So, honest people just

send the emails they want to send and

spammers would like to send

some particular spam email but it will get filtered out.

And so they try to modify it in some way and

maybe it's less effective but it gets

through the spam filter or something.

Yes. And again there you can think about

personalization versus everyone's data together.

>> [inaudible]

>> Yes, great question. Yes. So, does

the fraudster know the exact classifier that we're using?

So, in principle, there's no reason they would,

like if you're thinking about

credit card detection or spam filtering.

There's no real reason that we would publish,

for everyone to see, the exact algorithm

we're using at the current time.

But we're thinking of

that a little bit as security through

obscurity and that that's not

a good defense to assume

that they don't know the classifier.

You might be worried that,

if you think about like spam emails or something,

they send a ton of spam,

they can quickly start learning what

your classifier's doing at the current time.

So, this is a sort of worst

case assumption where they just

know the exact classifier and does respond to it.

So, we're thinking of it as

a worst case security assumption.

Question?

Okay, cool.

So now, I want to talk about measuring performance,

and why it's not obvious how to do that,

but how regret can be used.

So, the key ideas is that we're assuming that

these fraudulent agents are best responding, okay?

So, if they're honest, as I said,

they just make the transaction they want, but if not,

so let's define this function x hat t of

beta to be the best response.

So, the thing that the agent actually does,

the transaction they actually make, right?

So for the honest agent,

they just make the one they wanted,

but the dishonest agent is

doing some utility maximization.

So first, let's just define our performance.

So, this would be something pretty familiar and

normal in these machine learning settings.

So, we're going to take an average over all time,

capital t time steps,

of the loss that we experienced at that time step,

which is measuring how well did our classifier beta t do,

when the agent played x hat of

t of beta t. They responded to it,

and this was the true label.

Okay, so that's fine.

But normally, in this kind of model,

you look at regret as

compared to the best thing

you could have done in hindsight.

And now, there's a small wrinkle.

So, it's not really fair to

look at all the fraudulent transactions that were made,

and put them as dots like I did and then just say, "Oh,

If we had drawn this classifier,

we would have classified them all correctly."

So that's not fair because if

we had deployed this classifier,

they would have been modifying

their locations to respond to it.

So we can't just look at the past data and say,

something that did well in the past data,

would have actually done well if

deployed because if deployed,

the agents would have responded.

Okay. And so, this leads to this idea

that we want to compare to

the best fixed classifier in hindsight,

as we often do in these learning settings

but if we had used a different one,

they would have responded differently.

So, we need to compare to how it would have

performed had the agents responded to it.

What's the best thing in hindsight

knowing the agents would respond.

Okay. So, this was our performance

and then this is what OPT should look like.

The key point of this discussion is

that if we have deployed beta star the whole time,

we would have experienced loss of

beta star on the agent's responses to beta star.

So, this looks a little bit different

than the standard online learning where you might say,

"Take all of these x-y pairs and

plug in the performance of

beta star on those same x-y pairs.

This is taking into account that the

x's would actually look different.

The difference between these two performances

is just called the average Regret.

In the paper, we use this term, Stackelberg Regret.

The term comes from Stackelberg games in game theory

where a leader commits to

an action and the follower best responds.

So in this case, the follower is

best responding to the classifier.

So this notion of regret,

you can actually find in

previous papers in other literatures, though.

So things like security games and probably

if you're looking at online pricing

and auctions or something,

they would be using this kind of notion.

Although, we haven't seen it given a name

but we think Stackelberg Regret is a good name.

>> Is it still [inaudible] partial monitoring because I think the

enterprise literature on [inaudible].

>> Okay.

>> Like they have generalisation matrix.

>> Okay. Yes. So, it

is natural if you're thinking about this problem.

So, it's not surprising that people

have defined it in different places.

Good. But an interesting point about it is

that you can't- Yes, go ahead. Sorry.

>> So then the y's are also different, right?

>> No. So, that's important.

So, think of each agent that arrives, each x-y pair,

they're either a fraudster or they're

honest but they change.

If they're fraudster, they change their x.

>> They change their transaction.

>> They change their transaction x but their y is still,

were they honest were they not.

Yes, and one thing

that's interesting about our model

is that it's really baked in,

like their label is whether or not they're strategic.

So, that's an interesting aspect.

>> Do you need any assumptions on what fraction is fraud?

>> Good question.

Yes. So, we don't

need assumptions on what fraction are fraud versus not.

One thing we tried to do in

an extension is perform better

if there's a small amount of

fraud and you'll see briefly why.

So, a key point that's interesting

is if you just have the past data in front of you,

you can't really compute this OPT.

You have to know what these agent's utility functions

were in order to compute how they

would have responded if you had done something different.

You don't actually know that. You don't

know what their utility functions were,

you just got to see what they actually did on that day.

But we'll show how to compete with it,

how we can have an algorithm

whose performance competes with OPT,

or at least under some

assumptions under which you can do that.

I guess, I wouldn't be

surprised if everyone knows what average Regret is.

But just to be sure,

the point is that if average regret is going to zero,

then your loss per day is approaching that of OPT.

So that's good. So, you want

this average regret to go to zero.

If it goes to zero very fast, that's better.

By fast, I mean with t. So we're

often interested in average Regret as a function of time.

Okay. All good?

Okay. So now, let me tell you.

So, I tried to set up this relatively general question.

Let me tell you this specific setting

that we're able to solve for this paper.

So, we're assuming some linear predictor.

So, beta t is also

a vector in r to the d. The prediction,

you can think of as this number, beta.x.

Okay, so you can think of

positive means honest and negative means dishonest.

Very positive means I'm very

confident that this agent is honest.

Very negative, very confident they're dishonest.

So, we think about

logistic loss and we can also do hinge loss,

but we just focus on those two loss functions.

Okay. So this, I guess,

most people have seen this before but again,

this beta.x, this number,

if it's positive it thinks they're honest.

So if it matches y, if the sign of this is y,

then you have low loss.

If the sign doesn't match y,

so you predicted positive but they're actually fraud,

then you have a high loss, roughly.

But it's sort of a soft version

of zero one or it's trying to

be some soft version of zero one. All good.

Okay. So now, a very key assumption that we make,

what does the utility look like of these fraudsters?

What are they trying to do?

Okay, we assume that their utility of the form beta.x.

So, being classified as more positive is better for them.

So, the more honest

that they're perceived to be, the better.

But they pay some cost.

So notice this is beta.x-hat.

So, this is what the classifier predicts for

them but they suffer

some costs for having to modify from x to xn.

So, they have to trade off these two things.

We make some assumption on the class of

distance functions that are

measuring the cost for making this modification.

>> [inaudible]

>> Yes, good. So yes.

The reason is that we want to allow

heterogeneity among time steps.

So, an agent today

may care about different things than the agent tomorrow.

This is measuring that to say that,

like the agent today doesn't really mind switching

between chocolate and strawberry,

but the agent tomorrow minds very much.

So they have different distance functions.

>> What is the rationale of putting

this kind of disconnect when

you are putting it through the link,

one link for one agent but the other agent?

>> Do you mean,why one, the honest agent?

>> Looks linear.

>> Sorry so.

>> So, like the other loss

has this kind of hinge like form.

>> This one.

>> Right, whereas the process utility

is just fully of linear.

>> Yes. I see. I see.

So, maybe it wouldn't make

sense for their utility to be like the negative

of our loss or something like the worse we do the

happier they are or something.

>> [inaudible].

>>Yes, so it'll turn out that this is very

important for our whole approach, this particular form.

Yes, it would be

very interesting to be

able to solve something else like that.

I think one thing to mention though is,

one thing that keeps us really interesting is that it's

not a zero-sum game,

which your proposal would also not

be zero-sum because of this, I guess.

If it's a zero-sum game,

I think things are just less interesting because, yeah.

>> Is it realistic to assume this a utility function?

>> That's a great question, yes.

>> So one reason is to say like okay this

the best we can do technically.

>> Yeah.

>> Now does this merge anything.

>> Yeah.

Yeah, so we don't really know.

We don't have a particular scenario or like

we looked at these agents and they

definitely have this utility.

We can try to understand when this would be reasonable.

We don't think the answer is never.

So we think there are some scenarios.

So the most reasonable thing.

The first thing you would think of

probably could be like a 1 0.

So like my utility is one

if I got off my transaction went through.

So I was incorrectly classified and 0 if it didn't.

That would be maybe the most natural positive thing here.

But yeah go ahead.

>> [inaudible].

>> Well, yes.

So one thing we were thinking about is if

the classifiers using some randomisation

that's proportional to this.

So I'll allow things to go

through maybe with probability proportional

to the certainty of a classifier

then that this actually does make sense as

a utility could be.

But yeah, the main if you'd

go for something else from this

we don't know technically how to solve it.

So it would be great to do that.

>> I just wanted to mention that the 0 1,

I'm not necessarily sure that that's

the most natural thing because the thing where you

said before I guess it's unreasonable to assume

that attackers have a precise knowledge of data, right?

Here is also reflecting

that like the attackers want to make

sure be confident about

transactions will be classified as-

>> Yeah, thanks. That's something I forgot to

mention but is a great justification.

So yeah if you don't know

but you know that the higher this is the more

likely you are to be to pass through the filter then you

just try to maximise this that's some justification.

>> I guess it's not obvious to me how this

captures just like the inherent value of x or x half.

If x is like, I don't know,

like a 51-foot yacht and x half is a 52-foot yacht.

I don't really care.

Like they both have incredibly high value.

>> Yeah.

Yeah. So in this project

we're at a very high level of abstraction

and it's not clear what

the space are to

the deep space these things live in mean.

>> But is there a way, I mean

like some x hat are just going to have

intrinsically higher value or

some xs are just going to

have intrinsically higher value.

It seems like the only place that can

be captured in the prediction part.

>> Yeah. Yeah. So or maybe another,

I was going to say maybe you

could just have a coefficient in front that's

like how much I like

getting it through but you're right because I get

different utility for

different transactions going through.

Yeah. So yeah that may

that might not be captured by

this. That's really interesting.

>> Couldn't the distance

metric also include dollar amount?

>> Yeah. So yeah we can at least make these comparable.

Yeah. So you could sort of say,

yeah actually maybe this coefficient can

depend only on my true xt.

And then the distance is saying like you know

my distance from 51-foot yacht had

the value of 51-foot yacht minus

something number of feet.

Yeah. So yeah, and I'm glad we're having

the discussion because this is

a great direction for future work

can you what else can you

do besides these assumptions we're

making is really interesting.

Okay, so and our approach isn't going to work for

all Dt but for

some relatively large set of especially norms

of the of x minus x set and so this is

a pretty general almost completely general example

of the kinds of distance functions we can have.

So maybe have some matrix A and you take Ax and you

take Ax hat and you look at the distance

between them in some p norm

and you raise that to a power.

It's actually important for us that

this power be strictly greater than one.

The reason for that is actually intuitive.

There's this idea of a convexity involved.

And convexity is capturing diminishing returns.

So if I move this

a certain distance away and I

move that same distance again,

it's even more costly

and we actually need that to get results.

Otherwise you could have some,

this might not have a maximum.

You might be able to get

arbitrarily high utility by

moving farther and farther away arbitrarily.

Okay, good. So the main result

is under these assumptions

and in this setting that I told

you about we can reduce to

some online convex optimization algorithms

and use that to guarantee some regret.

So let me just really briefly tell you

that result and the idea of it.

So think of this loss at time

t as just a function of the classifier that

you chose to deploy at time t.

So that you get that

by plugging in beta in several spots here.

But think of it as one big function of beta.

Okay.

Well, if we can get that function to

be convex as a function of beta,

then we can apply things we know how to do.

Convex optimization.

Okay, and so that's what we show is true and

main tools using some convex analysis

which details which I

don't think are important for this talk,

so I'll skip it.

But so the corollary is

by applying various convex optimization algorithms,

you can get regret that's going to zero

as t increases some very fancy one you can

even get I think one over root t or one over t to

the two-thirds with better dependence on dimension.

Okay and this is despite I should emphases not

knowing exactly what this lost function

is that arrive right.

So we don't know this whole,

wo we have now a sequence of loss functions at each time.

They're each different.

And we don't know what

that entire loss function is

because it depends on their utility.

We just know that it's convex.

And there are convex optimization algorithms

that actually work in that setting.

>> [inaudible].

>> Yeah, I think so.

Yeah. Yeah. Good. Yeah, and so, sure.

Yes. So just quick extensions and future work.

Okay, so an extension

is you can treat these honest updates and

fraudulent updates differently and

do a better job with your algorithm.

The reason is that when an honest agent

arrives you actually do

know this convict's last function,

because it's just your last function.

If you change beta the thing

the data they gave you is not going to change.

So you can computer gradient

and you can do it fast a good update stuff.

When the fraud fraudulent update occurs,

you can't do that and you have to do something that

doesn't use gradient information.

If for those of you who are familiar

with such algorithms if not it's not a big deal.

Okay, so we have somewhat more general agent utilities

and also hinge lost.

The ideas are the same

but there's tons of room for future work.

I think it's really interesting as we've been discussing.

So other loss functions definitely.

Other forms of agents utility

definitely really interesting and especially if,

so you know our whole general approach

here once we formulated the problem ended up being

to reduce to this convex optimization which we

know how to do in this online setting.

So I mean one way to push on this is to just

say well here's another case

where we can use convex optimization.

Here's another case to go outside of that I really

don't know what approach you can

take that would be really interesting as well.

Okay. Yeah, go ahead.

>> So I guess one thing that makes me slightly nervous [inaudible]

optimization is that it

actually becomes really crucial that

when the attacker is in fact best responding to

the real knowledge of your data, right.

And so if they were working

with some sort of imperfect knowledge

about what I'm doing or

if I just like one very common thing

that goes on in these algorithms is you

have your current set of parameters and

because you can't observe a gradient

you randomly perturb it.

So there is no way in hell they

can know what random [inaudible] I actually applied.

>> Yeah.

Yeah, no, that's a great point.

Yeah. I think that's a disconnect between

our model and how you would deploy this in reality.

So maybe for the recording

I should say so the question was,

this is really relying on agents exactly best

responding to the exact beta we deploy.

And you're correct, we're using

these convex optimization algorithms that perturb

data a little and see how the gradient or see

what happens and use that to guess the gradient.

So if agents aren't best responding to the exact data

including this perturbation then

it's not clear this is going to work.

Yeah, I think another,

maybe there can be some way to get

around that by if you make

some assumptions about agent utilities,

you could guess that maybe

the gradient like locally you can guess

that they won't change too much and

then just use the gradient of your own loss.

So yeah, it should be addressable I would hope,

but not in this paper.

It's a great question.

>> Anything else, yeah.

>> I was just trying to think as at

a current price or whatever my utility would be.

It seems like it would be like a lot related to

the dollar amount of the item or

the dollar amount you could get from selling it.

So probably some things are easier to sell than others.

>> Yeah.

>> So, maybe if you could run like, if it all comes out,

suppose it all comes down to one number like

the dollar amount and you can sell something for,

will that simplify it what you do with it?

>> Yeah. Maybe, yeah. I think this is really

similar to the previous question about,

could we sort of this utility

doesn't seem to directly depend on something

like dollar value amount but could we

somehow argue that that's

a special case where there is

a coefficient which is the dollar amount of x.

>> [inaudible] second terms, you could just put them in dollars.

>> Yeah. So it seems to me possible.

Well, okay.

So we wanted A to be invertible but if you think of maybe

mapping these purchases into dollar amounts, okay?

If it's mapping down to

one dimension maybe it's there's a problem with

our formalism but it seems possible.

But yeah, when I don't want to claim this till

he is capturing too much because it's

a pretty special form so I'm not sure.

Good. Okay great. So yeah.

So, while I have you I'd like to give you

a different example of this machine

learning in societal context which is going to be

privacy focused example and much briefer of course.

Yeah. So, strategic behavior

was one really interesting aspect of this idea.

Another one that I'm interested in is this idea

of, maybe the agents,

they don't have a particular concern

about what this classifier is

doing so it's not like you know it's

classifying their own transactions.

But there is this privacy concern that it's going to

reveal information about my data

and differential privacy is a field which turns out to

be some very nice theory

and also moving into practice for addressing this issue.

So, I think I can probably guess that

everyone has seen differential privacy

a few times or many times.

So, I hope you don't mind, I think I'll

skip the definition.

It's actually not that important

for the level I want to tell you about this project.

The high level idea is you take your algorithm should be

randomized and by adding noise or being randomized,

reveals very little about the data

because a change in one person's data,

doesn't make much difference in the output you

see in terms of probabilities of different up.

That's a high level if you know that,

then you know enough to follow.

So, let me just skip this thing. It's fine.

Okay.

So, what do we want to address

in this project I'm going to tell you about briefly.

So, differential privacy, it's pretty well

understood and practical I would argue for

answering statistical queries about a data set.

So, what is the meaning of this data set?

Or something like that or different queries

about some complex data

but even being used in the census which is pretty cool.

And the most basic approach is that, okay.

So there's this parameter Epsilon differential privacy,

if you add noise that has

standard deviation like one over Epsilon,

then you're preserving Epsilon

differential privacy intuitively.

So, Epsilon very small,

is very private but lots of noise.

So, we understand all that very well.

Machine learning queries and data is very interesting.

So, there's more recent work.

It's quite active but I would argue it's

not well understood at least not in

the way that statistical queries are.

Because the kind of theory bounds we

get might be quite loose because we're

using these complicated algorithms, right?

So this algorithm is,

take the true number, add some noise, okay?

And then we can analyze how accurate that is.

If you're talking about something like

a stochastic gradient descent with

extra noise added to preserve privacy,

understanding how private and accurate that really is,

it's going to be very difficult

and often these papers are written by theorists.

So, who are great and I love theorists, okay?

But their theorems tend to look like,

I should say, our theorems tend to look like this.

They tend to look like the following.

Let's say like you prove

some general theorem that says you run

our algorithm with privacy level Epsilon and I'll

guarantee you an accuracy Alpha with high probability.

And often, you might

prove something with a consonant that looks like this,

more often you prove a

constant that looks like this, right?

So why is this an issue?

I think it's an important idea.

So people do this in algorithms.

But you put the algorithm in

practice and it runs really well,

and it doesn't matter you know the fact that you couldn't

prove a low constant is not a big deal.

But here, I want to argue it is an issue.

So, let's say you're an engineer at

some data corporation and

you want to run a differentially private,

you want to run this algorithm.

Well, you have to choose the privacy level you're

guaranteeing the people, okay?

And the problem is that theorem

isn't really going to help you that much in

figuring out how you actually choose it.

If you want an Alpha.01,

should you set Epsilon to be 100,000?

That will guarantee you an Alpha.01

but there's no privacy at all basically.

We want Epsilon to be around one or less.

On the other hand, you could probably

could get away with setting up Epsilon to one

because probably there's all these

loose constants in there but

you don't know and it can be hard.

And what if accuracy is critical to

your system so you can't afford to take chances?

Okay. So this project and this is a paper it

nips last year with some wonderful co-authors and Steven.

I'm sorry, and that's the most wonderful of them all.

So the question is,

if we have an accuracy requirement,

can we run a learning algorithm that tries to find

the most private possible output that we can

have without having to

tell it what Epsilon and Alpha to shoot for,

or sorry what Epsilon to shoot for,

it will try to satisfy

this accuracy requirement but be really private, okay?

And so, the setting

is what we call empirical risk minimization.

So, you just have this data set and you

have a last function and you

just try to find an hypothesis

that minimizes loss on the data set, right?

Okay.

So, I'm just going to give you

an example of the idea that we are

deploying and they'll be about it for this project.

So that will be about all of them.

Okay. So, one kind

of private algorithm is called output perturbation.

You compute the true output.

So, for example, this is a hypothesis, let's say,

like a classifier or something that you found by doing

empirical risk minimization on your data set, okay?

And now, you add a bunch of noise to all these numbers.

And now, it's private. Okay. And the difference

here is that we're going to use random

walks to add this noise.

So if you run a random walk for a certain length of time,

you've added a certain variance of

noise to each number and that's fine.

And you get some some output.

Okay, but now,

the key idea is to check how accurate this is.

And if it's not accurate

enough you wanted a certain accuracy level,

then we're going to have to improve

the accuracy somehow, okay?

So, there's this idea,

this step called interactive

above threshold that checks the accuracy.

Sorry, I'm not going to talk about

but you have to worry about

leaking privacy when you do this.

Okay. But let's suppose we check the accuracy.

If it's not accurate enough, the idea is to rewind

the random walk. That's the key idea.

We got it from a similar idea in

a statistics paper on

privacy that kind of does the reverse of this.

So now, let's just rewind to here.

That's on average, less noisy.

Check the accuracy. Is it good enough?

If it's not accurate enough,

so if this hypothesis is not predicting well

enough on the data set, rewind some more.

Eventually, as you approach the left you're going

to approach the true best possible hypothesis.

So you will approach something

that's accurate enough, right?

And that accuracy requirement will tell you when

you can stop, okay?

And now,

it's kind of late in the talk but maybe I can try to

convey this intuition for why this is good.

So the key idea of why this is

good from a privacy standpoint

is what you'd worry about is this process,

this interactive process of

compute some output check its accuracy and so on,

is leaking lots and lots of privacy, right?

So, you'd worry that

all these outputs are

leaking information about the dataset.

And you can actually

analyze it as though in the worst case we

published all these outputs that

would leak possibly lots of information.

Okay.

This turns out not to be true because of the random walk.

So, the idea is,

this noisy estimate here.

So, if you see this noisy estimate,

you're inferring something about the data.

But once I reveal this less noisy estimate,

that noisier one does not give

you any more information about the data.

>> Why? Because of these random walks.

So, if I tell you this number, and this number,

and you're trying to guess

or infer this thing, which is private.

This number will not help you at all.

It's independent condition on this one.

That's the idea of the random walk.

>> [inaudible]

>> Exactly. Yeah. So, that's why,

we saved the result of this

random walks secretly and rewind it.

Instead of just try again with less noise.

Yeah.

>> You can't go to the other way,

supposed you have plenty of accuracy,

but you want more privacy?

>> Yeah. So, that's tough to reason about because,

you've leaked some sort of

information just from the fact that

you decided to increase the privacy,

like increase the noise.

And information leak depended on

this very accurate thing you

computed which was very non-private.

I don't want to say it's impossible,

but to analyze is really difficult. So, we don't know-

>> You have to start earliest in

your algorithm, or you have to start with-

>> The very private thing in them.

>> Made more noise than do you think you need.

>> Yeah, yeah.

>> I was thinking like this whole thing is

like done inside a private box.

>> Yes, that's true. That's true.

>> Or you're kind of assuming that average hack

on the right hand side,

leaks information to this public.

>> Yes.

>> You know like [inaudible] board that is true?

>> Yes. Yes. So, it's not necessarily true.

It's more about how we know how to analyze these things.

Often in privacy, we have

a very hard time understanding

how much information is leaked.

And so we'll say,

well, even if we publish this entire thing,

we can still show,

and we don't know how to prove

a better privacy bound for

the thing that doesn't publish it.

>> It could just-

>> It could-

>> Publish the end result right.

>> Yeah. So, in our analysis,

you could have just published

this hypothesis and the fact that it wasn't good enough.

But you don't say what its accuracy

was, you just say not good enough.

Publish this one, you say not good enough.

Publish this one and say that's our final.

>> Because given the final one,

all the previous one that are useless.

>> Yeah. And so it's kind of getting into,

is it just analysis or is it actually what you're doing.

Yeah. It could be that we're just not good at analyzing.

So, that's the key idea

for this project basically that I wanted to convey.

But you can use this trick in other places as

well or for lots of algorithms. Yes?

>> So, my question were also connected

to Alex's previous question.

I don't worry about, say I

only publish the final result and I don't

worry about the evaluation about

the earlier, noisier results.

Do I gain anything for keeping that random work

versus get them fresh renimis to re-round thing?

>> So, if you care about privacy loss,

this is going to be better than doing

a fresh noise each time.

Well, again,

we're not sure how to analyze like where you're

trying to keep things secret, but yeah.

>> Isn't there a little bit of poorly given permission

just from knowing the fact

that it wasn't close enough to the-

>> Yes. Yeah. So, the fact that

this thing's accuracy was not

below the threshold tells you something.

And that is analyzed

in this part of the puzzle that I didn't talk about.

So, in particular like

you don't want them make these like

billions and billions of checks because that will add up,

but you get some sort of logarithmic dependence

on the number of checks.

So, it's not too worrying.

Okay, so let me

just recap this and finish up, so we don't go over.

So, the general idea of this project is take

some known algorithms for

differentially-private machine learning and

just apply these two tricks,

so the accuracy checks that I didn't talk and

this noise reduction that I did.

And then, these algorithms,

you give them some accuracy goal

and they try to give you an answer that's that accurate,

but as private as it could be. So, that's the idea.

Okay. These are important

like we're lower than the others and that's good.

Theorists whatever, so, some quick future directions.

Yes, strategic behavior and learning,

and not even just

learning but you know there's

paradigm I also think of like,

prediction markets and this paradigm rightly.

There's information out there, we're trying to

gather it and make predictions.

So, this is in the area,

I think this general questions super

interesting and there's tons of directions,

from both practice to theory.

Privacy, I feel that

privacy is moving towards practicality and applications.

There are interesting theory questions,

but they start moving

farther and farther away from practice,

but there's a lot of very practical questions

that we need help with.

So, that we can deploy privacy to help people.

And it's like the time to do that.

Fairness, which I didn't

tell you about any fairness projects,

but I think it's the opposite actually.

We're really struggling in theory land

to get a hand along fairness.

And initially people expected some sort of one size fits

all definition which is what happened in privacy, almost.

And now we're seeing that's not really possible.

And there's tons of interesting directions there too.

Okay. So, I like some very high level closing side.

I think it's really important

when both thinking theoretically,

as well as in practice.

To be thinking about the value and structure of

information in these problems that's really important.

So, the information agents give us is really valuable.

And because of that, it's susceptible

to manipulation, and so on.

I think it's important that these systems

respects these issues like privacy,

fairness, and also strategic behavior.

I like to think it's interesting

to take that perspective that agents have

the right to be strategic and act on their own behalf.

And so we should design things that work well,

taking that into account rather than

slapping them on the wrist for being strategic.

Okay. And this issue of data, ownership, and control,

I think is really important and I hope that we can do

some research in our communities to understand it better.

So, that's it. Thanks.

Không có nhận xét nào:

Đăng nhận xét