MIT LIBRARIES DUPL
^
DEWEY
'<>•'Bu§in«ss#MIT
MIT
Sloan School
of
Management
Sloan
Working
Paper
4180-01
eBusiness@MIT Working
Paper
101
October
2001
BUILDING
TRUST
ON-LINE:
THE
DESIGN
OF
RELIABLE
REPUTATION
REPORTING
BUILDING
TRUST
ON-LINE:
THE
DESIGN
OF
RELIABLE
REPUTATION
REPORTING
Chrysanthos
Dellarocas
This paper
is availablethrough
theCenter
foreBusiness@MIT web
site atthefollowing
URL:
http://ebusiness.mit.edu/research/papers.htiTil
This paper
alsocan be
downloaded
without charge
from
theSocial
Science Research
Network
Electronic
Paper
Collection:
Building
Trust
On-Line:
The
Design
of Reliable
Reputation
Reporting
Mechanisms
for
Online
Trading
Communities
Chrysanthos
Dellarocas
SloanSchool of
Management
MassachusettsInstituteof
Technology
RoomE53-315
Cambridge,
MA
02139
dell(a!mit.cdu
Abstract:
Several propertiesofonlineinteraction arechallenging theaccumulated
wisdom
oftradingcommunitieson
how
toproduceandmanage
trust. Onlinereputationreportingsystemshaveemerged
asapromisingtrustmanagement mechanism
insuchsettings.The
objectiveofthispaperistocontributetotheconstructionofonhne
reputation reportingsystemsthatarerobustinthepresenceofunfairanddeceitful raters.The
paper setsthestageby
providingacriticaloverviewofthecurrentstateoftheartin thisarea.Followingthat,itidentifiesa
number
ofimportantways
inwhich
the reliabilityofthecurrent generationofreputationreportingsystemscan beseverely
compromised by
unfairbuyers andsellers.The
centralcontributionofthepaperisa
number
of novel"immunizationmechanisms"
foreffectivelycounteringtheundesirableeffectsof suchfraudulent behavior.
The
paperdescribesthemechanisms, provestheirpropertiesande.\plains
how
variousparametersofthemarketplacemicrostructure,most
notably theanonymity
andauthenticationregimes,caninfluencetheireffectiveness. Finally,itconcludesbydiscussingthe
implicationsofthefindingsforthemanagers andusersofcurrentandfutureelectronicmarketplacesand
identifies
some
importantopen
issuesforfiitureresearch.1. Introduction
At
theheartof anybilateralexchangethereisatemptation,forthepartywho
moves
second,todefectfromtheagreed
upon
termsinways
that result inindividualgainsforit(andlossesfortheotherparty).Forexample, intransactions
where
thebuyer paysfirst,the selleristemptedtonotprovidetheagreedupon
goodsor services ortoprovide
them
ataqualitywhich
isinferior towhat was
advertisedtothe buyer.Unlessthere are
some
other guarantees, thebuyerwould
thenbe temptedtoholdbackon
her sideoftheexchangeas well. Insuchsituations,thetradewillnevertakeplaceand bothparties will
end up
beingworse
off.Unsecured
bilateralexchangesthushavethestructureofa Prisoner'sDilemma.
Our
societyhasdevelopedawide
range ofinformalmechanisms
and formalinstitutionsform.anagingsuchthelikelihoodthatonepartywillend up empty-handed.Writtencontracts,commercial law,creditcard
companies and escrowservices are additionalexamples ofinstitutionswithexactly the
same
goals.Although
mechanism
designandinstitutionalsupportcanhelpreducetransactionrisks, theycan nevereliminate
them
completely.One
example
isthe riskinvolvingtheexchange of goodswhose
"real" qualitycan only beassessed
by
thebuyera relativelylong timeafter atradehasbeen completed(e.g.usedcars).Even where
societydoes provideremedialmeasurestocoverrisks insuchcases(forexample,theMassachusetts
"lemon
law"),theseareusuallyburdensome
andcostlyandmost
buyerswould
verymuch
rathernothavetoresort tothem. Generallyspeaking, the
more
thetwo
sidesofa transaction are separatedintimeandspace,thegreater therisks.Inthosecases,
no
transactionwilltakeplace unless the partywho
moves
firstpossessessome
sufficientdegreeoftnistthatthe partywho
moves
secondwillindeed honoritscommitments.
The
productionoftrust,therefore,isapreconditionfortheexistenceofany market andcivilizedsocietyingeneral(Dunn, 1984;Gambetta, 1990).
In"bncks and mortar" communities,theproductionoftrustisbased
on
several cues, oftenrationalbutsometimespurelyintuitive.For example,
we
tendto trustordistrust potential trading partnersbasedon
theirappearance,thetoneoftheirvoice ortheir
body
language.We
also askouralreadytrustedpartnersabouttheirpriorexperienceswiththe
new
prospect,undertheassumptionthatpastbehaviorisa relativelyreliablepredictoroffuturebehavior.
Taken
together,theseexperiencesformthe reputationof ourprospectivepartners.
The emergence
ofelectronicmarketsandother typesofonline tradingcommunitiesarechangingthe ruleson
many
aspectsof doingbusiness. Electronicmarketspromisesubstantialgainsinproductivityandefficiency
by
bringing together amuch
larger setof buyers andsellersandsubstantiallyreducingthesearchandtransaction costs(Bakos, 1997;Bakos, 1998).Intheory,buyerscan then lookforthe best possible deal
andend
up
transactingwithadifferent selleron everysingletransaction.None
ofthesetheoreticalgainswillberealized,however,unlessmarket
makers
andonlinecommunity managers
findeffectiveways
toproducetrust
among
theirmembers.
The
productionoftrustisthusemerging
asanimportantmanagement
challengeinany organizationthatoperates orparticipates inonline tradingcommunities.
Several propertiesofonlinecommunitieschallenge theaccumulated
wisdom
of oursocietieson
how
toproducetrust.Formalinstitutions,suchas legalguarantees,are less effective inglobal electronic markets,
which
spanmultiple jurisdictions with, oftenconflicting, legalsystems (Johnson and Post, 1996). Forexample,itisverydifficult,andcostly, for abuyer
who
resides intheU.S.A.toresolveatrading disputewitha seller
who
livesin Indonesia.The
difficultyiscompounded by
the factthat, inmany
electronicmarkets,itisrelativelyeasyfortrading partners tosuddenly"disappear"and reappear underadifferent
Furthermore,
many
ofthecuesbasedon
which
we
tendto trustordistrustother individuals are absentinelectronicmarkets
where
face-to-facecontactistheexception. Finally,one ofthemotivatingforcesbehindelectronicmarketsisthedesiretoopen uptheuniverseofpotentialtrading partnersand enabletransactions
among
partieswho
have neverworked
togetherinthepast.Insuchalargetrading space,most
ofone's alreadytrustedpartners are unlikelytobeabletoprovidemuch
informationaboutthe reputationofmany
of theotherprospectsthatonemay
be considering.As
acounterbalancetothose challenges, electroniccommunitiesarecapableofstoringcomplete andaccurate informationaboutalltransactionsthey mediate. Several researchersandpractitionershave,
therefore, started tolookat
ways
inwhich
thisinformationcanbe aggregated and processedby
themarketmakers
or othertrusted third parties inordertohelp onlinebuyersand sellersassesseachother's trustworthiness.Thishas leadto anew
breedofsystems,which
arequicklybecoming
anindispensablecomponent
of everysuccessful online tradingcommunity:
electronic reputation reporting systems.We
arealready seeing thefirstgenerationof such systemsintheform
ofonlineratings,feedback
orrecommender
systems (Resnick andVarian, 1997;Schaferet.al.,2001).The
basic ideaisthatonlinecommunity
members
aregiventheability torateorprovidefeedback abouttheirexperiences withothercommunity members. Feedback
systemsaim
tobuildtrustby
aggregatingsuchratingsofpastbehavior oftheirusersand
making them
availabletoother usersaspredictorsoffiiturebehavior.eBay
(www.ebay.com),
forexample,encourages bothpartiesof eachtransactionto rateone another witheitherapositive(+1),neutral(0) oranegative(-1) ratingplusa short
comment.
eBay makes
thecumulativeratings ofitsmembers,
aswellasallindividualcomments
publicly availabletoeveryregistereduser.The
majorityofthe current generationofonlinefeedback systems havebeen
developedby
Internetentrepreneursandtheir reliabilityhas not yetbeensystematically researched. In fact,thereis
ample
anecdotal evidence,aswellasone recentlegal case', related to the ability to effectivelymanipulatepeople's
actions
by
using onlinefeedbackforums(stockmessage
boardsinthiscase)tospreadfalseopinions.As
more
andmore
organizationsparticipate inelectronicmarketplaces, online reputation reportingsystemsdeserve
new
scrutinyandthestudyoftrustmanagement
systemsindigitalcommunitiesdeservestobecome
anew
additiontotheburgeoningfieldofManagement
Science.The
objectiveofthispaperistocontributetotheconstructionofonline reputation reportingsystemsthat arerobustinthepresenceof
unfairanddeceitftil raters.The
papersetsthe stageby
providinga criticaloverview ofthecurrentstateoftheartin thisarea(Section2).Followingthat,itidentifiesa
number
ofimportant
ways
inwhich
thepredictive valueofthecurrentgenerationofreputation reportingsystems can beseverelycompromised by
unfairbuyersandsellers(Section3).The
centralcontributionofthepaperisanumber
of novel "immunizationmechanisms"
foreffectivelycounteringtheundesirableeffectsof suchfraudulent behavior.
The
paperdescribesthemechanisms, provestheirpropertiesandexplainshow
variousinfluencetheireffectiveness (Section4). Finally, itconcludes
by
discussing the implicationsofthe findingsforthemanagers
andusersofcurrentand futureelectronicmarketplacesandidentifiessome
openissues for futureresearch (Section5).
2. Reputation reporting
mechanisms
inonlinecommunities
The
relativeeasewithwhich
computers cancapture,storeandprocesshuge amounts
ofinformationaboutpasttransactions,
makes
pastbehavior(reputational)information aparticularlypromisingway
on which
tobasetheproductionoftrust inonlinecommunities. Thisfact,togetherwiththe fact thatthe othertraditional
ways
of producingtrust (institutionalguarantees,indirectcues)do
notwork
aswellincyberspace,hasprompted
researchersandpractitioners tofocustheirattentionon
developingonlinetrustbuildingmechanisms
basedon
reputationalinfonnation.Thissectionprovidesacriticalsurveyofthe state-of-the-art inthis field.A
repuiaiion, asdefinedby
Wilson(Wilson, 1985) isa "characteristicorattributeascribedtoone personby
another. Operationally,thisisusuallyrepresented asapredictionaboutlikelyfuture behavior.Itis,
however,primarilyanempirical statement. Itspredictive
power
dependson
thesuppositionthatpastbehaviorisindicativeoffuturebehavior". Reputationhasbeenthe objectofstudyofthesocialsciencesfor
along time (Schmalensee, 1978; Shapiro, 1982;
Smallwood
andConlisk, 1979). Severaleconomistsandgame
theoristshave demonstratedthat, inthepresenceofimperfect infonnation, theformationofreputations isanimportant forcethathelpsbuyers
manage
transactionrisks,but alsoprovidesincentivestosellers toprovide
good
servicequality.Having
interactedwithsomeone
inthe pastis,ofcourse,themost
reliablesourceofinformationaboutthatagent's reputation. But, relyingonly
on
directexperiencesisbothinefficientand dangerous. Inefficient,because anindividualwillbelimitedinthe
number
ofexchangepartnersheorshe hasand dangerousbecauseonewilldiscoveruntrustworthypartnersonly through hard experience(Kollock, 1999).These
shortcomingsareespecially severeinthecontextofonlinecommunities
where
thenumber
ofpotentialpartnersis
huge
andtheinstitutionalguaranteesincaseofnegativeexperiences areweaker.Greatgains are possibleifinformationaboutpastinteractionsisshared and aggregatedwithinagroupin
theform ofopinions, ratingsorrecommendations. Inthe"bricksand mortar"
communities
thiscantakemany
forms:informal gossipnetworks,institutionalizedratingagencies,professionalcritics, etc. Incyberspace,they take theform ofonline reputation reportingsystems, also
known
as onlinerecommender
systems (Resnick andVarian, 1997).
The
following sectionsprovidea brief discussionofthemostimportant design challengesandcategoriesofthese systems.
Althoughtheeffectiveaggregationofother
community members'
opinionscanbeaveryeffectiveway
togatherinformationaboutthereputationofprospective tradingpartners,itisnotwithoutpitfalls.
The
followingparagraphsdescribe
two
importantissues thatneedtobe addressedby
opinion-basedreputation reportingmechanisms:Subjectivelymeasurableattributes.
Intherestofthepaper
we
willusetheterm "agent"to refertoa participant(buyerorseller,human
orsoftware)ofan online tradingcommunity.
We
say thatanattributeQ
of anagentsissubjectivelymeasurableifidenticalbehavior ofagent5 vis-a-vistwodifferentagents b^ and b_^
may
result intwo
differentratings
R'
^
R'
for attributeQ
bytherespectiveraters.The most
common
example
ofaA, A,
subjeciivelymeasurableattributeisthenotionof productorservice"quality". In
most
transactiontypes,some
ofthe attributesofinterestaresubjectively measurable.Inorderforagentbto
make
useofother agents'ratingsforsubjectivelymeasurableattributesasabasisforcalculatingagent5'sreputation,it
must
firsttryto "translate"eachofthem
intoitsown
value system. Intraditionalcommunities
we
address theaboveissueby
primarily acceptingrecommendations
from peoplewhom
we
know
already.Inthosecases,ourpriorexperiencewiththesepeoplehelpsusgauge
theiropinions
and
"translate"them
intoourvalue system.For example,we may
know
from
pastexperiencethatBillisextremely
demanding
andsoa ratingof"acceptable"on
hisscalewould
correspondto "brilliant"onourscale.
As
afurtherexample,we
may know
thatMary
andwe
havesimilartastes inmoviesbut notinfood,so
we
follow her opinionson movies whilewe
ignoreherrecommendations
onrestaurants.Due
tothemuch
largernumber
ofpotential tradingpartners, inonlinecommunities
itis,onceagain,lesslikely thatourimmediate"friends"willhave haddirectexperiences with several oftheprospects
considered Itis,therefore,
more
likely thatwe
willhavetorelyontheopinionsofstrangerssogaugingsuchopinions
becomes
much
more
difficult.Intentionally falseopinions
Fora
number
ofreasons (see Section3)agentsmay
deliberatelyprovidefalseopinions about anotheragent,thatis,opinions,
which
bearno
relationshipto their truthfulassessment oftheirexperienceswiththatotheragent. Incontrasttosubjective opinions, for
which
we
haveassumed
thattherecan bea possibilityof"translation"to
somebody
else'svalue system,falseopinionsareusually deliberately constructedtomisleadtheirrecipientsandtheonlysensible
way
to treatthem
istoignorethem.Inordertobeabletoignorethem, however, onehastofirstbeableto identifythem. Beforeaccepting opinions,ratersmust,
therefore, alsoassess the trustworthinessofotheragentswithrespecttogivinghonestopinions.
(Yahalom
orthogonaltoitstrustworthinessas aservice provider.Inotherwords, anagentcan beahigh-quahty
service providerandavery unreliable
recommendation
provider or viceversa.Intherestofthesection
we
will brieflysurveythevarious classesof proposedonline reputation reportingsystems andwilldiscuss
how
each ofthem
fares inaddressing theaboveissues.2.2
Recommendation
repositoriesRecommendation
repositoriesstoreandmake
availablereconmiendationsfrom
a largenumber
ofcommunity members
withoutattemptingtosubstantiallyprocess or qualifythem.The
Web
isobviously verywell suitedforconstructingsuchrepositories.Infact,most
current-generationweb-based
recommendation
systems (messageboards,opinionforums,etc.) fallinto thiscategory.A
typicalrepresentativeofthis classof systemsisthefeedback
mechanism
ofauctionsiteeBay.Other popularauctionsites,suchas
Yahoo
andAmazon
employ
very similarmechanisms.eBay
encouragesthebuyer andsellerofaneBay-mediatedtransactiontoleavefeedbackforeachother.Feedback
consistsofanumericalrating,which
iscanbe +1 (praise), (neutral)or -1 (complaint) plus ashort(80 charactersmax.)text
comment.
eBay
thenmakes
thelistofallsubmitted feedbackratingsandcomments
accessibletoanyother registered userofthesystem.eBay
doescalculatesome
rudimentarystatisticsofthesubmittedratings foreachuser(the
sum
ofpositive,neutral andnegativeratingsinthelast7days, past
month
and 6 months)but,otherwise,itdoesnotfilter,modify
orprocessthesubmittedratings.Recommendation
repositories are a step intheright direction.They
make
lotsofinformationaboutother agents availabletointerested users,but theyexpectusersto"make
sense"ofthoseratingsthemselves anddraw
theirown
conclusions.On
theone hand,thisisconsistentwiththeviewpointthattheassessmentofsomebody's
trustworthinessisanessentiallysubjectiveprocess(Boon
and Holmes, 1991).On
the other hand, however,thisbaselineapproach doesnotscaleverywell.In situationswhere
therearedozensorhundredsof,possiblyconflicting, ratings,usersneedtospendconsiderableeffortreading"betweenthe
lines"ofindividualratingsinorderto "translate"other people'sratings to their
own
valuesystemorinordertodecidewhetheraparticular ratingishonestornot.
What's
more,incommunitieswhere most
ratersarecompletestrangerstooneanother, thereisnoconcreteevidencethat reliable"reading betweenthe
lines"ispossibleatall. Infact,as
we
mentioned,thereisample
anecdotalevidence of people being misledby
followmg
therecommendations
offalsemessages postedon
Internetfeedbackforums.2.3Professional(specialist)ratingsites
Specialist-based
recommendation
systemsemploy
trustedand knowledgeablespecialistswho
thenengagein first-handtransactions witha
number
ofserviceprovidersandthen publishtheir"authoritative"ratings.Otherusersthenuse theseratingsas abasisforformingtheir
own
assessmentofsomeone's
credit-ratingagencies(Moody's) and
e-commerce
professionalratingagencies,suchasGomez
Advisors,Inc.(www.gomez.com).
The
biggestadvantageofspecialist-basedrecommendation
systemsisthatitaddresses theproblem
offalseratingsmentionedabove.In
most
casesspecialistsareprofessionalsandtake great paintobuildandmaintaintheirtrustworthinessas disinterested,fairsourcesofopinions (otherwise theywillquickly find
themselvesoutofbusiness).
On
theother hand, specialist-basedrecommendation
systemshaveanumber
ofshortcomings,
which become
evenmore
severeinonlinecommunities:First,specialistscan onlytesta relativelysmall
number
ofservice providers.There istimeandcost involvedinperformingthesetestsand,the largerandthemore
volatilethepopulationofonecommunity,
thelowerthepercentage ofcertifiedproviders.Second,specialists
must
beabletosuccessfullyconcealtheiridentityorelsethereisadangerthatproviderswillprovideatypically
good
servicetothespecialist forthepurposeofreceiving
good
ratings. Third,specialistsare individualswiththeirown
tastesandinternalratings scale,
which
donot necessarilymatch
thatofanyotheruserofthe system. Individual usersofspecialist ratingsstillneedtobeabletogaugeaspecialist'srecommendation, inordertoderivetheir
own
likelyassessment. Last but notleast,specialiststypicallybasetheirratings
on
averysmallnumber
ofsampleinteractionswiththeservice providers (oflenjustone).This
makes
specialistratingsaveryweak
basisfi-om
which
toestimatethe variabilityof someone'sserviceattributes,which
isanimportant aspectofsomeone's
trustworthiness, especiallyindynamic,time-varyingenvironments.2.4 Collaborativefilteringsystems
Collaborativefilteringtechniques(Goldberget.al., 1992; Resnicket. al., 1994;
Shardanand
and Maes,1995;BillsusandPazzani, 1998)attempttoprocess"raw"ratingscontainedina
recommendation
repositoryinordertohelpraters focustheirattentiononlyonasubsetofthoseratings,which
aremostlikely tobeusefultothem.
The
basic ideabehindcollaborativefilteringistouse pastratingssubmittedby
anagent i asa basisforlocating other agents
b
,b^,...whose
ratingsarelikely tobemost
"usefiji" toagent b inordertoaccurately predictsomeone'sreputationfromits
own
subjective perspective.Thereareseveral classesofproposedtechniques:
Classificationor clusteringapproachesrely
on
theassumptionthatagentcommunities forma relativelysmallsetoftasteclusters,withthepropertythatratingsofagentsofthe
same
cluster forsimilar thingsaresimilartoeachother.Therefore,ifthe taste clusterof anagent b canbeidentified,thenratingsofother
members
ofthat cluster foranattributeQ
ofagentscanbe usedasstatisticalsamplesforcalculating theThe
problem ofidentifyingthe"right"tastecluster for agivenagentreducestothewell-studiedproblem ofclassification/dataclustering
(Kaufman
andRousseeuw,
1990; Jainet, al. 1999;Gordon, 1999). Inthecontextofcollaborativefiltering,the similarityoftwo buyersisafunctionofthe distanceoftheirratings for
commonly
ratedsellers.Collaborativefilteringresearchershave experimented withavarietyofapproaches,based
on
statisticalsimilaritymeasures (Resnicket. al., 1994;Breseeet. al., 1998)aswellasmachine
learningtechniques(BillsusandPazzani, 1998).Regression approachesrelyontheassumptionthatthe ratingsof anagent
b
canoftenberelated to theratingsof anotheragent
b
throughalinearrelationshipoftheformR'
=a
R'+B
forallagents.y (1)fc V i>, 'J
Thisassumptionismotivated
by
thebelief,widely acceptedby
economists (Arrow, 1963; Sen, 1986)that,even
when
agentshave"similar"tastes,one user's internal scaleisnotcomparabletoanotheruser's scale.Accordingto this belief, inagiven
community
thenumber
ofstrictnearestneighborswillbe verylimitedwhiletheassumption of(1)opensthe possibilityofusingthe
recommendations
ofamuch
largernumber
ofagents as the basisforcalculatingan agent'sreputation. In that case,if
we
canestimate theparametersa
,p
foreachpairofagents,we
canuseformula(1)to"translate"theratingsofagents b tothe"internalscale"ofagent b and thentreatthe translated ratings asstatisticalsamplesforestimatingthe
reputation R' fromtheperspectiveofagent b .
The problem
ofestimating thoseparameters reducestotheb '
well-Studied
problem
oflinearregression.Thereisahuge
literatureon
thetopicandalotofefficienttechniques,
which
areapplicabletothiscontext(Malinvaud, 1966;Pindyck andRubinfeld, 1981).Bothclassificationandregressionapproachesrelatebuyerstoone another based
on
their ratings for acommon
setofsellers.Iftheuniverseofsellersislargeenough,evenactivebuyersmay
haveratedavery small subsetofsellers.Accordingly,classificationandregressionapproachesmay
be unabletocalculate estimated reputationsformany
seller-buyerpairs. Furthermore,theaccuracyof
suchreputation estimatesmay
be poor becausefairlylittleratingsdatacan be usedtoderivethem.Thisproblem
isknown
asreducedcoverageandisduetothesparse natureofratings.
Such
weaknessesarepromptingresearcherstoexperiment withtheuse oftechniquesfrom
thefieldofKnowledge
Discoveryin Databases(Fayyadet.al. 1996),which
discoverlatentrelationshipsamong
elementsofsparsedatabasesinthecontextofonline reputation reporting systems.
The
promisinguseofone suchtechnique, SingularValueDecomposition
(SVD),
hasbeenreportedin(Billsusand Bazzani 1998;Sarwaret.al. 2000).
Of
thevarious classesofsystemssurveyedintheprevioussection,we
beheve
thatrecommendation
repositorieswithcollaborativefilteringhavethe best potential for scalabilityandaccuracy. Nevertheless,
whilethesetechniques address issuesrelated tothe subjective natureofratings,they
do
not addresstheproblem
ofunfairratings.Thissectionlooksatthisprobleminmore
detail.More
specifically,ourgoalistostudya
number
ofunfairratingscenariosand analyzetheireffects incompromising
thereliabilityofacoUaborative-filtering-based reputation reporting system.
To
simplifythediscussion,inthe restofthepaperwe
aremaking
thefollowingassumptions:We
assume
a tradingcommunity whose
participantsaredistinguishedintobuyersandsellers.We
furtherassume
thatonlybuyers canratesellers. Ina futurestudy
we
willconsidertheimplicationsofbi-directional ratings. Ina typicaltransactioni,abuyer bcontracts withasellersfortheprovisionofa service.
Upon
conclusionofthe transaction,bprovidesanumericalrating R'^ (I),reflecting
some
attributeQ
ofthe service offeredby
sasperceivedbyb(ratingscan onlybe submittedinconjunctionwitha transaction).Again,forthesakeof
simplicity
we
assume
that R'^ (i)isascalar quantify, although,inmost
transactions there aremore
thanonecriticalattributesand ^^ (0
would
bea vector.We
furtherassume
theexistenceofanonline reputation reportingmechanism, whose
goalisto storeandprocess pastratingsinordertocalculatereliablepersonalized reputation estimates 7?^^(/) for seller.?
upon
requestofaprospectivebuyerb.In settings
where
thecriticalattributeQ
forwhich
ratings areprovidedissubjectivelymeasurable,thereexistfourscenarios
where
buyersand/orsellerscanintentionallytry to "rig thesystem", resultinginbiased reputation estimates,which
deviatefroma "fair"assessment ofattributeQ
foragivenseller:
a.Unfairratings
by
buyers• Unfairlyhighratings ("ballotstuffing"):
A
sellercolludeswithagroup of buyers inordertobe givenunfairiyhighratings
by
them.Thiswillhavethe effectofinflatingaseller'sreputation, thereforeallowingthat seller toreceive
more
ordersfrom buyers andatahigherprice than she deserves.• Unfairlylowratings("bad-mouthing"):Sellerscancolludewithbuyersinorderto
"bad-mouth"
othersellersthatthey
want
todriveoutofthemarket.In suchasituation,theconspiringbuyers provideunfairlynegativeratings to thetargetedsellers,thusloweringtheirreputation.
b.Discriminatorysellerbehavior
• Negativediscrimination:Sellersprovide
good
servicetoeveryoneexceptafewspecificbuyersthatthey "don'tlike". Ifthe
number
of buyers beingdiscriminatedupon
isrelativelysmall,thecumulative• Positivediscrimination:Sellersprovideexceptionally
good
servicetoafew selectindividualsandaverageservicetotherest.
The
effectofthisisequivalentto ballot stuffing.Thatis,ifthefavoredgroupissufficiently large, theirfavorableratings will inflate thereputationofdiscriminatingsellers
andwillcreateanexternalityagainsttherestofthebuyers.
The
observableeffectofallfourabovescenariosisthattherewillbea dispersionofratings for agivenseller. Iftheratedattributeisnot objectivelymeasurable,itwillbe verydifficult,orimpossibleto
distinguishratingsdispersionduetogenuinetastedifferences fromthat
which
isduetounfairratingsor discriminatory behavior.Thiscreatesamoral
hazard,which
requires additionalmechanisms
inordertobeeitheravoided, or detectedandresolved.
Inthe followinganalysis,
we
assume
theuse ofcollaborativefilteringtechniquesinordertoaddress theissueofsubjectiveratings.
More
specifically,we
assume
that, inordertoestimate thepersonalizedreputationof5
from
the perspectiveofb,some
collaborativefilteringtechniqueisusedtoidentifythe nearestneighborsetN
oi^b.N
includesbuyerswho
havepreviously rated.sandwho
are thenearestneighborsofi,based
on
thesimilarityoftheirratings,forothercommonly
ratedsellers,with thoseofb.Sometimes,thisstepwillfilteroutallunfairbuyers. Suppose, however,thatthe colludershavetaken
collaborativefiltering intoaccountand havecleverlypicked buyers
whose
tastesaresimilartothoseofb
ineverythingelseexcepttheirratingsof5.In thatcase, theresultingsetA'willinclude
some
fairratersandsome
unfairraters.Effects
when
reputationissteadyoverlimeThe
simplest scenariotoanalyzeisonewhere
we
canassume
thatagent behavior,andthereforereputation,remainssteadyovertime.That
means
that,collaborativefilteringalgorithmscantakeintoaccountallratings in theirdatabase,
no
matterhow
old.Inorderto
make
ouranalysismore
concrete,we
willmake
theassumptionthatfairratingscan rangebetween!/?
,R
1andthatthey followa distributionofthe general form: tnin maxTUR)
=
max{R,mm{R
,z))where
z~
N{^,a)
(2)b
which
inthe restofthepaperwillbe approximatedto T'(R)
~
N{jU,
(T).The
introductionofminimum
and
maximum
ratingbounds
correspondsnicelywithcommon
practice.The
assumption of nomiallydistributedfairratings,requires
more
discussion. Itisbasedon
thepreviousassumptionthatthoseratingsbelong to thenearestneighborsetofagivenbuyer,andtherefore representasingletaste cluster. Withina
taste cluster,itisexpectedthatfairratings willberelativelyclosely clusteredaround
some
valueand henceInthispaper
we
will focusonthe reliableestimationofthereputationmean.Given
alltheaboveassumptions,thegoalofareliablereputationreportingsystem should bethe calculationofafair
mean
reputationestimate
(MRE)
R'^which
isequaltoor very closeto jU,themean
ofthefairratingsbjatr '
On
theotherhand,thegoalofunfairraters istostrategicallyintroduce unfairratingsin ordertomaximizethe distance
between
theactualMRE
R
*, calculated
by
the reputationsystemandthefairMRE.
More
specificallytheobjectiveofballot-stuffingagentisto
maximize
theMRE
while bad-mouthingagentsaimtominimizeit.
Note
that,incontrasttothecaseoffairratings,itisnot safetomake
any
assumptions abouttheform ofthe distributionofunfairratings.Therefore,allanalysesinthe restofthispaperwillcalculate
system behavior underthe
most
disruptive possible unfairratings strategy.We
willonly analyzethecaseofballot-stuffingsince thecaseof bad-mouthingissymmetrical.Assume
thattheinitialcollaborativefilteringstepconstructsa nearestneighborsetN,in
which
theproportionofunfairratersis5andthe proportionoffairratersis i5. Finally,ourbaseline analysisin thissectionassumes
thattheactual
MRE
/?' istakentobethesamplemean
ofthemost
recent ratinggivento sby
eachb.aclual
qualifyingrater inA^.This simpleestimatorisconsistentwiththepracticeof
most
current-generationcommercial
recommender
systems(Schaferet. al. 2001).In that case,the actualMRE
willapproximate:b.actual ^ ' '^ '^ u
where
U
isthemean
valueofunfairratings.The
strategy,which maximizes
theabove
MRE
isonewhere
u
=
R
,i.e.where
allunfairbuyers givethemaximum
possiblerating totheseller.'^
u max
We
define themean
reputation estimate biasforacontaminatedsetofratings to be:B
=
Rl
-R'.
(5)b.uctual bjair
Inthe
above
scenario,themaximum
MRE
biasisgivenby:B
={\-S)H
+
dR
-^
=
5{R
-jU)
(6)max ' max
' max
Figure 1 tabulates
some
valuesofB
forseveraldifferentvaluesLiand6,inthe special casewhere
*^ max
ratingsrange
from
[0,9]. Forthepurpose of comparingthisbaselinecasewith the"immunizationmechanisms"
describedinSection4,we
havehighlighted biasesabove5%
ofthe ratingsrange(i.e.biases greater than±0.5 pointsonratingswhich
range from0-9).As
can beseen,formula(6)canresult inverysignificant inflationofa seller's
MRE,
especiallyforsmall(Iandlarge6.Percentage of
selleras a solution. Inenvironments
where
reputation estimatesuseallavailableratings, thissimplestrategyensiires thateventually6 can neverbe
more
than the actualfractionofunfairratersinthecommunity,
usuallyavery smallfraction.However,
thestrategybreaksdown
inenvironmentswhere
reputation estimates arebasedon
ratingssubmittedwithina relativelyshorttimewindow
(orwhere
olderratingsareheavily discounted).
The
followingparagraphexplainswhy.
Let us
assume
thattheinitialnearestneighborsetN,„i,icicontainsm
fairratersand
nunfairraters.Inmost
casesn«
m.Assume
further thattheaverageinterarrivaltimeoffairratings for agivensellerisX
and thatpersonalizedMREs
R'
(t) arebased onlyon
ratings for 5submittedby
buyers ue A',,,,,/,,/within thetime
window
W
=[t-
kA, t].Based
ontheabove
assumptions,theaveragenumber
offairratingssubmittedwithin fF
would
beequaltok.To
ensure accurate reputation estimates, thewidth ofthetimewindow
W
shouldberelativelysmall; thereforek shouldgenerallybeasmall
number
(say,between
5 and20).Fork«
mwe
canassume
thateveryratingsubmittedwithin fFisfroma distinctfair rater.Assume
now
thatunfairratersfloodthesystem withratingsata frequency
much
higher than thefrequencyoffairratings. Ifthe unfairratingsfrequencyishighenough, every one ofthenunfairraterswillhave submittedatleastone
ratingwithinthetime
window
W.As
suggestedby
Zachariaet. al.,we
keep onlythelastrating sentby
eachrater.
Even
usingthat rule,however,theabovescenariowould
result inanactivenearestneighborsetofraters
where
thefractionofunfairratersis5=
n/(n+k). TTiisexpressionresults in5>
0.5forn>k,
independentof
how
smalln isrelative tom. For example, ifn=10
and k=5, 6=
10/(10+5)=
0.67.We
thereforeseethat,for relativelysmalltime
windows,
even asmall(e.g.5-10)number
ofcolludingbuyers can successfullyuseunfairratingsfloodingtodominatethesetofratingsusedtocalculateMREs
andcompletelybias theestimateprovided
by
thesystem.The
resultsofthissection indicatethatevena relativelysmallnumber
ofunfairraterscansignificantlycompromise
thereliabilityofcollaborative-filtering-based reputation reporting systems. Thisrequires thedevelopment ofeffectivemeasuresforaddressing theproblem.
Next
sectionproposesand
analyzes severalsuch measures.
4.
Mechanisms
forimmuntzing
onlinereputationreportingsystems against unfairraterbehaviorHaving
recognizedtheproblem
ofunfairratings as a realandimportant one,thissectionproposesanumber
ofmechanisms
foreliminating orsignificantlyreducingitsadverseeffectson
the reliabilityofonline reputation reporting systems.
4.1 Avoidingnegative unfairratingsusing controlledanonymity
The main argument
ofthissection isthattheanonymity regime of anonlinecommunity
caninfluence the kindsofreputationsystemattacksthatarepossible.A
slightlysurprisingresultisthe realization that a fullytransparentmarketplace,
where
everybodyknows
everybodyelse's true identity incursmore
dangersofreputationsystemfraud thanamarketplace
where
the true identitiesoftradersare carefullyconcealedfrom eachotherbutareknown
toatrusted third entity(usually themarket-maker).Bad-mouthing
andnegative discriminationarebasedon
theabilitytopickafew
specific"victims"andgive
them
unfairlypoorratingsorprovidethem
withpoorservice respectively. Usually, victims are selectedbased onsome
real-life attributesoftheirassociated principalentities(forexample, becausetheyareourcompetitors orbecauseofreligiousorracialprejudices).This adverseselectionprocesscan be
avoidedifthe
community
conceals thetrue identitiesofthebuyers andsellersfrom eachother.Insucha "controlled
anonymity"
scheme, themarketplaceknows
the true identityofallmarketparticipantsby applying
some
effective aiithenlicationprocessbeforeitallows accesstoanyagent (Huttet. al. 1995).Inaddition,itkeepstrackofalltransactionsandratings.
The
marketplacepublishes the estimatedreputationof buyers andsellersbutkeepstheir identitiesconcealed
from
eachother(orassignsthem
pseudonyms which
change from onetransactiontothenext, inordertomake
identitydetectionverydifficult).In thatway, buyersandsellers
make
theirdecisions solelybasedon
theofferedtermsoftradeaswellasthepublishedreputations. Becausetheycan
no
longeridentify their"victims",bad-mouthingandnegative discriminationcanbeavoided.
Itisinterestingtoobservethat,while,in
most
cases,theanonymity
ofonlineconmiunities has been viewedas asourceofadditionalrisks(Kollock 1999;Friedman and Resnick1999), here
we
have anexample
ofasituation
where
some
controlleddegreeof anonymity can be usedtoeliminatesome
transactionrisks.Concealingthe identitiesof buyers andsellersisnot possibleinalldomains. For example, concealingthe
identityofsellersisnot possibleinrestaurantandhotelratings(althoughconcealingtheidentityof buyers
is).Inotherdomains,it
may
requirethecreative interventionofthe marketplace.For example,inamarketplaceofelectronic
component
distributors,itmay
requirethemarketplaceto act asanintermediaryshipping
hub
that willhelp erase informationaboutthe seller'saddress.Ifconcealingtheidentitiesof bothpartiesfrom eachotherisnot possible, thenit
may
stillbeusefultoconcealthe identityofoneparty only.
More
specifically,concealingtheidentityof buyersbut notsellersavoids negative discrimination against
hand
picked buyersbutdoesnotavoid bad-mouthing ofhand
pickedsellers.Inananalogousmanner, concealingtheidentityofsellersbut notbuyers avoids bad-mouthingbut not negative discrimination.
These
resultsaresummarized
inFigure2.Generallyspeaking,concealingtheidentitiesof buyersisusually easierthanconcealingtheidentitiesof
sellers (asimilarpointis
made
inCranor and Resnick1999). Thismeans
thatnegative discriminationiseasier toavoidthan"bad-mouthing". Furthermore, concealingtheidentitiesofsellersbefore a serviceis
performedisusually easier than afterwards. In
domains
withthisproperty, controlledanonymity
canbesubsequent bad-mouthing. For example,intheabove-mentioned marketplace ofelectronic
component
distributors,one couldconcealthe identitiesofsellers untilaftertheclosingofadeal.
Assuming
that thenumber
ofdistributors for agivencomponent
typeisrelatively large, thisstrategywould
make
itdifficult,or impossible,formalevolent buyersto intentionallypickspecific distributors forsubsequent
bad-mouthing.
The
samplemedian
F
of « orderedobservationsY
12/1
<Y
<...<¥
isthemiddleobservation Y,where
Ak=
(n+I)/2if«isodd.When
n iseventhenY
isconsideredtobeanyvaluebetween
thetwo
middleobservations
Y
andY
where
k=n/2, althoughitismostoftentakentobetheiraverage.Intheabsenceofunfairratings(i.e.
when
5=0)we
havepreviouslyassumed
that ^^(^)
=
^{Mi
<^)• '^'^well
known
(Hojo, 1931)that,as the size« ofthesampleincreases,themedian
ofasampledrawn
fromanormaldistributionconvergesrapidlyto anormaldistributionwith
mean
equal tothemedian
oftheparentdistribution.Innormaldistributions,the
median
isequaltothemean.Therefore,insituationswhere
therearenounfairraters,theuseofthesample
median
results inunbiasedfairMREs:
bjair '
Let us
now
assume
thatunfairratersknow
thatMREs
arebasedon
thesample median.They
willstrategically try tointroduce unfairratings
whose
valueswillmaximize
the absolute biasbetween
thesample
median
ofthefairsetand
thesamplemedian
ofthecontaminatedset.More
specifically, "ballotstuffers" will try to
maximize
thatbiaswhile"bad-mouthers"will try tominimizeit.In thefollowinganalysis
we
consider the caseofballot stuffing.The
caseof bad-mouthingissymmetric, withthesigns reversed.Assuming
thatthenearestneighborsetconsistsof«
=
{\—
5)n
fairratingsandn
=
S
n
unfairratings,
where
<
(J<
0.5,themostdisruptive unfairratings strategy, intermsofinfluencingthesamplemedian,isone
where
allunfairratingsarehigher than thesamplemedian
ofthecontaminatedset.In thatcaseandfor
J
<
0.5,allthe ratings,which
arelowerthan or equaltothesamplemedian
willhavetobefairratings.Then,thesample
median
ofthecontaminatedset,willbeidentical totheA"'orderstatisticofthesetof
n
fairratings,where
k=(n+l)/2.Ithasbeen
shown
(Cadwell 1952)that,as the sizen ofthesampleincreases,thek"'orderstatisticofasample
drawn
fromanormaldistributionN{/U,
O)
convergesrapidlytoanormaldistributionwithmean
equaltotheq"'quantileoftheparentdistribution
where
q=k/n. Therefore,forlargeratingsamplesn,undertheworstpossible unfairratings strategy, thesample
median
ofthecontaminatedset willconvergetoX
where
X
isdefined by:k n
+
\ (n+
iwhere
q=
—
=
1
2 (!-(?)
and
O
(q) isthe inversestandardnormalCDF.
1
2 (1-tJ)
(9)
Given
thatR'
=
U
theasymptotic formulafortheaveragereputation bias achievableby
bjair ^
S
100%
unfairratingswhen
fairratings aredrawn
fromanormaldistribution N{/U.,o)
and
unfairratingsfollow the
most
disruptivepossible unfairratings distribution,isgivenby;1
E[B
- max]-*= E[Rl
p.actual-R'
b.jto-
0-'(-:) (10)
2
(\-sy
Figure3
shows
some
ofthe valuesofE\B
1forvariousvaluesofS
and <7 inthe specialcasewhere
° "- max-•
ratingsrange
from
to9.Given
thatwe
haveassumed
thatallratings in thenearestneighborsetcorrespondtousersinthe
same
taste cluster, itisexpectedthatthe standard deviationofthefairratings willberelativelysmall. Therefore,
we
didnot consider standard deviations higher than10%
ofthe ratingsrange. It isobviousthatthe
maximum
biasincreaseswiththepercentage ofunfairratingsand isdirectlyproportional tothe standard deviationofthefairratings.
As
before,we
havehighlightedmaximum
averagebiasesof
5%
oftheratingrangeormore.Figure3clearlyshows
thattheuseofthesamplemedian
as a thebasisofcalculating
MREs
manages
toreducethemaximum
averagebiastobelow
5%
oftheratingrange forunfairrater ratiosofupto30-40%
andawide
range offairratingstandard deviations.Percentage
of
4.3 Using frequencyfilteringtoeliminate unfairratingsflooding
Formulas(6)and(10)confirmtheintuitive factthatthereputation biasduetounfairratingsincreaseswith
the ratio
5
ofunfairratersinagivensample.In settingswhere
a seller'squalityattributescan vary overtime (mostrealisticsettings),calculationofreputationshouldbebased
on
recentratingsonly usingtimediscounting ora
time-window
approach.Inthosecases.Section3demonstratedthatby
"flooding"thesystem withratings,arelativelysmall
number
ofunfairraterscanmanage
to increasethe ratio5
ofunfairratings inany given time
window
above50%
and completelycompromise
thereliabilityofthe system.Thissectionproposes an approachfor effectively
immunizing
areputationreportingsystemagainst unfairratings flooding.
The main
ideaistofilterraters in thenearestneighborsetbasedon
theirratingssubmissionfrequency.
Description
of
frequencyfilteringStep 1:Frequencyfilteringdepends
on
estimating theaverage frequency ofratingssubmittedby
each buyerforagivenseller. Sincethis frequencyisatime-varying quantity(sellerscan
become
more
orlesspopularwiththepassageoftime),it,tooneedstobeestimatedusingatime
window
approach.More
specifically:1
.
Calculatethe set
F'
{t)o^buyer-specificaverageratingssubmissionfrequenciesf^
(t) for sellers,foreachbuyer bthathassubmittedratings for sduringthe ratingssubmission frequencycalculation
time
window
W
.=[t-E, i].More
precisely,/'
(/)= (number
ofratingssubmittedforshy
bduringW
)IE (1 1)2. Setthecutofffrequency
/'
(?) tobeequaltothe A-thorderstatisticofthesetF'
(/)where
k
=
(\-D)
•«
,n isthe
number
of elements ofF'
(t)andD
isaconservative estimateofthefractionofunfairraters inthetotalbuyerpopulationfor sellers.For example,if
we
assume
thatthereareno
more
than 10%
unfairratersamong
allthebuyersfor sellers,thenD=0
1.Assuming
further thatn=100,i.e. thatthe
setF"
(?)containsaverageratingssubmissionfrequenciesfrom
100buyers, thenthecutofffrequency
would
beequal tothe 90-th smallestfrequency(the 10-thlargestfrequency)presentintheset
F\t)
.The
widthE
ofthe ratingssubmission frequencycalculationtimewindow
W
shouldbelargeenough
inStep2: Duringthecalculationoi~a
MRE
for sellerj',eliminateallratersb inthenearestneighborsetforwhom
f^> f'
. Inotherwords,elimmateallbuyerswhose
averageratingssubmission frequency forsellersisabovethecutoff frequency.
Analysis
of
frequencyfilteringWe
willshow
thatfrequencyfilteringprovideseffectiveprotection against unfair ratings floodingby
guaranteeingthatthe ratioofunfairraters inthe
MRE
calculationsetcannotbemore
than twiceaslarge asthe ratioofunfairratersin thetotalbuyerpopulation.
As
before,we
willassume
thattheentirebuyerpopulationisn,unfairratersareS
n
«
n
andthewidth ofthe reputation estimationtimewindow
isa relativelysmall ff (sothat,eachratingwithin fFtypicallycomes
fromadifferentrater).Then,afterapplyingfrequencyfiltering to thenearestneighborsetofraters,inatypicaltime
window
we
expecttofindW
(\-5)
n \u (p{u) dii fairratings,where
(p(u)isthe probabilitydensityftinctionoffairratingsfrequencies,andat
most
W
5
na
f unfairratings,where
<
«
<
1 isthe fractionofunfairraterswith submission ^cutoff °frequencies
below
/^^,„.
Therefore, theunfair/fairratings ratio in the final set
would
beequalto:unfairratings
_
S'_
5
^'Jcuiaff_
5
,.-.fairratings
\-5'
\-
5
''« 1-
8
\u (p{u)du
a-f
where
/=
'^^ denotestheinflationoftheunfair/fairratings ratio inthefinalset relative toits\u(p{u) du
valueinthe originalset.
The
goalofunfairratersisto strategically distribute their ratingsfrequenciesabove and
below
thecutofffrequencyinordertomaximize
/.Incontrast,thegoalofthemarketdesigneristopick the cutofffrequency
/__^,,^ soas tominimize/.
The
cutofffrequencyhasbeendefinedas the(/-D/n-thorderstatisticofthesample of buyerfrequencies,where
D>5.
Forrelativelylargesamples,thisconvergestothe^-th quantileofthefairratingfrequencies(\-D)
n=
qi\-S)
n+
aS
n =>q=
1-D
+ (a-l)
(13)1-0
From
thispoint on,theexact analysis requiressome
assumptions aboutthe probability densityfijnctionoffairratingsfrequencies.
We
startby
assumingaunifonndistributionbetween F^^^=
f^/(I+
s) andF
=
/
(1+
5). Let5 =
F
-
F
.Then,by
applyingthepropertiesofunifonnprobabilitydistributions toequation(12),
we
getthefollowing expressionofthe inflation/ofunfairratings:/=^""V
where/
-F
^^iS^Z^s
(14)cutuj/ min
After
some
algebraicmanipulationwe
find that—
>
and—
>
.Thismeans
that,unfairraters willda
dD
wantto
maximize
a,the fractionofratings thatarelessthan orequalto f^^^^^,whilemarket makerswillwanttominimize D,i.e. set
D
ascloseas possibletoanaccurateestimateoftheratioofunfairraters inthetotalpopulation.Therefore,atequilibrium,
a
=
\,D=
S
and;2(F
-eS)
5
1=
!!2iwhere
f=
(15)(\-£)(F
^ ' ^ mmmin+F
mm
max-£
S)l-S
The
aboveexpressionfortheunfair/fairratings inflationdependson
thespreadS offairratings1 2
frequencies.
At
the limitingcaseswe
get lim/=
and lim/=
5-»o
i-£
s^-i-e
By
substituting theabove
limiting valuesof/inequation(12),we
get thefinal fonnulaforthe equilibrium relationshipbetween6,the ratioofunfairraters inthetotalpopulationof buyers and5' the final ratioofunfairratings
m
thenearestneighborsetusingtimewindowing
and frequency filtering:S/(\-S)<S'<2S
(16)Equation(16)
shows
that,nomatterhow
hardunfairratersmay
tryto"fiood"thesystem withratings,thepresenceof frequencyfilteringguaranteesthattheycannotinflatetheirpresencein the final
MRE
calculationset
by more
than afactorof2.This concludestheproof
In
most
onlinecommunities,the exactratio5 ofunfairraterswillnotbeknown
exactly. Insuchcases,ifwe
haveabeliefthat5<0.1, thensettingD=0.1
hasbeenexperimentallyprovento result in inflation ratios,which
alsofallwithin thebounds
ofequation(16).A
more
realisticassumption aboutfairratingsfrequenciesisthattheyfollow alognormaldistributionwithmean
/
andvariance relatedtothefrequencyspreadS.This assumptionisconsistentwiththe findingsofgiveninclosed form.
However,
anumericalsolutionyields results,which
approximatevery closely those obtainedanalytically foruniformlydistributedfairratingfrequencies (Figure4).0.001 0.01 0.1 1 10
Frequency
spread100 1000
-B—
Uniformdistributionx
Log-normaldistributionFigure4.
Maximum
unfairratings inflation factorswhen
frequencyfilteringisused(S
=
D
=
0.l).Given
thatmedian
filteringguarantees reputation biases oflessthan5%
oftheratingsscale(e.g.lessthan±0.5 points
when
ratingsrange from 1-10)forcontaminationratiosofupto30-40%
and frequencyfilteringguaranteesthatunfairraterscannotuse floodingto inflate theirpresence
by more
thana factoroftwo,thecombinationof frequencyfilteringand
median
filteringofguarantees reputation biasesoflessthan5%
when
the ratioofunfairratersisupto15-20%
ofthe totalbuyerpopulationforagivenseller.One
possiblecriticismofthefrequencyfilteringapproachisthatitpotentiallyeliminatesthosefairbuyerswho
transactmost
frequentlywithagivenseller. Infact, intheabsence ofunfairraters,allraterswho
would
befilteredoutbasedon
theirhighratingssubmission frequencywould
befairraters.Nevertheless,we
do
not believethat thispropertyconstitutesaweakness
ofthe approach.We
arguethatthe "bestcustomers" ofagivenselleroften receivepreferentialtreatment,
which
isinaway
aform ofpositivediscrimination
on
behalfoftheseller. Therefore,we
believethat the potentialeliminationof suchratersfromthefinalreputation estimateinfactbenefitstheconstructionof
more
unbiasedestimatesforthebenefitoffirst-timeprospective buyers.
4.4 Issuesincommunities
where
buyeridentityisnot authenticatedThe
effectivenessof frequencyfiltering reliesontheassumptionthatagivenprincipalentitycan only haveone buyeragent actingonitsbehalfinagiven marketplace.
The
techniqueisalso validinsituationswhere
principalentitieshavemultiplebuyeragentswithauthenticatedidentifiers. In that case,frequencyfiltering
Innon-aulhenticated onlinecommunities (communities
where
"pseudonyms"
are"cheap",tousetheterm ofFriedman
and Resnick) withtime-windowed
reputation estimation, unfairbuyers canstillmanage
to"flood"thesystem withunfairratingsbycreating a large
number
ofpseudonymously
known
buyeragents actingon
theirbehalfIn thatcasethetotalratio5 ofunfairagentsrelative tothe entirebuyerpopulationcanbe
made
arbitrarilyhigh. Ifeachofthe unfairagents takes careofsubmitting unfairratings forsellerswithfrequency
J
'< J
,because 6willbehigh,eveninthepresenceof frequencyfiltering,imfairraterscanstill
manage
toseverelycontaminatea seller'sfairreputation.Furtherresearchisneededinordertodevelopimmunizationtechniquesthatareeffective incommunities
where
the"true"identityofbuyeragentscannotbeauthenticated.Inthemeantime,theobservationsofthissection
make
astrongargumentforusingsome
reasonablyeffectiveauthenticationregimeybr
buyers(forexample,requiringthatall
newly
registeringbuyers supplyavalidcreditcardforauthentication purposes)inall onlinecommunities
where
trustisbasedonreputationalinfomiation.5.Conclusions
and
Management
ImplicationsWe
beganthispaperby
arguingthatmanagers ofonlinemarketplacesshould pay specialattention tothedesignofeffectivetrust
management mechanisms
thatwillhelpguaranteethestability,longevityandgrowth oftheirrespectivecommunities.
We
pointed outsome
ofthe challengesof producingtrust inonlineenviroiunentsand arguedthatonline reputation reporting systems,an emergingclassofinformation systems,holdthepotentialof
becoming
aneffective,scalable,andrelatively low-costapproachforachievingthisgoal,especially
when
thesetof buyers andsellersislargeandvolatile.Understandingtheproper implementation,usageandlimitationsof such systems(inordertoeventually
overcome
them)isthereforeimportant,both forthe
managers
aswellas for theparticipantsofonlinecommunities.Thispaperhas contributedin this direction,first,
by
analyzing thereliabilityofcurrent-generationreputationreportingsystemsinthepresenceof buyers
who
intentionallygive unfairratings to sellersand, second,by proposingandevaluating asetof "immunization mechanisms",which
eliminate orsignificantlyreducetheundesirableeffectsof such fraudulentbehavior.
InSection3,themotivations forsubmitting unfairratings
were
discussedandthe effectsof suchratingsonbiasinga reputation reportingsystem's
mean
reputation estimateofa sellerwere
analyzed.We
haveconcludedthatreputationestimation
methods
basedontheratingsmean, which
arecommonly
usedincommercial
recommender
systems, areparticularlyvulnerabletounfair ratingattacks,especiallyincontexts
where
aseller'sreputationmay
varyovertime.techniquesproposed
by
thiswork
willprovideauseful basisthat willstimulatefurtherresearchinthe importantand promisingfieldofonline reputation reporting systems.References
Arrow,
Kenneth
(1963). SocialChoiceand
Individual Values.YaleUniversityPress.Bakos,Y. (1997).
Reducing
Buyer
SearchCosts: ImplicationsforElectronicMarketplaces.Management
Science,
Volume
43, 12,December
1997.Bakos,Y. (1998).
Towards
Friction-FreeMarkets:The Emerging
RoleofElectronicMarketplaceson
theInternet.
Communications of
theACM,
Volume
41,8 (August1998),pp. 35-42.Billsus,D.andPazzani,M.J. (1998).Learningcollaborativeinformationfilters. InProceedings
of
the15'''InternationalConference
on
Machine
Learning,July 1998,pp.46-54.Boon, SusanD.,
&
Holmes, John G.(1991).The dynamics
ofinterpersonaltrust: resolving uncertaintyinthefaceofrisk. Pages 190-211
of
Hinde,RobertA.,&
Groebel, Jo(eds),Cooperationand
ProsocialBehaviour.
Cambridge
UniversityPress.Bresee,J S,
Heckerman,
D.,and Kadie, C.(1998)EmpiricalAnalysisofPredictiveAlgorithmsfor CollaborativeFiltering.InProceedings ofthe14'''Conference on UncertaintyinArtificialIntelligence(UAI-98),pp.43-52,SanFrancisco, July 24-26, 1998.
Cadwell,J.H.(1952)
The
distributionofquantilesofsmall samples. Biometrika. Vol. 39,pp.207-211.Cranor, L.F.andResnick, P.(2000). Protocolsfor
Automated
NegotiationswithBuyer Anonymity
andSellerReputations.
To
appearinNetnomics.Dellarocas,C. (2000).
The
Design ofReliable TrustManagement
Systems forOnline TradingCommunities.
Working
Paper, availablefromhttp://ccs.mit.edu/dell/trustm(;t.pdfDunn,
John.(1984)The
concept of'trust' inthe politicsof John Locke. Chap. 12,pages279-301
of
Rorty, Richard,Schneewind,J.B.,&
Skinner,Quentin(eds),PhilosophyinHistory.Cambridge
UniversityPress.Fayyad,U.M.,Piatetsky-Shapiro,G.,Smyth, P. and Uthurusamy,R.eds.(1996)
Advances
inKnowledge
Discovery
and
Data
Mining,MIT
Press, Cambridge, Mass.Friedman,E.J.andResnick,P. (1999)
The
SocialCostofCheap Pseudonyms. Working
paper.An
eariierversion
was
presentedattheTelecommunicationsPolicyResearch Conference, Washington,DC,
October1998.
Gambetta,
Diego
(ed).(1990). Trust.Oxford:Basil Blackwell.Gordon, A.D. (1999)Classification.
Boca
Raton:Chapman
&
Hall/CRC.Goldberg,D.,Nichols,D.,Oki,B.M.,andTerry,D.(1992) UsingCollaborativeFiltering to
Weave
anInfonnation Tapestry.
Communications
oftheACM
35(12), pp. 61-70,December
1992.Hojo,T.(1931). Distributionofthemedian,quartilesandinterquartiledistanceinsamples fromanormal
population. Biometrika, Vol.23, pp. 315-360.
Hutt,A.E., Bosworth,S.andHoyt.D.B.eds.(1995).
Computer
SecurityHandbook
(3"^edition).Wiley,New
York.Jain,A.K., Murty,
M.N.
andFlynn, P.J.(1999) Dataclustering:areview.ACM
Computing
Surveys, Vol.31, 3(Sep. 1999),pages
264 -323.
Johnson,D. R.and PostD.G.(1996).
Law And
Borders—The
RiseofLaw
inCyberspace. StanfordLaw
Review,Vol. 48.
Kaufman,
L.andRousseeuw,
P.J.(1990).FindingGroups
inData:An
IntroductiontoCluster Analysis.Wiley,
New
York.Kollock,P.(1999)
The
Production ofTrustinOnlineMarkets.InAdvances
inGroup
Processes(Vol. 16),eds. E.J.Lawler,
M. Macy,
S. Thyne, andHA.
Walker,Greenwich,CT:
JAIPress.Lawrence,R.J. (1980)
The Lognormal
DistributionofBuying
FrequencyRates.Journalof
MarketingResearch.Vol.XVII,
May
1980,pp.212-226Malinvaud, E.(1966).Statistical
Methods of
Econometrics.Paris:NorthHolland.Pindyck,R.andRubinfeld,D.L.(1981).Econometric
Models
and
Economic
Forecasts(2"''Edition).McGraw-Hill,
New
York.Resnick,P.,lacovou,N.,Suchak, M., Bergstrom,P.,andRiedl, J.(1994) Grouplens:
An
Open
ArchitectureforCollaborativeFilteringof Netnews.InProceedings
of
theACM
1994
Conferenceon Computer
Supported Cooperative Work,pp. 175-186,
New
York,NY:
ACM
Press.Resnick,P.and Varian,H.R.(1997).
Recommender
Systems.Communications of
theACM,
Vol.40
(3),pp. 56-58.
Sarwar,B.M.,Karypis,G.,Konstan,J.A.,andRiedl,J.(2000)Applicationof Dimensionality Reduction
in
Recommender
System
-A
Case
Study.InACM
WebKDD
2000
Web
Mining
forE-Commerce
Workshop.Schmalensee,R. (1978).Advertisingand?rod.vic\QMa\\Ky.Journal
of
PoliticalEconomy,
Vol.86, pp.485-503.
Sen,A. (1986). Social choicetheory. In
Handbook
of
Mathematical Economics,Volume
3. ElsevierSciencePublishers.
Schafer,J.B.,Konstan,J.,andRiedl,J., (2001)Electronic
Commerce Recommender
Applications.Journal ofData Mining
and
Knowledge
Discovery. January,2001 (expected).Shapiro, C.(1982)
Consumer
Infonnation,ProductQuality,andSellerReputation. BellJournalof
Economics
13 (1),pp
20-35,Spring1982.Shardanand,U.and Maes,P.(1995). Social informationfiltering: Algorithmsforautomating
"word
ofmouth". InProceedings
of
theConferenceon
Human
FactorsinComputing
Systems (CHI95), Denver,CO,
Smallwood,
D.andConlisk,J.(1979).ProductQualityinMarketsWhere
Consumers Are
Imperfectly Infomied.QuarterlyJournal of Economics.Vol.93, pp. 1-23.Wilson,Robert(1985). Reputationsin
Games
andMarkets. InGame-TheoreticModels of
Bargaining.edited
by
AlvinRoth,Cambridge
UniversityPress, pp.27-62.Yahalom,
R., Klein.B
,andBeth,T. (1993). Trust RelationshipsinSecureSystems -A
Distributed Authentication Perspective.InProceedingsof
theIEEE
Symposium on
ResearchinSecurityand
Privacy,Oakland, 1993.
Zacharia,G.,
Moukas,
A.,and Maes,P.(1999)CollaborativeReputationMechanisms
inOnlineMarketplaces.InProceedings
of
32"''Hawaii
InternationalConferenceon
System Sciences (HICSS-32),Maui, Hawaii, January 1999.
Footnotes
'
Jonathan
Lebed,
a15-year-old
boy was
sued
by
theSEC
forbuying
largeblocks
of
inexpensive,
thinlytraded
stocks,posting
falsemessages
promoting
thestocks
on
Internetmessage
boards
and
then
dumping
thestocks
afterprices rose, partlyas
a resultof
hismessages.
The
boy
allegedly
earned
more
than
aquarter-million
dollarsin lessthan
sixmonths and
settledthelawsuit
on September
20,2000
for$285,000
(Source:
Associated
0£C
^^00?
^
MIT LIBRARIES