2026年人工智能如何侵蚀人类能动性:形式模型研究报告(英文版)_第1页
2026年人工智能如何侵蚀人类能动性:形式模型研究报告(英文版)_第2页
2026年人工智能如何侵蚀人类能动性:形式模型研究报告(英文版)_第3页
2026年人工智能如何侵蚀人类能动性:形式模型研究报告(英文版)_第4页
2026年人工智能如何侵蚀人类能动性:形式模型研究报告(英文版)_第5页
已阅读5页,还剩51页未读 继续免费阅读

下载本文档

版权说明:本文档由用户提供并上传,收益归属内容提供方,若内容存在侵权,请进行举报或认领

文档简介

ALVINMOONANDBENJAMINBOUDREAUX

AFormalModel

ofHowAIErodes

HumanAgency

iii

AboutThisReport

Inthisreport,wedevelopaformalframeworkformeasuringhowartificialintelligence(AI)affectscollec-tivehumanagencyindecisionmakingprocesses.Wedefinecollectiveagencyashumanity’scapacitytomakeconsequentialchoicesaboutsharedresources,institutions,andthesocialenvironment.

Drawingonsocialchoicetheory,wemodeldecisionmakingintermsofcoalitions:groupsthathavedeci-siveinfluenceoveroutcomes.WeproposequantitativemetricsfortrackinghowagencyconcentratesovertimeandidentifythreemechanismsthroughwhichAIsystemsreducehumanagency:humandisenfran-chisement,AIenfranchisement,andAIagendacontrol.

Thisreportisintendedforpolicymakers,AIresearchers,AIdeployers,andotherstakeholderswhoneedrigoroustoolsforanticipatingandrespondingtoAI’seffectsonhumandecisionmakingcapacity.

CenterfortheGeopoliticsofArtificialGeneralIntelligence

RANDGlobalandEmergingRisksisadivisionofRANDthatdeliversrigorousandobjectivepublicpolicyresearchonthemostconsequentialchallengestocivilizationandglobalsecurity.Thisworkwasundertakenbythedivision’sCenterfortheGeopoliticsofArtificialGeneralIntelligence(AGI),whichiscommittedtohelpingdecisionmakersunderstand,anticipate,andpreparetonavigatethenationalsecurityandgeopoliticalimplicationsofAGI.Thecenterconvenesleadingtechnologists,strategists,economists,politicalscientists,andoutsideexpertstoconsiderthefeasibilityandeffectivenessofprospectiveAGI-enabledcapabilities;thedomesticandinternationalimplicationsoftheiruse;andthestrategiesandpoliciesthatgovernments,busi-nesses,andcivilsocietycouldadopttorespondtonewrealities.Formoreinformation,visit

/

geopolitics-of-agi

.

Funding

ThisresearchwasindependentlyinitiatedandconductedwithintheCenterfortheGeopoliticsofArtifi-cialGeneralIntelligenceusingincomefromoperationsandgiftsfromRANDsupporters,includingphil-anthropicgiftsmadeorrecommendedbyDALHAPInvestmentsLtd.,ErgoImpact,FoundersPledge,Char-lottesochFredriksStiftelse,GoodVentures,Longview,andCoefficientGiving.RANDdonorsandgrantorshavenoinfluenceoverresearchfindingsorrecommendations.

Acknowledgments

WethankJoelPreddandtheGeopoliticsofAGIteamformanyhelpfulconversations.WeareespeciallythankfultoEdwardGeistfordiscussionsonmodelingandagency.ThisreportalsobenefitedgreatlyfromcommentsandhelpfulsuggestionsfromreviewersRobertLempert,ofRAND,andRaymondDouglas,ofACSResearch.Allerrorsareourown.

v

Summary

Asartificialintelligence(AI)systemsassumemoredecisionmakingrolesingovernment,theeconomy,andsociety,aquestionemerges:Willhumansretainthecapacitytoshapecollectiveoutcomes?Severaltheo-riessuggestthat,oncehumandecisionmakingerodespastacertainthreshold,theskills,institutions,andpoliticalstandingneededtoreclaimthatdecisionmakingcapacitymaynolongerexist.However,nowidelyacceptedmetricsexistfortrackingthiserosion.Inthisreport,wedrawonsocialchoicetheorytodevelopaformalmodelofhowAIerodescollectivehumanagency;wealsomodeldecisionmakingintermsofcoali-tionsandproposequantitativemetricsfortrackingshiftsinthedistributionofdecisionmakingpowertoidentifythepointbeyondwhichthoseshiftscouldbecomeirreversible.

KeyFindings

Thefollowingkeyfindingsarosefromourresearch:

•Agencyerosionismeasurableacrossdomains.Wedefinethreemetrics—distributionofdecisivecoalitions,minimalcoalitionsize,andcompositionofminimalcoalitions—thatareapplicabletoawidevarietyofdecisionmakingprocesses.Thesemetricsprovideaframeworkfortrackinghumanagencyimpactsacrossdomains,comparingchanges,anddetectingnonlinearaccelerationofagencyerosion.

•Distinctmechanismsdriveagencyerosion.WeidentifyandformallyvalidatethreepathwaysthroughwhichAIreduceshumanagency:humandisenfranchisement(fewerhumansindecisionmakingroles),AIenfranchisement(AIentitiesgainingdecisionmakingpowerandchangingthecompositionofdeci-sivegroups),andAIagendacontrol(AIsystemsshapingwhichalternativesreachhumandecisionmak-erstoconsolidatepowerinunintendedways).

•Aterminalstateexists.Themathematicalstructureofourmodelidentifiesaformalendstateofagencyerosion:asingleminimalcoalitionthatisdecisiveforallchoices.Thisprovidesatargetformonitoringhowfarthetrajectoryisfromthispointofirreversibility.

Recommendations

Wemakethefollowingrecommendations:

•Developagencyevaluations.ExistingAIevaluationsassesscapabilities,safety,andalignment,buttheydonotassessstructuraleffectsonhumandecisionmaking.ResearchersshoulddesignbenchmarksthatmeasurewhenAIsystemsreducethenumberofhumansindecisivecoalitions,influenceoutcomes,orshapewhichalternativesreachhumandecisionmakers.

•Establishhumanparticipationthresholds.Forhigh-stakesdomains(suchasdemocraticgovernance,militaryapplications,andcriticalinfrastructure),policymakersshouldconsiderminimumrequire-mentsforhumanpresenceindecisivecoalitions.Thesethresholdsshouldreflectdomain-specificrequirementsforlegitimacyandreversibility.Organizations,suchastheNationalInstituteofStandardsandTechnology,couldincorporatecoalitioncompositionasameasurabledimensionofAIriskalong-sideexistingmetricsforreliabilityandsecurity.

AFormalModelofHowArtificialIntelligenceErodesHumanAgency

vi

•Monitorcoalitioncompositionlongitudinally.Thedangerofgradualdisempowermentisthatnosinglechangeappearscatastrophic.Organizationsandgovernmentsshouldtrackthehumancomposi-tionofdecisivecoalitionsacrossdomainsovertime.

•Benchmarkreversibilitycapacity.OrganizationsshouldassesswhethertheycouldrestorehumandecisionmakingifAI-drivenagencylossaccelerates.Doingsowouldrequiremaintainingthehumanexpertise,institutionalknowledge,anddeliberativeinfrastructureneededtoreversecourse.

vii

Contents

AboutThisReport iii

Summary v

Figures ix

AFormalModelofHowArtificialIntelligenceErodesHumanAgency 1

ResearchQuestionsandScope 1

ReportStructure 3

DynamicsofHumanAgencyLoss 3

IrreversibilityofAgencyLoss:DefinitionsandImplications 4

ACoalition-BasedModelofCollectiveAgency 5

CollectiveAgencyasaModelFeature 6

AnEndStateofAgencyLoss 9

ValidatingMechanismsThatReduceCollectiveHumanAgency 10

LimitationsoftheModel 13

Applications 14

FindingsandRecommendations 16

APPENDIXES

A.MathematicalDetails 19

B.CalculationsforEconomicDisenfranchisementScenario 21

C.AIAgendaControlThroughtheIntroductionofAlternatives 23

References 25

AbouttheAuthors 29

ix

Figures

1.DecisiveCoalitionsinaMajorityRuleSituation 8

2.DecayintheDistributionofDecisiveCoalitionsforMajorityRuleSystems 8

3.DecayinMinimalSizeofDecisiveCoalitionsforMajorityRuleSystems 9

4.DecisiveCoalitionsinanExampleVotingSystemwithThreeIndividuals 10

5.AnOrganizationwithSharedAgenticAIResources 13

6.PossibleChangestoDecisionmakersinGroupG 15

1

AFormalModelofHowArtificialIntelligenceErodesHumanAgency

Theadoptionofpowerfulartificialintelligence(AI)systemswillchangewhomakesdecisionsandhowtheydoso.AsAIisdeployedinthegovernmentandtheeconomyinwaysthatautomatedecisions,ahardpolicyquestionemerges:Howcanchangestohumanity’scollectivecapacitytoshapethefuturebemeasured?

SeveralAIresearchersarguethatthetransitiontopowerfulAIsystemscouldpermanentlydiminishhumanagency.SomewarnofagradualdisempowermentinwhichAIsystemsslowlydisplacehumanpar-ticipationacrosseconomic,cultural,andpoliticaldomainswithouttherebeinganysinglecatastrophicevent(Kulveitetal.,2025).OthersarguethatsufficientlycapableAIsystemswillacquireresourcesandpowerasinstrumentalgoals,potentiallyculminatinginrapidhumandisplacement(Carlsmith,2023;Ngo,Chan,andMindermann,2024).Thesetheoriesdifferinmechanismandtimelines,buttheyshareatroublingpredictionthatonceagencyislost,recoverymaybeimpossible.

Despitethisconvergenceonirreversibility,existingframeworksdonotoffertheprecisionnecessaryforstrategicpolicyplanning.Withoutacommonmetric,policymakerscannotcomparescenarios,tracktrends,oridentifyinterventionpointsbeforechangesbecomepermanent.Meanwhile,novelandcapableAIsystemsmaybringsignificantbenefits,especiallytofirstmovers,whichcouldincentivizeAIdevelopersandotherstakeholderstoacceleratedevelopment(Abraham,Kavner,andMoon,2025)andunderscoretheurgencyofunderstandinghowagencymaybeaffected.

Inthiscontext,themotivationforrigorouslystudyingAI’seffectsonhumanagencyisclear.AformaltheoryofagencymayinformdecisionsabouthowtobestbenefitfromadoptingAItechnologieswhilehedg-ingagainstunintendedorundesirableoutcomes,suchasirreversiblelossofhumanagencyinkeydomains.

ResearchQuestionsandScope

Inthisreport,weconsidercollectiveagencytobethecapacityforhumanstomakedecisionsaboutsharedresources,institutions,andthephysicalandsocialenvironment.Agencydiminisheswhenhumanslosethecapacitytocollectivelyinfluenceoutcomesbasedontheirchosenandexistingdecisionmakingprocesses.Toexploreconditionsforirreversibleagencyloss,weaskthefollowingthreequestionsaboutcollectiveagency:

1.Howdochangestodecisionmakinggroups(suchaschangestotheirsizes,compositions,andavail-ablealternatives)translateintomeasurablelossesofcollectivehumanagency?

2.Whatindicatorsmightallowearlydetectionandinterventionbeforechangesbecomeirreversible?

3.Isthereaterminalstatetowardwhichthesedynamicstrend,andcanwecharacterizeitformally?

Answeringthesequestionsisurgent;somedeploymentdecisionsincriticalinfrastructure,governance,andeconomicinstitutionsmaybeeffectivelypermanent(wefurtherdiscussthisissueinthesectiontitled“Irre-versibilityofAgencyLoss:DefinitionsandImplications”).

Tobeginaddressingthesequestions,weproposeaformalframeworkformeasuringcollectivehumanagencyincertaindecisionmakingprocesses.Wedrewonsocialchoicetheorytodevelopacoalition-basedmodelthattranslatesintuitionsaboutagencylossintoquantifiablemetrics.Then,givenadecisionmakingprocessthatfallswithinthescopeofourframework,thismodelcanassesshowchangestoparticipants,theirrelativeinfluence,orthealternativesthattheyconsiderwillaffectcollectivehumanagency.

AFormalModelofHowArtificialIntelligenceErodesHumanAgency

2

Whatprocessesaremodeledbyourframework?Decision-theoreticanalysestypicallydividedecisionproblemsintotwodistinctsetsoftasks(Gongetal.,2017).Decisionstructuringtasksdefinetheproblem,identifyobjectives,andassembleamenuofoptionsforachievingthoseobjectives.1Choicetaskstheninvolveselectingthebestoptionfromthatmenu,givenestimatesofconsequences.Humanagencyencompassesbothtasks:theabilitytoassesswhatmattersandtheabilitytochooseamongalternatives.

AIsystemsmightchangeagencythroughbothchannels.AnAIthatcontrolsdecisionstructuring(e.g.,byshapinghowproblemsaredefined,whichobjectivesareprioritized,orwhichoptionsareconsidered)candetermineoutcomesbeforeanychoicetaskbegins.AnAIthatcontrolsthechoicetask(e.g.,byreplacinghumanparticipantsintheprocessthatselectsamongoptions)erodesagencyevenwhentheframingisdonebyhumans.

Theformalmodelwedevelopinthisreportprimarilyaddressesthechoicetasksideofthisdistinction.Specifically,wemodeldecisionmakingprocesseswiththefollowingtwofeatures:

•Decision-structuringtasksarelimitedtochangingwhichoptionsareconsidered,suchasbyeliminatingoraddingoptions.Wedonotconsidertasksthatidentifyproblemsandobjectivesorotherwiseinflu-encehowproblemsandobjectivesareset.

•Choicetaskscanbemodeledasaggregatingprocessesthatusemultipleindividualpreferencesandinputs.

Thisscopeisdeliberatelynarrow.Theagendacontrolmechanism(discussedinlatersectionsandAppen-dixC)partiallybridgesthegapbymodelinghowAIcanmanipulatewhichalternativesreachhumandeci-sionmakers,butitdoesnotcapturethedeeperdynamicthroughwhichAImightreshapehowhumanscon-ceptualizeproblemsinthefirstplace.Extendingformalagencymeasurestodecisionstructuringremainsanimportantdirectionforfuturework.

Withinourscope,theclearestexampleofamodeledprocessisavotingsystem,althoughwediscusshowtheframeworkgeneralizestosystemsoutsideavotingsetting,suchashuman-in-the-loopsystems(i.e.,semi-automatedorfullyautomatedsystemsinwhichhumandecisionmakersinteractwithautomatedprocesses).Anexampleofanout-of-scopeprocessisahumanusinganAIsystem(suchasalargelanguagemodelchat-bot)toinformindividualdecisions:Itschoicetasksdonotinvolveaggregationacrossmanyindividualprefer-ences,andsuchaprocessisthereforeoutofscope.AlthoughAI-humanchatbotinteractioncouldinfluenceaggregatedecisions,theimpactsonindividualsaredistinct(Treyger,Matveyenko,andAyer,2025)andlikelyrequireadifferentcharacterization(Sharmaetal.,2026).

Withinthescopeofourframework,wefindthefollowingthreepreliminaryanswerstoourquestions:

1.Inseveralimportantexamplescenarios,ourquantifiablemetrics(whichmeasuresize,composition,anddistributionofdecisivegroupswithinthemodel)decreasewithchangestothesystemthatshouldheuristicallydecreasecollectiveagency.

2.Quantitativeorqualitativethresholds,suchasratesofchangeinourmetricsinresponsetochangestodecisionmakingprocesses,mayindicatetippingpointsatwhichcollectiveagencylossisirreversible.

3.Ourframeworkfeaturesaterminalstateofhumanagencyloss—specifically,theexistenceofasingleminimallysizeddecisivegroup.

Becausethescopeofourframeworkisnarrowerthanthefullgeneralityrequiredtodiscussallaspectsofdecisionmaking,weleaveseveralinterestingandopenquestionsforfutureresearch.Anobviousbutimpor-

1Incomparison,ethicists,suchasNussbaum(2000;2011),recognizethesetasksaspracticalreasonorthecapacitytocriti-callyreflectandformaconceptionofthegood.

AFormalModelofHowArtificialIntelligenceErodesHumanAgency

3

tantquestioniswhatotherframeworkscanextendourformalunderstandingofcollectiveagencytoencom-passbroadergroupsofdecisionmakingprocesses.Anotherimportantquestion,whichwediscussinourrecommendationstopolicymakers,isdetermininghowmuchhumanagencyshouldberequiredindifferentdecisionmakingdomains.

Ultimately,ourmodelcontributesafirststeptowardformallyunderstandingtheeffectsofAIdeploy-mentsonhumanagency.

ReportStructure

Thisreportisorganizedasfollows.WereviewliteratureonAI-drivendisempowerment,AIpower-seekingbehavior,humancognitiveatrophy,anddecisionmakinglegitimacy.Weestablishirreversibilityasakeyfea-tureofourframework.Then,wederiveourcoalition-basedmodelofcollectiveagencyfromsocialchoicetheory.Wecarefullycircumventknownproblemswiththemodelstemmingfromafeatureofpreferencescalledtransitivityandthecontextofvotingsystems,whichwediscussinthesectiononlimitations.

Next,weapplyourmodeltovalidatethreedynamicsthroughwhichAIreducescollectivehumanagency:humandisenfranchisement(fewerhumansindecisionmakingroles),AIenfranchisement(AIentitiesgain-ingdecisionmakingpower),andAIagendacontrol(AIsystemsshapingwhichalternativesreachhumandecisionmakers).WeconcludewithabriefdiscussionofapplicationstoAIevaluationsandpublicpolicy,followedbyrecommendationsanddirectionsforfuturework.Wealsoaddressfurthermathematicaldetailsintheappendixes.

DynamicsofHumanAgencyLoss

SeveraldynamicssuggestthatAIdevelopmentmaydriveirreversiblelossesofhumanagency.Inthissec-tion,wereviewtherelevantliteraturebeforeestablishingirreversibilityasakeydesignprincipleforourformalmodel.

ThewidespreaddeploymentofAIscoulddrivechangestosocialsystems.Thispossibilityisobviousinscenariosinwhichamisalignedartificialsuperintelligencerapidlyseizespowerandseekstocontroloreradi-catehumanity(seeYudkowskyandSoares[2025],Bostrom[2016],Yampolskiy[2015],andmanyothers).ButwidespreadAIadoptionmayalternativelydecouplesocietalsystemsfromhumanintereststhroughmore-gradualreplacementofhumanparticipationacrosseconomic,cultural,andpoliticaldomains.Kulveitetal.

(2025)arguethatthesesystemssupporthumanprioritiesonlysofarashumanscontributetothem,and,onceAIsystemsoffermore-competitivemachinealternatives,decisionmakerswillfacepressuretominimizehumaninvolvement.Thisdynamicmayunfoldthroughlaborsubstitution,algorithmiccontentgeneration,orautomationofpoliticaljudgment.Similarly,thedegreeofautonomyofAIsystems(i.e.,thescopeandsig-nificanceofdecisionsmadewithouthumanintervention)determineswhetherAIsystemsfunctionaspar-ticipantsormerelyastools.Theoriesofinstrumentalconvergence(Omohundro,2008;Bostrom,2016;Rus-sell,2020;Dung,2025;Carlsmith,2023)furthershowhowgoal-drivenautonomycouldleadtowardresourceorpoweracquisition.Theriskofsuchdisplacementisalsocompoundedbycognitiveatrophy.Automation’s“irony,”asdescribedbyBainbridge(1983)andconfirmedbyLeeetal.(2025),suggeststhatrelianceonAIdullscriticalthinkingandexceptionhandling;Koeszegi(2023)foundthat,withcertaindesigns,algorithmicmanagementcandisempowerworkers.Overtime,lostcompetencemakesmeaningfuloversightmoredif-ficult,andaffectedpartiesmayloseopportunitiesforcontestationandreason-giving.

ThesedynamicssuggestthatAIsystemsmaybecomeeffectiveparticipantsincertaindecisiondomains.Critically,thesedynamicsdonotrequireAIsystemstoactcontrarytohumaninterests.Evenaperfectly

AFormalModelofHowArtificialIntelligenceErodesHumanAgency

4

alignedAIthatreliablyproducesdecisionsthathumanswouldendorsecanerodecollectiveagencybydis-placinghumanparticipation.Thiserosioncanoccurbecausebroadhumanparticipationinconsequentialdecisionsnotonlyisinstrumentallyvaluablebutalsopreservestheskillsandinstitutionalknowledgeneededtogoverneffectively,sustaindiscursivelegitimacy(Habermas,1996;Cohen,1997;GutmannandThompson,2004),andmaintainoptionality,thecapacitytochangecourseifcircumstancesorvaluesshift.AnAIthatconsistentlymakesgooddecisionswhileremovinghumansfromthedecisionmakingprocessmayoptimizeforpresentoutcomeswhileforeclosingfutureones.Tocapturethesepossibilities,ourmodelisagnosticaboutwhetherAIsystemsshareordivergefromhumaninterests;itmeasuresthestructuralcapacityofhumanstoinfluenceoutcomes.

IrreversibilityofAgencyLoss:DefinitionsandImplications

Pressureonsocietalsystemstodecouplefromhumaninterests,power-seekingbehaviorofautonomousAIsystems,andhumancognitiveatrophymayallcontributetofuturedynamicsthatdecreasecollectivehumanagency.Thetheoryofirreversibledecisionsiswell-troddengroundineconomics,startingwithkeystudiesbyHenry(1974)(andbyArrowandFisher[1974])thatestablishedthateconomicdecisionsareoftenirreversible.Henrydefinesanirreversibledecisionasonethat“significantlyreducesforalongtimethevarietyofchoicesthatwouldbepossibleinthefuture”(1974,p.1006).AlthoughHenryoriginallydefinedthistermforeco-nomicanalysis,wefinditsufficientlygeneraltoapplytootherdomains,suchascultureandpolitics,andsoitisdirectlyrelevanttodecisionsthatchangecollectiveagency.Giventhisdefinitionofanirreversibledeci-sion,acriticalquestionarises:Whenarelossestocollectivehumanagencylikelytobeirreversible?TheAIagencyliteraturesuggeststhatthiswillbethecaseforsomescenariosofsignificantagencyloss,sowebuildouragencyevaluationframeworktofocusonwhatdrivesdecisionmakingpowertoconcentrateinsmallergroupsasdeploymentofAIsystemsincreases.Inthissetting,wecanexploretheirreversibilityquestionbyprovidinginsightintowhichconditionsdriveandacceleratetheconcentrationofdecisionmakingpower.

Ourfocusonirreversibilityisbasedonacoupleofobservations.Thefirstobservationisthatmanyofthesechangeswillbeeconomicinnature.Forexample,automationthroughAImaybringdramaticchangestoeconomicsectorsacrossawiderangeofscales,fromindividualworkerstotheglobaleconomy.Bytheprinciplethatsignificanteconomicdecisionsaregenerallyirreversible(Henry,1974;ArrowandFisher,1974),automationthroughAIatthisscalewillbeirreversibletoo.Arecentexamplethatsupportsthistrendisinautomatedstockmarkettrading:Automatedtradingmachineshavesteadilyreplacedhumantradersandarebeingdesignedwithgreaterautonomyintheirroles(BorchandMin,2022).Giventheeconomicadvan-tagesofautomatedtradersandthefactthattradingcompanieshaveinvestedsignificantlyinsoftwareandhardware,aswellasinfindingandexploitingoptimalphysicallocationsfortrading(Wissner-GrossandFreer,2010),wefinditunlikelythattradingcompanieswouldbeabletoeasilyreverttohuman-basedornon-algorithmictrading.Effectively,thischangeisirreversibleaccordingtoHenry’sdefinition.

AsecondobservationisthatAIimplementationinorganizationsandgovernmentswilllikelyeliminatehumanrolesindecisionmakingprocesses.AcorevaluepropositionforAIautomationisthatAIsystemswillbeabletoaccomplishtasksandfillavarietyofroles.Alikelyconsequenceofthisautomationisthatsomehumanpositionswillbeeliminated,2workflowswillchange,andprocesseswillrelymoreheavilyonAIsys-tems.ReversingAIautomationwouldrequirethereconstructionofdismantledcapacityforhumaninvolve-ment,whichwefindisunlikelytobereversible.Forexample,consideringthepreviousdiscussiononhuman

2Thedegreetowhichhumanpositionswillbeeliminatedbecauseofautomationisanactiveareaofresearch.SeeworkbySytsma(2025)thatanalyzedpossibleeffectsofautomationpolicyunderuncertainty.

AFormalModelofHowArtificialIntelligenceErodesHumanAgency

5

cognitiveatrophyandagencyloss,wehypothesizethatthehumancapitalrequiredformeaningfuloversightofAI-drivenprocesseswilldegradewithdisuse.Oncethatknowledgebaseislost,buildingupaworkforcethatiscapableofreversingchangestokeysocietalsystemswillbeacostlyactivity.

ACoalition-BasedModelofCollectiveAgency

Intheprevioussection,webrieflyreviewedtheliteraturesurroundingAIandhumanagencytoidentifycriticalconceptsaboutagencyloss.Inthissection,weborrowfromthesocialchoiceliteraturetoproposeamodel-theoreticdefinitionofcollectiveagencythatisexpressiveenoughtorepresenttheseconcepts.Thesocialchoiceliteratureisvastandinterdisciplinary,so,tomatchthescopeofthisshortreport,wepresentonlyalimitedexpositionofthefieldasitintersectswithvotingsystems(seeSen’s[1977;1999]reviewsonsocialchoiceforamorecomprehensiveoverviewofsocialchoicetheory).

Weconsideragroupofindividualswhoareparticipatinginadecisionmakingprocess.Thisprocess,iden-tifiedasavotingsystem,ischaracterizedbythefollowingthreeadditionalconcepts:

•alternatives,whichlistpossibleoutcomesofthedecision

•preferences,whichindicateindividualandgrouppreferencesbetweenalternatives

•constraints,whichdescribeconventionsandrulesthatpreferencesshouldfollow.

Followingstandardconventions(CampbellandKelly,2000),werepresentthesefourelementsofavotingsystemusingthefollowingmathematicalobjects,withtheintentionofextendingtheformalismtonon–voting-basedprocessesaswell(seeAppendixAforamoreaxiomatictreatment):

•Individualsarelabeledfrom1toN.Whenwerefertotheseindividualsasagroup,especiallyinrefer-encetoadecisionmakingprocess,weusetheshorthandG.

•Alternativesarelabeledasuppercaseletters(X,Y,Z,etc.).

•Preferencesarerepresentedbythesymbol≤.X<YmeansthatYisstrictlypreferredtoX,andX~YmeansthatYandXareequivalentlypreferredasoutcomes.

•Constraintsarerulesonhowpreferencesshouldbehaveandrepresentthefactthatindividualsarepar-ticipatinginaprocesswithwell-definedproceduresandintendedfeatures.

Muchofthesocialchoiceliteratureisconcernedwiththerelationshipbetweenpreferencesandcon-straints.Wesupposeeachindividualhasapreference,labeled≤iforIndividuali.Letpdenoteanarbitrarylistofallindividualpreferences.Anaggregatepreferenceisamethodofusingtheinputpofindividualprefer-encestoproduceanaggregatedpreference≤p.3Implicitly,adecisionbetweenalternativesshouldfollowtheaggregatedpreference.Wealsoadoptthefollowingdefinitionfromthesocialchoiceliterature,whichidenti-fiesindividualswhohavedecisiveinfluenceoverhowaggregatedpreferencesaredetermined(Hansson,1976,asreferencedinCampbellandKelly,2000):

AsubgroupofindividualsHisadecisivecoalitionif,foralllistspofindividualpreferencesinscopeofthe

decisionmakingprocess,wheneverX<jYforallindividualsjinH,X≤pY.

Inotherwords,adecisivecoalition’spreferencesarealwaysrepresentedintheaggregatepreference.Someresultsintheliteraturerequireanarrowerdefinitionofdecisivecoalitionsthatforcesstrictpreferenceinthe

3

温馨提示

  • 1. 本站所有资源如无特殊说明,都需要本地电脑安装OFFICE2007和PDF阅读器。图纸软件为CAD,CAXA,PROE,UG,SolidWorks等.压缩文件请下载最新的WinRAR软件解压。
  • 2. 本站的文档不包含任何第三方提供的附件图纸等,如果需要附件,请联系上传者。文件的所有权益归上传用户所有。
  • 3. 本站RAR压缩包中若带图纸,网页内容里面会有图纸预览,若没有图纸预览就没有图纸。
  • 4. 未经权益所有人同意不得将文件中的内容挪作商业或盈利用途。
  • 5. 人人文库网仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对用户上传分享的文档内容本身不做任何修改或编辑,并不能对任何下载内容负责。
  • 6. 下载文件中如有侵权或不适当内容,请与我们联系,我们立即纠正。
  • 7. 本站不保证下载资源的准确性、安全性和完整性, 同时也不承担用户因使用这些下载资源对自己和他人造成任何形式的伤害或损失。

最新文档

评论

0/150

提交评论