Introduced:
Apr 3, 2025
Policy Area:
Science, Technology, Communications
Congress.gov:
Bill Statistics
3
Actions
17
Cosponsors
0
Summaries
1
Subjects
1
Text Versions
Yes
Full Text
AI Summary
AI Summary
No AI Summary Available
Click the button above to generate an AI-powered summary of this bill using Claude.
The summary will analyze the bill's key provisions, impact, and implementation details.
Error generating summary
Latest Action
Apr 3, 2025
Referred to the House Committee on Energy and Commerce.
Actions (3)
Referred to the House Committee on Energy and Commerce.
Type: IntroReferral
| Source: House floor actions
| Code: H11100
Apr 3, 2025
Introduced in House
Type: IntroReferral
| Source: Library of Congress
| Code: Intro-H
Apr 3, 2025
Introduced in House
Type: IntroReferral
| Source: Library of Congress
| Code: 1000
Apr 3, 2025
Subjects (1)
Science, Technology, Communications
(Policy Area)
Cosponsors (16 of 17)
(R-FL)
Oct 24, 2025
Oct 24, 2025
(R-NJ)
Oct 24, 2025
Oct 24, 2025
(D-NJ)
Oct 8, 2025
Oct 8, 2025
(R-IN)
Oct 8, 2025
Oct 8, 2025
(D-IL)
Oct 8, 2025
Oct 8, 2025
(D-MA)
Oct 8, 2025
Oct 8, 2025
(D-FL)
Oct 8, 2025
Oct 8, 2025
(D-CA)
Oct 8, 2025
Oct 8, 2025
(R-NC)
Oct 8, 2025
Oct 8, 2025
(R-NY)
Oct 8, 2025
Oct 8, 2025
(R-VA)
Oct 8, 2025
Oct 8, 2025
(R-GA)
Apr 3, 2025
Apr 3, 2025
(R-PA)
Apr 3, 2025
Apr 3, 2025
(R-IA)
Apr 3, 2025
Apr 3, 2025
(D-NY)
Apr 3, 2025
Apr 3, 2025
(D-WA)
Apr 3, 2025
Apr 3, 2025
Showing latest 16 cosponsors
Full Bill Text
Length: 27,926 characters
Version: Introduced in House
Version Date: Apr 3, 2025
Last Updated: Nov 14, 2025 6:05 AM
[Congressional Bills 119th Congress]
[From the U.S. Government Publishing Office]
[H.R. 2657 Introduced in House
(IH) ]
<DOC>
119th CONGRESS
1st Session
H. R. 2657
To require large social media platform providers to create, maintain,
and make available to third-party safety software providers a set of
real-time application programming interfaces, through which a child or
a parent or legal guardian of a child may delegate permission to a
third-party safety software provider to manage the online interactions,
content, and account settings of such child on the large social media
platform on the same terms as such child, and for other purposes.
_______________________________________________________________________
IN THE HOUSE OF REPRESENTATIVES
April 3, 2025
Ms. Wasserman Schultz (for herself, Mr. Carter of Georgia, Ms. Schrier,
Mrs. Miller-Meeks, Mr. Suozzi, and Mr. Fitzpatrick) introduced the
following bill; which was referred to the Committee on Energy and
Commerce
_______________________________________________________________________
A BILL
To require large social media platform providers to create, maintain,
and make available to third-party safety software providers a set of
real-time application programming interfaces, through which a child or
a parent or legal guardian of a child may delegate permission to a
third-party safety software provider to manage the online interactions,
content, and account settings of such child on the large social media
platform on the same terms as such child, and for other purposes.
Be it enacted by the Senate and House of Representatives of the
United States of America in Congress assembled,
[From the U.S. Government Publishing Office]
[H.R. 2657 Introduced in House
(IH) ]
<DOC>
119th CONGRESS
1st Session
H. R. 2657
To require large social media platform providers to create, maintain,
and make available to third-party safety software providers a set of
real-time application programming interfaces, through which a child or
a parent or legal guardian of a child may delegate permission to a
third-party safety software provider to manage the online interactions,
content, and account settings of such child on the large social media
platform on the same terms as such child, and for other purposes.
_______________________________________________________________________
IN THE HOUSE OF REPRESENTATIVES
April 3, 2025
Ms. Wasserman Schultz (for herself, Mr. Carter of Georgia, Ms. Schrier,
Mrs. Miller-Meeks, Mr. Suozzi, and Mr. Fitzpatrick) introduced the
following bill; which was referred to the Committee on Energy and
Commerce
_______________________________________________________________________
A BILL
To require large social media platform providers to create, maintain,
and make available to third-party safety software providers a set of
real-time application programming interfaces, through which a child or
a parent or legal guardian of a child may delegate permission to a
third-party safety software provider to manage the online interactions,
content, and account settings of such child on the large social media
platform on the same terms as such child, and for other purposes.
Be it enacted by the Senate and House of Representatives of the
United States of America in Congress assembled,
SECTION 1.
This Act may be cited as the ``Sammy's Law''.
SEC. 2.
It is the sense of Congress that--
(1) parents and legal guardians should be empowered to use
the services of third-party safety software providers to
protect the children of such parents and legal guardians from
certain harms on large social media platforms; and
(2) dangers like cyberbullying, human trafficking, illegal
drug distribution, sexual harassment, and violence perpetrated,
facilitated, or exacerbated through the use of certain large
social media platforms have harmed children on such platforms.
SEC. 3.
In this Act:
(1) Child.--The term ``child'' means any individual under
the age of 17 years who has registered an account with a large
social media platform.
(2) Commerce.--The term ``commerce'' has the meaning given
such term in
section 4 of the Federal Trade Commission Act (15
U.
U.S.C. 44).
(3) Commission.--The term ``Commission'' means the Federal
Trade Commission.
(4) Large social media platform.--The term ``large social
media platform''--
(A) means a service--
(i) provided through an internet website or
a mobile application (or both);
(ii) the terms of service of which do not
prohibit the use of the service by a child;
(iii) with any feature or features that
enable a child to share images, text, or video
through the internet with other users of the
service whom such child has met, identified, or
become aware of solely through the use of the
service; and
(iv) that has more than 100,000,000 monthly
global active users or generates more than
$1,000,000,000 in gross revenue per year,
adjusted yearly for inflation; and
(B) does not include--
(i) a service that primarily serves--
(I) to facilitate--
(aa) the sale or provision
of professional services; or
(bb) the sale of commercial
products; or
(II) to provide news or
information, where the service does not
offer the ability for content to be
sent by a user directly to a child; or
(ii) a service that--
(I) has a feature that enables a
user who communicates directly with a
child through a message (including a
text, audio, or video message) not
otherwise available to other users of
the service to add other users to that
message that such child may not have
otherwise met, identified, or become
aware of solely through the use of the
service; and
(II) does not have any feature or
features described in subparagraph
(A)
(iii) .
(5) Large social media platform provider.--The term ``large
social media platform provider'' means any person who, for
commercial purposes in or affecting commerce, provides,
manages, operates, or controls a large social media platform.
(6) State.--The term ``State'' means each State of the
United States, the District of Columbia, each commonwealth,
territory, or possession of the United States, and each
federally recognized Indian Tribe.
(7) Third-party safety software provider.--The term
``third-party safety software provider'' means any person who,
for commercial purposes in or affecting commerce, is authorized
by a child (if the child is 13 years of age or older) or a
parent or legal guardian of a child to interact with a large
social media platform to manage the online interactions,
content, or account settings of such child for the sole purpose
of protecting such child from harm, including physical or
emotional harm.
(8) User data.--The term ``user data'' means any
information needed to have a profile on a large social media
platform or content on a large social media platform, including
images, video, audio, or text, that is created by or sent to a
child on or through the account of such child with such
platform, but only--
(A) if the information or content is created by or
sent to such child while a delegation under
(3) Commission.--The term ``Commission'' means the Federal
Trade Commission.
(4) Large social media platform.--The term ``large social
media platform''--
(A) means a service--
(i) provided through an internet website or
a mobile application (or both);
(ii) the terms of service of which do not
prohibit the use of the service by a child;
(iii) with any feature or features that
enable a child to share images, text, or video
through the internet with other users of the
service whom such child has met, identified, or
become aware of solely through the use of the
service; and
(iv) that has more than 100,000,000 monthly
global active users or generates more than
$1,000,000,000 in gross revenue per year,
adjusted yearly for inflation; and
(B) does not include--
(i) a service that primarily serves--
(I) to facilitate--
(aa) the sale or provision
of professional services; or
(bb) the sale of commercial
products; or
(II) to provide news or
information, where the service does not
offer the ability for content to be
sent by a user directly to a child; or
(ii) a service that--
(I) has a feature that enables a
user who communicates directly with a
child through a message (including a
text, audio, or video message) not
otherwise available to other users of
the service to add other users to that
message that such child may not have
otherwise met, identified, or become
aware of solely through the use of the
service; and
(II) does not have any feature or
features described in subparagraph
(A)
(iii) .
(5) Large social media platform provider.--The term ``large
social media platform provider'' means any person who, for
commercial purposes in or affecting commerce, provides,
manages, operates, or controls a large social media platform.
(6) State.--The term ``State'' means each State of the
United States, the District of Columbia, each commonwealth,
territory, or possession of the United States, and each
federally recognized Indian Tribe.
(7) Third-party safety software provider.--The term
``third-party safety software provider'' means any person who,
for commercial purposes in or affecting commerce, is authorized
by a child (if the child is 13 years of age or older) or a
parent or legal guardian of a child to interact with a large
social media platform to manage the online interactions,
content, or account settings of such child for the sole purpose
of protecting such child from harm, including physical or
emotional harm.
(8) User data.--The term ``user data'' means any
information needed to have a profile on a large social media
platform or content on a large social media platform, including
images, video, audio, or text, that is created by or sent to a
child on or through the account of such child with such
platform, but only--
(A) if the information or content is created by or
sent to such child while a delegation under
section 4
(a) is in effect with respect to the account; and
(B) during a 30-day period beginning on the date on
which the information or content is created by or sent
to such child.
(a) is in effect with respect to the account; and
(B) during a 30-day period beginning on the date on
which the information or content is created by or sent
to such child.
SEC. 4.
(a) Duty of Large Social Media Platform Providers.--
(1) In general.--Not later than 30 days after the effective
date of this Act (in the case of a service that is a large
social media platform on such effective date) or not later than
30 days after a service becomes a large social media platform
(in the case of a service that becomes a large social media
platform after such effective date), the large social media
platform provider shall create, maintain, and make available to
any third-party safety software provider registered with the
Commission under subsection
(b)
(1) a set of third-party-
accessible real-time application programming interfaces,
including any information necessary to use such interfaces, by
which a child (if the child is 13 years of age or older) or a
parent or legal guardian of a child may delegate permission to
the third-party safety software provider to--
(A) manage the online interactions, content, and
account settings of such child on the large social
media platform on the same terms as such child; and
(B) initiate secure transfers of user data from the
large social media platform in a commonly-used and
machine-readable format to the third-party safety
software provider, where the frequency of such
transfers may not be limited by the large social media
platform provider to less than once per hour.
(2) Revocation.--Once a child or a parent or legal guardian
of a child makes a delegation under paragraph
(1) , the large
social media platform provider shall make the application
programming interfaces and information described in such
paragraph available to the third-party safety software provider
on an ongoing basis until--
(A) the child (if the child made the delegation) or
the parent or legal guardian of such child revokes the
delegation;
(B) the child or a parent or legal guardian of such
child revokes or disables the registration of the
account of such child with the large social media
platform;
(C) the third-party safety software provider
rejects the delegation; or
(D) one or more of the affirmations made by the
third-party safety software provider under subsection
(b)
(1)
(A) is no longer true.
(3) Secure transfer of user data.--A large social media
platform provider shall establish and implement reasonable
policies, practices, and procedures regarding the secure
transfer of user data pursuant to a delegation under paragraph
(1) from the large social media platform to a third-party
safety software provider in order to mitigate any risks related
to user data.
(4) Disclosure.--In the case of a delegation made by a
child or a parent or legal guardian of a child under paragraph
(1) with respect to the account of such child with a large
social media platform, the large social media platform provider
shall--
(A) disclose to such child and (if the parent or
legal guardian made the delegation) the parent or legal
guardian the fact that the delegation has been made;
(B) provide to such child and (if such parent or
legal guardian made the delegation) such parent or
legal guardian a summary of the user data that is
transferred to the third-party safety software
provider; and
(C) update the summary provided under subparagraph
(B) as necessary to reflect any change to the user data
that is transferred to the third-party safety software
provider.
(5) Limitation.--Any management by a third-party safety
software provider of online interactions, content, and account
settings of a child under this subsection shall be limited to
such management that protects such child from harm, including
the optimization of the privacy settings of the account, stated
user age, and marketing settings of the child.
(b) Third-Party Safety Software Providers.--
(1) Registration with commission.--A third-party safety
software provider shall register with the Commission as a
condition of accessing an application programming interface and
any information under subsection
(a) . As a condition of such
registration, the third-party safety software provider shall--
(A) satisfactorily demonstrate to the Commission
that the third-party safety software provider--
(i) is a company based in the United
States;
(ii) is not a subsidiary of any foreign-
owned company or otherwise controlled by a
foreign person or persons;
(iii) will solely use any user data
obtained under subsection
(a) for the purpose
of protecting a child from harm in accordance
with any applicable terms of service and the
provisions of this Act;
(iv) will only disclose user data obtained
under subsection
(a) as permitted by subsection
(f) ;
(v) will process and maintain all user data
obtained under subsection
(a) and copies of any
communications generated in relation thereto
exclusively on hardware and devices located
within the territorial boundaries of the United
States;
(vi)
(I) will delete any user data obtained
under this section as soon as possible but not
later than 14 days after receiving such data
from the large social media platform, not
including any data the third-party safety
software provider discloses under subsection
(f) ;
(II) for any data disclosed under
subsection
(f)
(1)
(C) , will maintain such data
until the child or a parent or legal guardian
of the child who made a delegation under
subsection
(a) and whose data is at issue
requests that the third-party safety software
provider delete such data; and
(III) in the event that the child or a
parent or legal guardian of the child who made
a delegation under subsection
(a) cancels their
account with the third-party safety software
provider, will delete all applicable user data
no later than 30 days after the request for
account cancellation has been made; and
(vii) will disclose, in an easy-to-
understand, human-readable format, to each
child with respect to whose account with a
large social media platform the service of the
third-party safety software provider is
operating and (if a parent or legal guardian of
the child made the delegation under subsection
(a) with respect to the account) to the parent
or legal guardian, sufficient information
detailing the operation of the service and what
information the third-party safety software
provider is collecting to enable such child and
(if applicable) such parent or legal guardian
to make informed decisions regarding the use of
the service; and
(B) as part of the registration process, undergo a
security review in such form as the Commission may
proscribe but which may include requiring that a
qualified independent auditing firm conduct such a
review to independently verify and confirm via a
written report (which shall be exempt from disclosure
under
section 552
(b)
(3) of title 5, United States Code)
that the third-party safety software provider--
(i) is in compliance, or has the ability to
comply, with the requirements of subparagraph
(A) ;
(ii) is able to provide services in
accordance with any applicable terms of service
and any relevant disclosures made to any
consumer, including whether such terms and
disclosures are clear and conspicuous and are
written in plain and easy-to-understand
English;
(iii) has taken appropriate steps to assess
potential risks and to protect the
confidentiality, integrity, and security of any
user data, including a determination of the
adequacy of business and technology-related
controls, policies, procedures, and other
safeguards employed by the third-party safety
software provider based on guidance issued by
the Commission and other industry standards and
best practices; and
(iv) assesses compliance with applicable
Federal law, including the requirements of this
Act.
(b)
(3) of title 5, United States Code)
that the third-party safety software provider--
(i) is in compliance, or has the ability to
comply, with the requirements of subparagraph
(A) ;
(ii) is able to provide services in
accordance with any applicable terms of service
and any relevant disclosures made to any
consumer, including whether such terms and
disclosures are clear and conspicuous and are
written in plain and easy-to-understand
English;
(iii) has taken appropriate steps to assess
potential risks and to protect the
confidentiality, integrity, and security of any
user data, including a determination of the
adequacy of business and technology-related
controls, policies, procedures, and other
safeguards employed by the third-party safety
software provider based on guidance issued by
the Commission and other industry standards and
best practices; and
(iv) assesses compliance with applicable
Federal law, including the requirements of this
Act.
(2) Annual audit.--
(A) Audit process; audit report.--For each year or
partial year during which a third-party safety software
provider is registered with the Commission under
paragraph
(1) , the third-party safety software provider
shall retain the services of a qualified independent
auditing firm to complete an annual audit and write an
audit report (which shall be exempt from disclosure
under
section 552
(b)
(3) of title 5, United States
Code), and such audit report shall--
(i) include a review and assessment of the
third-party safety software provider's initial
security review and any subsequent written
reports, including whether the third-party
safety software provider has remained in
compliance with the requirements described in
paragraph
(1)
(B) ; and
(ii) identify whether the third-party
safety software provider has made any material
changes in how the third-party safety software
provider provides services, and in the event of
any such material changes, provide an
explanation as to how such changes have
impacted users.
(b)
(3) of title 5, United States
Code), and such audit report shall--
(i) include a review and assessment of the
third-party safety software provider's initial
security review and any subsequent written
reports, including whether the third-party
safety software provider has remained in
compliance with the requirements described in
paragraph
(1)
(B) ; and
(ii) identify whether the third-party
safety software provider has made any material
changes in how the third-party safety software
provider provides services, and in the event of
any such material changes, provide an
explanation as to how such changes have
impacted users.
(B) Submission to commission.--Not later than 30
days after the date on which an audit report is written
under subparagraph
(A) , a third-party safety software
provider shall submit to the Commission--
(i) a full copy of such audit report; and
(ii) a summary of such audit report that
may contain redactions to protect the
proprietary information and trade secrets of
the third-party safety software provider.
(C) Audit review by commission.--The Commission
shall--
(i) review each audit report submitted by a
third-party safety software provider under
subparagraph
(B)
(i) to verify compliance;
(ii) make a copy of the summary of such
audit report submitted by a third-party safety
software provider under subparagraph
(B)
(ii) available to the public; and
(iii) in the event an audit required under
subparagraph
(A) detects an unusual finding,
direct a third-party safety software provider
to promptly investigate and resolve the matter.
(3) Additional authority of commission.--In addition to the
jurisdiction, powers, and duties of the Commission otherwise
provided under this Act and any other provision of law, the
Commission may take an adverse action against a third-party
safety software provider, including by--
(A) denying an initial registration for the third-
party safety software provider under paragraph
(1) ;
(B) permanently de-registering the third-party
safety software provider; and
(C) suspending the registration of the third-party
safety software provider due to an audit finding of a
material risk to the security of the data or safety of
the public, including for--
(i) willful misconduct or gross negligence
by the third-party safety software provider;
(ii) a material misrepresentation made by a
third-party safety software provider to the
Commission or to any consumer;
(iii) failure by the third-party safety
software provider to comply with any
requirements of this Act or failure to operate
in accordance with the affirmations,
assertions, representations, or terms of any
security review, audit, terms of services, or
consumer disclosures;
(iv) failure by the third-party safety
software provider to respond to an unusual
finding in an annual audit completed under
paragraph
(2)
(A) ; and
(v) failure by the third-party safety
software provider to adhere to or implement
guidance issued by the Commission.
(4) Rights of third-party safety software providers.--
(A) In general.--In the event the Commission takes
an adverse action against a third-party safety software
provider under paragraph
(3) , the Commission shall give
the third-party safety software provider--
(i) the opportunity to appeal the findings
of the auditor or such action of the
Commission; and
(ii) the opportunity to remediate any
deficiencies, except in the case of a finding
of--
(I) willful misconduct;
(II) gross negligence; or
(III) a demonstrated history of
multiple failures in relation to the
types of material risk described in
paragraph
(3)
(C)
(ii) through
(3)
(C)
(v) .
(B) Exception.--The rights described in
subparagraph
(A) shall not prevent the Commission from
suspending the registration of a third-party safety
software provider to protect the public from ongoing
material risk for the period during which the third-
party safety software provider is in the process of
exercising the rights described in paragraph
(4) .
(c) Authentication.--Not later than 180 days after the date of the
enactment of this Act, the Commission shall issue guidance to
facilitate the ability of a third-party safety software provider to
obtain user data or access under subsection
(a) in a manner that
ensures that a request for user data or access on behalf of a child is
a verifiable request.
(d) Guidance and Consumer Education.--The Commission shall--
(1) not later than 180 days after the date of the enactment
of this Act, issue guidance for large social media platform
providers and third-party safety software providers regarding
the maintenance of reasonable safety standards to protect user
data; and
(2) educate consumers regarding the rights of consumers
under this Act.
(e) Indemnification.--In any civil action in Federal or State court
(other than an action brought by the Commission), a large social media
platform provider may not be held liable for damages arising out of the
transfer of user data to a third-party safety software provider under
subsection
(a) , if the large social media platform provider has in good
faith complied with the requirements of this Act and the guidance
issued by the Commission under this Act.
(f) User Data Disclosure.--
(1) Permitted disclosures.--A third-party safety software
provider may not disclose any user data obtained under
subsection
(a) to any other person except--
(A) pursuant to a lawful request from a government
body, including for law enforcement purposes or for
judicial or administrative proceedings by means of a
court order or a court-ordered warrant, a subpoena or
summons issued by a judicial officer, or a grand jury
subpoena;
(B) to the extent that such disclosure is required
by law and such disclosure complies with and is limited
to the relevant requirements of such law;
(C) to the child or a parent or legal guardian of
the child who made a delegation under such subsection
and whose data is at issue, with such third-party
safety software provider making a good faith effort to
ensure that such disclosure includes only the user data
necessary for a reasonable parent or caregiver to
understand that such child is experiencing (or is at
foreseeable risk to experience) the following harms--
(i) suicide;
(ii) anxiety;
(iii) depression;
(iv) eating disorders;
(v) violence, including being the victim of
or planning to commit or facilitate assault;
(vi) substance abuse;
(vii) fraud;
(viii) severe forms of trafficking in
persons (as defined in
section 103 of the
Trafficking Victims Protection Act of 2000 (22
U.
Trafficking Victims Protection Act of 2000 (22
U.S.C. 7102));
(ix) sexual abuse;
(x) physical injury;
(xi) harassment;
(xii) sexually explicit conduct or child
pornography (as defined in
U.S.C. 7102));
(ix) sexual abuse;
(x) physical injury;
(xi) harassment;
(xii) sexually explicit conduct or child
pornography (as defined in
section 2256 of
title 18, United States Code);
(xiii) terrorism (as defined in
title 18, United States Code);
(xiii) terrorism (as defined in
(xiii) terrorism (as defined in
section 140
(d) of the Foreign Relations Authorization
Act, Fiscal Years 1988 and 1989 (22 U.
(d) of the Foreign Relations Authorization
Act, Fiscal Years 1988 and 1989 (22 U.S.C.
2656f
(d) )), including communications with or in
support of a foreign terrorist organization (as
designated by the Secretary of State under
Act, Fiscal Years 1988 and 1989 (22 U.S.C.
2656f
(d) )), including communications with or in
support of a foreign terrorist organization (as
designated by the Secretary of State under
section 219
(a) of the Immigration and
Nationality Act (8 U.
(a) of the Immigration and
Nationality Act (8 U.S.C. 1189
(a) ));
(xiv) academic dishonesty, including
cheating, plagiarism, and other forms of
academic dishonesty that are intended to gain
an unfair academic advantage; and
(xv) sharing personal information, limited
to--
(I) home address;
(II) phone number;
(III) social security number; and
(IV) personal banking information;
(D) in the case of a reasonably foreseeable serious
and imminent threat to the health or safety of any
individual, if the disclosure is made to a person or
persons reasonably able to prevent or lessen the
threat; or
(E) to a public health authority or other
appropriate government authority authorized by law to
receive reports of child abuse or neglect.
(2) Disclosure reporting.--A third-party safety software
provider that makes a disclosure permitted by paragraph
(1)
(A) ,
(1)
(B) ,
(1)
(D) , or
(1)
(E) shall promptly inform the child with
respect to whose account with a large social media platform the
delegation was made under subsection
(a) and (if a parent or
legal guardian of the child made the delegation) the parent or
legal guardian that such a disclosure has been or will be made,
except if--
(A) the third-party safety software provider, in
the exercise of professional judgment, believes
informing such child or parent or legal guardian would
place such child at risk of serious harm; or
(B) the third-party safety software provider is
prohibited by law (including a valid order by a court
or administrative body) from informing such child or
parent or legal guardian.
SEC. 5.
(a) Enforcement.--
(1) Unfair or deceptive acts or practices.--A violation of
this Act shall be treated as a violation of a rule defining an
unfair or deceptive act or practice prescribed under
section 18
(a)
(1)
(B) of the Federal Trade Commission Act (15 U.
(a)
(1)
(B) of the Federal Trade Commission Act (15 U.S.C.
57a
(a)
(1)
(B) ).
(2) Powers of commission.--
(A) In general.--The Commission shall enforce this
Act in the same manner, by the same means, and with the
same jurisdiction, powers, and duties as though all
applicable terms and provisions of the Federal Trade
Commission Act (15 U.S.C. 41 et seq.) were incorporated
into and made a part of this Act.
(B) Privileges and immunities.--Any person who
violates this Act shall be subject to the penalties and
entitled to the privileges and immunities provided in
the Federal Trade Commission Act (15 U.S.C. 41 et
seq.).
(3) Preservation of authority.--Nothing in this Act may be
construed to limit the authority of the Commission under any
other provision of law.
(b) FTC Guidance.--Not later than 180 days after the date of the
enactment of this Act, the Commission shall issue guidance to assist
large social media platform providers and third-party safety software
providers in complying with this Act.
(c) Compliance Assessment.--The Commission, on a biannual basis,
shall assess compliance by large social media platform providers and
third-party safety software providers with the provisions of this Act.
(d) Complaints.--The Commission shall establish procedures under
which a child, or the parent or legal guardian of such child, a large
social media platform provider, or a third-party safety software
provider may file a complaint alleging that a large social media
platform provider or a third-party safety software provider has
violated this Act.
SEC. 6.
(a) In General.--No State or political subdivision of a State may
maintain, enforce, prescribe, or continue in effect any law, rule,
regulation, requirement, standard, or other provision having the force
and effect of law of the State, or political subdivision of a State,
related to requiring large social media platform providers to create,
maintain, and make available to third-party safety software providers a
set of real-time application programming interfaces, through which a
child or a parent or legal guardian of a child may delegate permission
to a third-party safety software provider to manage the online
interactions, content, and account settings of such child on a large
social media platform on the same terms as such child.
(b) Rule of Construction.--This section may not be construed to--
(1) limit the enforcement of any consumer protection law of
a State or political subdivision of a State;
(2) preempt the applicability of State trespass, contract,
or tort law; or
(3) preempt the applicability of any State law to the
extent that the law relates to acts of fraud, unauthorized
access to personal information, or notification of unauthorized
access to personal information.
SEC. 7.
This Act shall take effect on the date on which the Commission
issues guidance under
section 5
(b) .
(b) .
<all>