联通手机营业厅自动做任务、签到、领流量、领积分等。

Overview

UnicomTask

简介

👯 😄 📫

联通手机营业厅自动完成每日任务,领流量、签到获取积分等,月底流量不发愁。

功能

  • 沃之树领流量、浇水(12M日流量)
  • 每日签到(1积分+翻倍4积分+第七天1G流量日包)
  • 天天抽奖,每天三次免费机会(随机奖励)
  • 游戏中心每日打卡(连续打卡,积分递增至最高7,第七天1G流量日包)
  • 游戏中心宝箱100M任务(100M日流量+随机奖励并翻倍)
  • 4G流量包看视频、下软件任务(90M+150M七日流量)
  • 每日领取100定向积分
  • 积分抽奖,每天最多抽30次(中奖几率渺茫)
  • 冬奥积分活动(第1和7天,可领取600定向积分,其余领取300定向积分,有效期至下月底)
  • 邮件推送运行结果

Github Actions 部署方法

1.fork本项目

项目地址:srcrs/UnicomTask

2.准备需要的参数

手机号、服务密码、appID。

其中appId的获取:

  • 安卓用户可在文件管理 --> Unicom/appid 文件中获取。

  • 苹果用户可抓取客户端登录接口获取。

https://m.client.10010.com/mobileService/login.htm

3.将必要参数填到Secrets

Secrets中的NameValue格式如下:

Name Value 说明
USERNAME_COVER 18566669999 手机号(必须)
PASSWORD_COVER 123456 服务密码(必须)
APPID_COVER xxxxxxxxx appId(必须)
EMAIL_COVER [email protected] 邮箱(可选)
LOTTERY_NUM 填写正整数 抽奖次数(可选)

4.开启Actions

默认Actions处于禁止状态,在Actions选项中开启Actions功能,把那个绿色的长按钮点一下。如果看到左侧工作流上有黄色!号,还需继续开启。

5.进行一次push操作

push操作会触发工作流运行。

删除掉README.md中的 😄 即可。完成后,每天早上7:30将自动完成每日任务。

同步上游代码

在最新的代码中,已经加上自动同步上游代码的action,将会定时在每周五16点执行,文件地址在.github/workflows/auto_merge.yml

同时您也可以安装pull应用,也可实现自动同步上游代码。

申明

本项目仅用于学习。

参考项目

mixool/HiCnUnicom,感谢该项目对于登录部分的思路

You might also like...
Releases(UnicomTask-v1.7)
Scrapes all articles and their headlines from theonion.com

The Onion Article Scraper Scrapes all articles and their headlines from the satirical news website https://www.theonion.com Also see Clickhole Article

0 Nov 17, 2021
The core packages of security analyzer web crawler

Security Analyzer 🐍 A large scale web crawler (considered also as vulnerability scanner tool) to take an overview about security of Moroccan sites Cu

Security Analyzer 10 Jul 03, 2022
A distributed crawler for weibo, building with celery and requests.

A distributed crawler for weibo, building with celery and requests.

SpiderClub 4.8k Jan 03, 2023
✂️🕷️ Spider-Cut is a Network Mapper Framework (NMAP Framework)

Spider-Cut is a Network Mapper Framework (NMAP Framework) Installation | Usage | Creators | Donate Installation # Kali Linux | WSL

XforWorks 3 Mar 07, 2022
Web Crawlers for Data Labelling of Malicious Domain Detection & IP Reputation Evaluation

Web Crawlers for Data Labelling of Malicious Domain Detection & IP Reputation Evaluation This repository provides two web crawlers to label domain nam

1 Nov 05, 2021
Amazon web scraping using Scrapy Framework

Amazon-web-scraping-using-Scrapy-Framework Scrapy Scrapy is an application framework for crawling web sites and extracting structured data which can b

Sejal Rajput 1 Jan 25, 2022
A leetcode scraper to compile all questions in leetcode free tier to text file. pdf also available.

A leetcode scraper to compile all questions in leetcode free tier to text file, pdf also available. if new questions get added, run again to get new questions.

3 Dec 07, 2021
An helper library to scrape data from Instagram effortlessly, using the Influencer Hunters APIs.

Instagram Scraper An utility library to scrape data from Instagram hassle-free Go to the website » View Demo · Report Bug · Request Feature About The

2 Jul 06, 2022
fork huanghyw/jd_seckill

Jd_Seckill 特别声明: 本仓库发布的jd_seckill项目中涉及的任何脚本,仅用于测试和学习研究,禁止用于商业用途,不能保证其合法性,准确性,完整性和有效性,请根据情况自行判断。 本项目内所有资源文件,禁止任何公众号、自媒体进行任何形式的转载、发布。

512 Jan 03, 2023
This is a webscraper for a specific website

This is a webscraper for a specific website. It is tuned to extract the headlines of that website. With some little adjustments the webscraper is able to extract any part of the website.

Rahul Siyanwal 1 Dec 13, 2021
Use Flask API to wrap Facebook data. Grab the wapper of Facebook public pages without an API key.

Facebook Scraper Use Flask API to wrap Facebook data. Grab the wapper of Facebook public pages without an API key. (Currently working 2021) Setup Befo

Encore Shao 2 Dec 27, 2021
Scrapy uses Request and Response objects for crawling web sites.

Requests and Responses¶ Scrapy uses Request and Response objects for crawling web sites. Typically, Request objects are generated in the spiders and p

Md Rashidul Islam 1 Nov 03, 2021
Proxy scraper. Format: IP | PORT | COUNTRY | TYPE

proxy scraper 🔎 Installation: git clone https://github.com/ebankoff/proxy_scraper Required pip libraries (pip install library name): lxml beautifulso

Eban'ko 19 Dec 07, 2022
Create crawler get some new products with maximum discount in banimode website

crawler-banimode create crawler and get some new products with maximum discount in banimode website. این پروژه کوچک جهت یادگیری و کار با ابزار سلنیوم

nourollah rezaei 2 Feb 17, 2022
This project was created using Python technology and flask tools to scrape a music site

python-scrapping This project was created using Python technology and flask tools to scrape a music site You need to install the following packages to

hosein moradi 1 Dec 07, 2021
自动完成每日体温上报(Github Actions)

体温上报助手 简介 每天 10:30 GMT+8 自动完成体温上报,如想修改定时运行的时间,可修改 .github/workflows/SduHealthReport.yml 中 schedule 属性。 如果当日有异常,请手动在小程序端/PC 端填写!

Teng Zhang 23 Sep 15, 2022
12306抢票脚本

12306抢票脚本

罐子里的茶 457 Jan 05, 2023
Poolbooru gelscraper - a simple python script for scraping images off gelbooru pools.

poolbooru_gelscraper a simple python script for scraping images off gelbooru pools. modules required:requests_html, and os by default saves files with

savantshuia 1 Jan 02, 2022
Web scraper build using python.

Web Scraper This project is made in pyhthon. It took some info. from website list then add them into data.json file. The dependencies used are: reques

Shashwat Harsh 2 Jul 22, 2022
🕷 Phone Crawler with multi-thread functionality

Phone Crawler: Phone Crawler with multi-thread functionality Disclaimer: I'm not responsible for any illegal/misuse actions, this program was made for

Kmuv1t 3 Feb 10, 2022