Compare commits

..

40 Commits

Author SHA1 Message Date
fdb8db4782 Merge pull request 'v4.1.0' (#189) from dev into main
Some checks failed
Analysis / SonarCloud (push) Successful in 46s
Tests / Build and Test (3.11) (push) Successful in 1m9s
Tests / Build and Test (3.12) (push) Successful in 1m7s
Tests / Build and Test (3.13) (push) Successful in 1m4s
Upload Python Package / release-build (release) Successful in 18s
Upload Python Package / gitea-publish (release) Failing after 20s
Upload Python Package / pypi-publish (release) Failing after 11s
Reviewed-on: #189
2025-02-16 18:37:38 +02:00
12f7cb6365 Added some basic tests for Cache
All checks were successful
Analysis / SonarCloud (push) Successful in 42s
Analysis / SonarCloud (pull_request) Successful in 49s
Tests / Build and Test (3.11) (pull_request) Successful in 1m18s
Tests / Build and Test (3.12) (pull_request) Successful in 1m9s
Tests / Build and Test (3.13) (pull_request) Successful in 1m6s
2025-02-16 17:16:05 +01:00
76ee24cd9e Added cache support
All checks were successful
Analysis / SonarCloud (push) Successful in 50s
2025-02-16 17:03:46 +01:00
6d56d9d0f9 Hopefully fixed a circular import
All checks were successful
Analysis / SonarCloud (push) Successful in 47s
2025-02-16 16:58:56 +01:00
554b522400 Added experimental cache support
All checks were successful
Analysis / SonarCloud (push) Successful in 51s
2025-02-16 16:45:22 +01:00
e9abed27f8 Removed caching action
All checks were successful
Analysis / SonarCloud (push) Successful in 52s
2025-02-09 18:55:26 +01:00
845a69491d Silly attempt to fix token issues
Some checks failed
Analysis / SonarCloud (push) Failing after 4s
2025-02-09 18:51:48 +01:00
df2b5efd88 Merge pull request 'Closes #187 and improves documentation' (#188) from bugfix/187 into dev
Some checks failed
Analysis / SonarCloud (push) Failing after 4s
Reviewed-on: #188
2025-02-09 19:41:30 +02:00
6b2be48052 Closes #187 and improves documentation
Some checks failed
Analysis / SonarCloud (pull_request) Failing after 5s
Tests / Build and Test (3.11) (pull_request) Successful in 1m16s
Tests / Build and Test (3.12) (pull_request) Successful in 1m9s
Tests / Build and Test (3.13) (pull_request) Successful in 1m7s
2025-02-09 18:40:30 +01:00
ad70648ea2 Update dependency mypy to v1.15.0
Some checks failed
Analysis / SonarCloud (pull_request) Failing after 4s
Tests / Build and Test (3.11) (pull_request) Successful in 1m6s
Tests / Build and Test (3.12) (pull_request) Successful in 1m4s
Tests / Build and Test (3.13) (pull_request) Successful in 1m0s
Analysis / SonarCloud (push) Failing after 4s
2025-02-05 06:31:00 +02:00
09b4d512a6 Merge pull request 'Update dependency pytest-asyncio to v0.25.3' (#184) from renovate/pytest-asyncio-0.x into dev
Some checks failed
Analysis / SonarCloud (push) Failing after 3s
Reviewed-on: #184
2025-01-29 10:48:23 +02:00
1473d34ca1 Merge pull request 'Update dependency black to v25' (#185) from renovate/black-25.x into dev
Some checks failed
Analysis / SonarCloud (push) Failing after 4s
Reviewed-on: #185
2025-01-29 10:43:45 +02:00
5fc8ae6a6e Update dependency black to v25
Some checks failed
Analysis / SonarCloud (pull_request) Failing after 4s
Tests / Build and Test (3.11) (pull_request) Successful in 1m9s
Tests / Build and Test (3.12) (pull_request) Successful in 1m5s
Tests / Build and Test (3.13) (pull_request) Successful in 1m4s
2025-01-29 06:29:48 +02:00
8562d7e84c Update dependency pytest-asyncio to v0.25.3
Some checks failed
Analysis / SonarCloud (pull_request) Failing after 4s
Tests / Build and Test (3.11) (pull_request) Successful in 1m19s
Tests / Build and Test (3.12) (pull_request) Successful in 1m20s
Tests / Build and Test (3.13) (pull_request) Successful in 1m18s
2025-01-28 21:04:58 +02:00
258b46d829 Update dependency pylint to v3.3.4
Some checks failed
Analysis / SonarCloud (pull_request) Failing after 4s
Tests / Build and Test (3.11) (pull_request) Successful in 1m12s
Tests / Build and Test (3.12) (pull_request) Successful in 1m7s
Tests / Build and Test (3.13) (pull_request) Successful in 1m6s
Analysis / SonarCloud (push) Failing after 4s
2025-01-28 15:50:37 +02:00
efec002667 Merge pull request 'Update dependency twine to v6.1.0' (#181) from renovate/twine-6.x into dev
Some checks failed
Analysis / SonarCloud (push) Failing after 3s
Reviewed-on: #181
2025-01-24 16:38:45 +02:00
475eaf9ff3 Update dependency twine to v6.1.0
Some checks failed
Analysis / SonarCloud (pull_request) Failing after 3s
Tests / Build and Test (3.11) (pull_request) Successful in 1m31s
Tests / Build and Test (3.12) (pull_request) Successful in 1m23s
Tests / Build and Test (3.13) (pull_request) Successful in 1m14s
2025-01-21 21:37:09 +02:00
0fcd9f2041 Merge pull request 'Update dependency tox to v4.24.0' (#179) from renovate/tox-4.x into dev
Some checks failed
Analysis / SonarCloud (push) Failing after 4s
Reviewed-on: #179
2025-01-21 21:12:20 +02:00
44d07dc56a Update dependency tox to v4.24.0
Some checks failed
Tests / Build and Test (3.11) (pull_request) Successful in 1m17s
Tests / Build and Test (3.13) (pull_request) Successful in 1m6s
Tests / Build and Test (3.12) (pull_request) Successful in 1m9s
Analysis / SonarCloud (pull_request) Failing after 4s
2025-01-21 20:32:47 +02:00
c5e83c17d3 Disabled pip cache for publish because dependencies are inline
All checks were successful
Analysis / SonarCloud (push) Successful in 42s
2025-01-10 11:23:35 +02:00
129cbd923b Updated cache path for tests
All checks were successful
Analysis / SonarCloud (push) Successful in 41s
2025-01-10 11:22:33 +02:00
1ca126829b Hopefully fixed caching
All checks were successful
Analysis / SonarCloud (push) Successful in 49s
2025-01-10 00:09:05 +01:00
974aebfd1a Bruh, works exactly as bad. I give up... Let's cache this shit.
Some checks failed
Analysis / SonarCloud (push) Has been cancelled
2025-01-09 23:08:09 +01:00
ed7fa50dbd SonarQube works like shit, switching back to the old SonarCloud action
All checks were successful
Analysis / SonarCloud (push) Successful in 9m49s
2025-01-09 22:54:21 +01:00
82542de0bb SonarQube doesn't seem to work, one more try with latest
Some checks failed
Analysis / SonarCloud (push) Has been cancelled
2025-01-09 22:40:32 +01:00
9021eac87b SonarCloud action is deprecated, replacing with sonarqube one and returning to ubuntu-24.04
Some checks failed
Analysis / SonarCloud (push) Has been cancelled
2025-01-09 22:35:53 +01:00
651022ab6e SonarCloud doesn't like 22.04 either, trying ubuntu-latest instead
Some checks failed
Analysis / SonarCloud (push) Has been cancelled
2025-01-09 22:33:38 +01:00
f8c6b782a1 SonarCloud doesn't like 24.04, trying 22.04 instead
Some checks failed
Analysis / SonarCloud (push) Has been cancelled
2025-01-09 22:32:24 +01:00
a1d0b98858 Replaced ubuntu-latest with ubuntu-24.04
Some checks failed
Analysis / SonarCloud (push) Has been cancelled
2025-01-09 22:18:57 +01:00
fec40b1c44 Update dependency pytest-asyncio to v0.25.2
All checks were successful
Analysis / SonarCloud (pull_request) Successful in 34s
Tests / Build and Test (3.11) (pull_request) Successful in 1m15s
Tests / Build and Test (3.12) (pull_request) Successful in 1m21s
Tests / Build and Test (3.13) (pull_request) Successful in 1m20s
Analysis / SonarCloud (push) Successful in 46s
2025-01-08 08:41:33 +02:00
508c48d22b Merge pull request 'v4.0.2' (#172) from dev into main
All checks were successful
Analysis / SonarCloud (push) Successful in 46s
Tests / Build and Test (3.11) (push) Successful in 1m16s
Tests / Build and Test (3.12) (push) Successful in 1m24s
Tests / Build and Test (3.13) (push) Successful in 1m24s
Upload Python Package / release-build (release) Successful in 18s
Upload Python Package / gitea-publish (release) Successful in 9s
Upload Python Package / pypi-publish (release) Successful in 15s
Reviewed-on: #172
2025-01-02 15:04:26 +02:00
kku
e9b9fc6ca1 TEST: Publishing Action
All checks were successful
Analysis / SonarCloud (push) Successful in 46s
Analysis / SonarCloud (pull_request) Successful in 41s
Tests / Build and Test (3.11) (pull_request) Successful in 1m19s
Tests / Build and Test (3.12) (pull_request) Successful in 1m24s
Tests / Build and Test (3.13) (pull_request) Successful in 1m24s
2025-01-02 13:58:23 +01:00
1da367ccb1 Update dependency pytest-asyncio to v0.25.1
All checks were successful
Analysis / SonarCloud (push) Successful in 41s
Analysis / SonarCloud (pull_request) Successful in 40s
Tests / Build and Test (3.11) (pull_request) Successful in 1m15s
Tests / Build and Test (3.12) (pull_request) Successful in 1m22s
Tests / Build and Test (3.13) (pull_request) Successful in 1m22s
2025-01-02 08:03:07 +02:00
kku
d5e390fe66 Optimized json_load usage, imports and typing
All checks were successful
Analysis / SonarCloud (push) Successful in 48s
Analysis / SonarCloud (pull_request) Successful in 39s
Tests / Build and Test (3.11) (pull_request) Successful in 1m22s
Tests / Build and Test (3.12) (pull_request) Successful in 1m27s
Tests / Build and Test (3.13) (pull_request) Successful in 1m29s
2025-01-01 22:34:38 +01:00
kku
ae54bd5cce Bump version to 4.0.2
All checks were successful
Analysis / SonarCloud (push) Successful in 42s
Analysis / SonarCloud (pull_request) Successful in 39s
Tests / Build and Test (3.11) (pull_request) Successful in 1m19s
Tests / Build and Test (3.12) (pull_request) Successful in 1m22s
Tests / Build and Test (3.13) (pull_request) Successful in 1m23s
2024-12-31 11:16:16 +01:00
kku
9ce251d733 Added a quick README for examples (belongs to #60)
All checks were successful
Analysis / SonarCloud (push) Successful in 41s
2024-12-31 11:10:06 +01:00
kku
5dd873d683 Closes #61
All checks were successful
Analysis / SonarCloud (push) Successful in 1m1s
2024-12-31 11:07:24 +01:00
b47bcbe513 Update dependency mypy to v1.14.1
All checks were successful
Analysis / SonarCloud (pull_request) Successful in 35s
Tests / Build and Test (3.11) (pull_request) Successful in 1m14s
Tests / Build and Test (3.12) (pull_request) Successful in 1m37s
Tests / Build and Test (3.13) (pull_request) Successful in 1m22s
Analysis / SonarCloud (push) Successful in 47s
2024-12-30 19:17:40 +02:00
kku
bbbec75f91 Fixed naming conventions
All checks were successful
Analysis / SonarCloud (push) Successful in 44s
2024-12-29 19:27:42 +01:00
kku
94553b602e Fixed imports in examples 2024-12-29 16:27:58 +01:00
28 changed files with 536 additions and 97 deletions

View File

@@ -6,19 +6,18 @@ on:
- main
- dev
pull_request:
types: [opened, synchronize, reopened]
types: [ opened, synchronize, reopened ]
jobs:
sonarcloud:
name: SonarCloud
runs-on: ubuntu-latest
container: catthehacker/ubuntu:act-latest
runs-on: ubuntu-24.04
steps:
- uses: actions/checkout@v4
with:
fetch-depth: 0
- name: SonarCloud Scan
uses: SonarSource/sonarcloud-github-action@master
- name: SonarQube Scan
uses: SonarSource/sonarqube-scan-action@v4.2.1
env:
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
SONAR_TOKEN: ${{ secrets.SONAR_TOKEN }}

View File

@@ -0,0 +1,67 @@
name: Upload Python Package
on:
release:
types: [ published ]
permissions:
contents: read
jobs:
release-build:
runs-on: ubuntu-24.04
steps:
- uses: actions/checkout@v4
- uses: actions/setup-python@v5
with:
python-version: "3.x"
- name: Build release distributions
run: |
python -m pip install build
python -m build
- name: Upload distributions
uses: christopherhx/gitea-upload-artifact@v4
with:
name: release-dists
path: dist/
gitea-publish:
runs-on: ubuntu-24.04
needs: release-build
permissions:
id-token: write
environment:
name: gitea
url: https://git.end-play.xyz/profitroll/-/packages/pypi/libbot
env:
GITHUB_WORKFLOW_REF: ${{ gitea.workflow_ref }}
INPUT_REPOSITORY_URL: https://git.end-play.xyz/api/packages/profitroll/pypi
steps:
- name: Retrieve release distributions
uses: christopherhx/gitea-download-artifact@v4
with:
name: release-dists
path: dist/
- name: Publish package distributions to TestPyPI
uses: pypa/gh-action-pypi-publish@release/v1
with:
password: ${{ secrets.PYPI_GITEA_API_TOKEN }}
repository-url: https://git.end-play.xyz/api/packages/profitroll/pypi
pypi-publish:
runs-on: ubuntu-24.04
needs: release-build
permissions:
id-token: write
environment:
name: pypi
env:
GITHUB_WORKFLOW_REF: ${{ gitea.workflow_ref }}
steps:
- name: Retrieve release distributions
uses: christopherhx/gitea-download-artifact@v4
with:
name: release-dists
path: dist/
- name: Publish package distributions to TestPyPI
uses: pypa/gh-action-pypi-publish@release/v1
with:
password: ${{ secrets.PYPI_PYPI_API_TOKEN }}

View File

@@ -11,18 +11,18 @@ on:
jobs:
test:
name: Build and Test
runs-on: ubuntu-latest
container: catthehacker/ubuntu:act-latest
runs-on: ubuntu-24.04
strategy:
matrix:
python-version: [ "3.11", "3.12", "3.13" ]
steps:
- uses: actions/checkout@v3
- name: Set up Python ${{ matrix.python-version }}
uses: actions/setup-python@v3
with:
python-version: ${{ matrix.python-version }}
cache: 'pip'
cache-dependency-path: './requirements/*'
env:
AGENT_TOOLSDIRECTORY: /opt/hostedtoolcache
- name: Install dependencies

View File

@@ -36,6 +36,8 @@ pip install libbot[pycord,speed]
### Pyrogram
```python
import sys
from libbot.pyrogram.classes import PyroClient
@@ -47,7 +49,7 @@ def main():
except KeyboardInterrupt:
print("Shutting down...")
finally:
exit()
sys.exit()
if __name__ == "__main__":
@@ -57,6 +59,9 @@ if __name__ == "__main__":
### Pycord
```python
import asyncio
from asyncio import AbstractEventLoop
from discord import Intents
from libbot.utils import config_get
from libbot.pycord.classes import PycordBot
@@ -76,7 +81,7 @@ async def main():
if __name__ == "__main__":
loop = asyncio.get_event_loop()
loop: AbstractEventLoop = asyncio.get_event_loop()
loop.run_until_complete(main())
```

4
examples/README.md Normal file
View File

@@ -0,0 +1,4 @@
# Examples
If you're looking for Pyrogram usage examples, please take a look at
the [PyrogramBotBase](https://git.end-play.xyz/profitroll/PyrogramBotBase) repository.

View File

@@ -31,6 +31,7 @@ dev = { file = "requirements/dev.txt" }
pycord = { file = "requirements/pycord.txt" }
pyrogram = { file = "requirements/pyrogram.txt" }
speed = { file = "requirements/speed.txt" }
cache = { file = "requirements/cache.txt" }
[project.urls]
Source = "https://git.end-play.xyz/profitroll/LibBotUniversal"

View File

@@ -1,2 +1,2 @@
aiofiles>=23.0.0
typing_extensions~=4.12.2
typing-extensions~=4.12.2

2
requirements/cache.txt Normal file
View File

@@ -0,0 +1,2 @@
pymemcache~=4.0.0
redis~=5.2.1

View File

@@ -1,12 +1,12 @@
black==24.10.0
black==25.1.0
build==1.2.2.post1
isort==5.13.2
mypy==1.14.0
pylint==3.3.3
pytest-asyncio==0.25.0
mypy==1.15.0
pylint==3.3.4
pytest-asyncio==0.25.3
pytest-cov==6.0.0
pytest==8.3.4
tox==4.23.2
twine==6.0.1
tox==4.24.0
twine==6.1.0
types-aiofiles==24.1.0.20241221
types-ujson==5.10.0.20240515

View File

@@ -1,4 +1,4 @@
__version__ = "4.0.1"
__version__ = "4.1.0"
__license__ = "GPL3"
__author__ = "Profitroll"

2
src/libbot/cache/__init__.py vendored Normal file
View File

@@ -0,0 +1,2 @@
# This file is left empty on purpose
# Adding imports here will cause import errors when libbot[pycord] is not installed

3
src/libbot/cache/classes/__init__.py vendored Normal file
View File

@@ -0,0 +1,3 @@
from .cache import Cache
from .cache_memcached import CacheMemcached
from .cache_redis import CacheRedis

44
src/libbot/cache/classes/cache.py vendored Normal file
View File

@@ -0,0 +1,44 @@
from abc import ABC, abstractmethod
from typing import Any, Dict
import pymemcache
import redis
class Cache(ABC):
client: pymemcache.Client | redis.Redis
@classmethod
@abstractmethod
def from_config(cls, engine_config: Dict[str, Any]) -> Any:
pass
@abstractmethod
def get_json(self, key: str) -> Any | None:
# TODO This method must also carry out ObjectId conversion!
pass
@abstractmethod
def get_string(self, key: str) -> str | None:
pass
@abstractmethod
def get_object(self, key: str) -> Any | None:
pass
@abstractmethod
def set_json(self, key: str, value: Any) -> None:
# TODO This method must also carry out ObjectId conversion!
pass
@abstractmethod
def set_string(self, key: str, value: str) -> None:
pass
@abstractmethod
def set_object(self, key: str, value: Any) -> None:
pass
@abstractmethod
def delete(self, key: str) -> None:
pass

View File

@@ -0,0 +1,89 @@
import logging
from logging import Logger
from typing import Dict, Any
from pymemcache import Client
from .cache import Cache
from ..utils._objects import _json_to_string, _string_to_json
logger: Logger = logging.getLogger(__name__)
class CacheMemcached(Cache):
client: Client
def __init__(self, client: Client):
self.client = client
logger.info("Initialized Memcached for caching")
@classmethod
def from_config(cls, engine_config: Dict[str, Any]) -> "CacheMemcached":
if "uri" not in engine_config:
raise KeyError(
"Cache configuration is invalid. Please check if all keys are set (engine: memcached)"
)
return cls(Client(engine_config["uri"], default_noreply=True))
def get_json(self, key: str) -> Any | None:
try:
result: Any | None = self.client.get(key, None)
logger.debug(
"Got json cache key '%s'%s",
key,
"" if result is not None else " (not found)",
)
except Exception as exc:
logger.error("Could not get json cache key '%s' due to: %s", key, exc)
return None
return None if result is None else _string_to_json(result)
def get_string(self, key: str) -> str | None:
try:
result: str | None = self.client.get(key, None)
logger.debug(
"Got string cache key '%s'%s",
key,
"" if result is not None else " (not found)",
)
return result
except Exception as exc:
logger.error("Could not get string cache key '%s' due to: %s", key, exc)
return None
# TODO Implement binary deserialization
def get_object(self, key: str) -> Any | None:
raise NotImplementedError()
def set_json(self, key: str, value: Any) -> None:
try:
self.client.set(key, _json_to_string(value))
logger.debug("Set json cache key '%s'", key)
except Exception as exc:
logger.error("Could not set json cache key '%s' due to: %s", key, exc)
return None
def set_string(self, key: str, value: str) -> None:
try:
self.client.set(key, value)
logger.debug("Set string cache key '%s'", key)
except Exception as exc:
logger.error("Could not set string cache key '%s' due to: %s", key, exc)
return None
# TODO Implement binary serialization
def set_object(self, key: str, value: Any) -> None:
raise NotImplementedError()
def delete(self, key: str) -> None:
try:
self.client.delete(key)
logger.debug("Deleted cache key '%s'", key)
except Exception as exc:
logger.error("Could not delete cache key '%s' due to: %s", key, exc)

89
src/libbot/cache/classes/cache_redis.py vendored Normal file
View File

@@ -0,0 +1,89 @@
import logging
from logging import Logger
from typing import Dict, Any
from redis import Redis
from .cache import Cache
from ..utils._objects import _string_to_json, _json_to_string
logger: Logger = logging.getLogger(__name__)
class CacheRedis(Cache):
client: Redis
def __init__(self, client: Redis):
self.client = client
logger.info("Initialized Redis for caching")
@classmethod
def from_config(cls, engine_config: Dict[str, Any]) -> Any:
if "uri" not in engine_config:
raise KeyError(
"Cache configuration is invalid. Please check if all keys are set (engine: memcached)"
)
return cls(Redis.from_url(engine_config["uri"]))
def get_json(self, key: str) -> Any | None:
try:
result: Any | None = self.client.get(key)
logger.debug(
"Got json cache key '%s'%s",
key,
"" if result is not None else " (not found)",
)
except Exception as exc:
logger.error("Could not get json cache key '%s' due to: %s", key, exc)
return None
return None if result is None else _string_to_json(result)
def get_string(self, key: str) -> str | None:
try:
result: str | None = self.client.get(key)
logger.debug(
"Got string cache key '%s'%s",
key,
"" if result is not None else " (not found)",
)
return result
except Exception as exc:
logger.error("Could not get string cache key '%s' due to: %s", key, exc)
return None
# TODO Implement binary deserialization
def get_object(self, key: str) -> Any | None:
raise NotImplementedError()
def set_json(self, key: str, value: Any) -> None:
try:
self.client.set(key, _json_to_string(value))
logger.debug("Set json cache key '%s'", key)
except Exception as exc:
logger.error("Could not set json cache key '%s' due to: %s", key, exc)
return None
def set_string(self, key: str, value: str) -> None:
try:
self.client.set(key, value)
logger.debug("Set string cache key '%s'", key)
except Exception as exc:
logger.error("Could not set string cache key '%s' due to: %s", key, exc)
return None
# TODO Implement binary serialization
def set_object(self, key: str, value: Any) -> None:
raise NotImplementedError()
def delete(self, key: str) -> None:
try:
self.client.delete(key)
logger.debug("Deleted cache key '%s'", key)
except Exception as exc:
logger.error("Could not delete cache key '%s' due to: %s", key, exc)

1
src/libbot/cache/manager/__init__.py vendored Normal file
View File

@@ -0,0 +1 @@
from .manager import create_cache_client

24
src/libbot/cache/manager/manager.py vendored Normal file
View File

@@ -0,0 +1,24 @@
from typing import Dict, Any, Literal
from ..classes import CacheMemcached, CacheRedis
def create_cache_client(
config: Dict[str, Any],
engine: Literal["memcached", "redis"] | None = None,
) -> CacheMemcached | CacheRedis:
if engine not in ["memcached", "redis"] or engine is None:
raise KeyError(f"Incorrect cache engine provided. Expected 'memcached' or 'redis', got '{engine}'")
if "cache" not in config or engine not in config["cache"]:
raise KeyError(
f"Cache configuration is invalid. Please check if all keys are set (engine: '{engine}')"
)
match engine:
case "memcached":
return CacheMemcached.from_config(config["cache"][engine])
case "redis":
return CacheRedis.from_config(config["cache"][engine])
case _:
raise KeyError(f"Cache implementation for the engine '{engine}' is not present.")

0
src/libbot/cache/utils/__init__.py vendored Normal file
View File

42
src/libbot/cache/utils/_objects.py vendored Normal file
View File

@@ -0,0 +1,42 @@
import logging
from copy import deepcopy
from logging import Logger
from typing import Any
try:
from ujson import dumps, loads
except ImportError:
from json import dumps, loads
logger: Logger = logging.getLogger(__name__)
try:
from bson import ObjectId
except ImportError:
logger.warning(
"Could not import bson.ObjectId. PyMongo conversions will not be supported by the cache. It's safe to ignore this message if you do not use MongoDB."
)
def _json_to_string(json_object: Any) -> str:
json_object_copy: Any = deepcopy(json_object)
if isinstance(json_object_copy, dict) and "_id" in json_object_copy:
json_object_copy["_id"] = str(json_object_copy["_id"])
return dumps(json_object_copy, ensure_ascii=False, indent=0, escape_forward_slashes=False)
def _string_to_json(json_string: str) -> Any:
json_object: Any = loads(json_string)
if "_id" in json_object:
try:
json_object["_id"] = ObjectId(json_object["_id"])
except NameError:
logger.debug(
"Tried to convert attribute '_id' with value '%s' but bson.ObjectId is not present, skipping the conversion.",
json_object["_id"],
)
return json_object

View File

@@ -1,9 +1,13 @@
import logging
from logging import Logger
from pathlib import Path
from typing import Any, Dict
from typing_extensions import override
from ...i18n.classes import BotLocale
from ...utils import json_read
try:
from apscheduler.schedulers.asyncio import AsyncIOScheduler
from apscheduler.schedulers.background import BackgroundScheduler
@@ -11,32 +15,21 @@ try:
except ImportError as exc:
raise ImportError("You need to install libbot[pycord] in order to use this class.") from exc
try:
from ujson import loads
except ImportError:
from json import loads
from ...i18n.classes import BotLocale
logger = logging.getLogger(__name__)
logger: Logger = logging.getLogger(__name__)
class PycordBot(Bot):
@override
def __init__(
self,
*args,
config: Dict[str, Any] | None = None,
config_path: str | Path = Path("config.json"),
locales_root: str | Path | None = None,
scheduler: AsyncIOScheduler | BackgroundScheduler | None = None,
**kwargs,
self,
*args,
config: Dict[str, Any] | None = None,
config_path: str | Path = Path("config.json"),
locales_root: str | Path | None = None,
scheduler: AsyncIOScheduler | BackgroundScheduler | None = None,
**kwargs,
):
if config is None:
with open(config_path, "r", encoding="utf-8") as f:
self.config: dict = loads(f.read())
else:
self.config = config
self.config: Dict[str, Any] = config if config is not None else json_read(config_path)
super().__init__(
debug_guilds=(self.config["bot"]["debug_guilds"] if self.config["debug"] else None),

View File

@@ -0,0 +1 @@
from .color import color_from_hex, hex_from_color

View File

@@ -0,0 +1,35 @@
from discord import Colour
def _int_from_hex(hex_string: str) -> int:
try:
return int(hex_string, base=16)
except Exception as exc:
raise ValueError("Input string must be a valid HEX code.") from exc
def _hex_from_int(color_int: int) -> str:
if not 0 <= color_int <= 0xFFFFFF:
raise ValueError("Color's value must be in the range 0 to 0xFFFFFF.")
return f"#{color_int:06x}"
def color_from_hex(hex_string: str) -> Colour:
"""Convert valid hexadecimal string to discord.Colour.
:param hex_string: Hexadecimal string to convert into Colour object
:type hex_string: str
:return: Colour object
"""
return Colour(_int_from_hex(hex_string))
def hex_from_color(color: Colour) -> str:
"""Convert discord.Colour to hexadecimal string.
:param color: Colour object to convert into the string
:type color: Colour
:return: Hexadecimal string in #XXXXXX format
"""
return _hex_from_int(color.value)

View File

@@ -2,6 +2,7 @@ import asyncio
import logging
import sys
from datetime import datetime, timedelta
from logging import Logger
from os import cpu_count, getpid
from pathlib import Path
from time import time
@@ -9,6 +10,12 @@ from typing import Any, Dict, List
from typing_extensions import override
from .command import PyroCommand
from .commandset import CommandSet
from ...i18n import _
from ...i18n.classes import BotLocale
from ...utils import json_read
try:
import pyrogram
from apscheduler.schedulers.asyncio import AsyncIOScheduler
@@ -35,42 +42,33 @@ try:
except ImportError:
from json import dumps, loads
from ...i18n.classes import BotLocale
from ...i18n import _
from .command import PyroCommand
from .commandset import CommandSet
logger = logging.getLogger(__name__)
logger: Logger = logging.getLogger(__name__)
class PyroClient(Client):
@override
def __init__(
self,
name: str = "bot_client",
owner: int | None = None,
config: Dict[str, Any] | None = None,
config_path: str | Path = Path("config.json"),
api_id: int | None = None,
api_hash: str | None = None,
bot_token: str | None = None,
workers: int = min(32, cpu_count() + 4),
locales_root: str | Path | None = None,
plugins_root: str = "plugins",
plugins_exclude: List[str] | None = None,
sleep_threshold: int = 120,
max_concurrent_transmissions: int = 1,
commands_source: Dict[str, dict] | None = None,
scoped_commands: bool | None = None,
i18n_bot_info: bool = False,
scheduler: AsyncIOScheduler | BackgroundScheduler | None = None,
**kwargs,
self,
name: str = "bot_client",
owner: int | None = None,
config: Dict[str, Any] | None = None,
config_path: str | Path = Path("config.json"),
api_id: int | None = None,
api_hash: str | None = None,
bot_token: str | None = None,
workers: int = min(32, cpu_count() + 4),
locales_root: str | Path | None = None,
plugins_root: str = "plugins",
plugins_exclude: List[str] | None = None,
sleep_threshold: int = 120,
max_concurrent_transmissions: int = 1,
commands_source: Dict[str, dict] | None = None,
scoped_commands: bool | None = None,
i18n_bot_info: bool = False,
scheduler: AsyncIOScheduler | BackgroundScheduler | None = None,
**kwargs,
):
if config is None:
with open(config_path, "r", encoding="utf-8") as f:
self.config: dict = loads(f.read())
else:
self.config = config
self.config: Dict[str, Any] = config if config is not None else json_read(config_path)
super().__init__(
name=name,
@@ -211,7 +209,7 @@ class PyroClient(Client):
@override
async def stop(
self, exit_completely: bool = True, scheduler_shutdown: bool = True, scheduler_wait: bool = True
self, exit_completely: bool = True, scheduler_shutdown: bool = True, scheduler_wait: bool = True
) -> None:
try:
await self.send_message(
@@ -310,7 +308,7 @@ class PyroClient(Client):
# in it, if there are any. Then adds them to self.commands
for handler in self.dispatcher.groups[0]:
if isinstance(handler, MessageHandler) and (
hasattr(handler.filters, "base") or hasattr(handler.filters, "other")
hasattr(handler.filters, "base") or hasattr(handler.filters, "other")
):
for entry in [handler.filters.base, handler.filters.other]:
if hasattr(entry, "commands"):
@@ -321,8 +319,8 @@ class PyroClient(Client):
return command_sets
def add_command(
self,
command: str,
self,
command: str,
) -> None:
"""Add command to the bot's internal commands list

View File

@@ -20,7 +20,7 @@ def config_get(key: str, *path: str, config_file: str | Path = DEFAULT_CONFIG_LO
### Args:
* key (`str`): Key that contains the value
* *path (`str`): Path to the key that contains the value
* *path (`str`): Path to the key that contains the value (pass *[] or don't pass anything at all to get on the top/root level)
* config_file (`str | Path`, *optional*): Path-like object or path as a string of a location of the config file. Defaults to `"config.json"`
### Returns:
@@ -59,7 +59,7 @@ async def config_get(key: str, *path: str, config_file: str | Path = DEFAULT_CON
### Args:
* key (`str`): Key that contains the value
* *path (`str`): Path to the key that contains the value
* *path (`str`): Path to the key that contains the value (pass *[] or don't pass anything at all to get on the top/root level)
* config_file (`str | Path`, *optional*): Path-like object or path as a string of a location of the config file. Defaults to `"config.json"`
### Returns:
@@ -98,7 +98,7 @@ def config_set(key: str, value: Any, *path: str, config_file: str | Path = DEFAU
### Args:
* key (`str`): Key that leads to the value
* value (`Any`): Any JSON serializable data
* *path (`str`): Path to the key of the target
* *path (`str`): Path to the key of the target (pass *[] or don't pass anything at all to set on the top/root level)
* config_file (`str | Path`, *optional*): Path-like object or path as a string of a location of the config file. Defaults to `"config.json"`
### Raises:
@@ -116,7 +116,7 @@ async def config_set(
### Args:
* key (`str`): Key that leads to the value
* value (`Any`): Any JSON serializable data
* *path (`str`): Path to the key of the target
* *path (`str`): Path to the key of the target (pass *[] or don't pass anything at all to set on the top/root level)
* config_file (`str | Path`, *optional*): Path-like object or path as a string of a location of the config file. Defaults to `"config.json"`
### Raises:
@@ -136,7 +136,7 @@ def config_delete(
### Args:
* key (`str`): Key to delete
* *path (`str`): Path to the key of the target
* *path (`str`): Path to the key of the target (pass *[] or don't pass anything at all to delete on the top/root level)
* missing_ok (`bool`): Do not raise an exception if the key is missing. Defaults to `False`
* config_file (`str | Path`, *optional*): Path-like object or path as a string of a location of the config file. Defaults to `"config.json"`
@@ -165,7 +165,7 @@ async def config_delete(
### Args:
* key (`str`): Key to delete
* *path (`str`): Path to the key of the target
* *path (`str`): Path to the key of the target (pass *[] or don't pass anything at all to delete on the top/root level)
* missing_ok (`bool`): Do not raise an exception if the key is missing. Defaults to `False`
* config_file (`str | Path`, *optional*): Path-like object or path as a string of a location of the config file. Defaults to `"config.json"`

View File

@@ -3,11 +3,11 @@ from typing import Any, Dict
from typing import Callable
def supports_argument(func: Callable, arg_name: str) -> bool:
def supports_argument(func: Callable[..., Any], arg_name: str) -> bool:
"""Check whether a function has a specific argument
### Args:
* func (`Callable`): Function to be inspected
* func (`Callable[..., Any]`): Function to be inspected
* arg_name (`str`): Argument to be checked
### Returns:
@@ -24,11 +24,13 @@ def supports_argument(func: Callable, arg_name: str) -> bool:
return False
def nested_set(target: dict, value: Any, *path: str, create_missing=True) -> Dict[str, Any]:
def nested_set(
target: Dict[str, Any], value: Any, *path: str, create_missing: bool = True
) -> Dict[str, Any]:
"""Set the key by its path to the value
### Args:
* target (`dict`): Dictionary to perform modifications on
* target (`Dict[str, Any]`): Dictionary to perform modifications on
* value (`Any`): Any data
* *path (`str`): Path to the key of the target
* create_missing (`bool`, *optional*): Create keys on the way if they're missing. Defaults to `True`
@@ -39,29 +41,29 @@ def nested_set(target: dict, value: Any, *path: str, create_missing=True) -> Dic
### Returns:
* `Dict[str, Any]`: Changed dictionary
"""
d = target
target_copy: Dict[str, Any] = target
for key in path[:-1]:
if key in d:
d = d[key]
if key in target_copy:
target_copy = target_copy[key]
elif create_missing:
d = d.setdefault(key, {})
target_copy = target_copy.setdefault(key, {})
else:
raise KeyError(
f"Key '{key}' is not found under path provided ({path}) and create_missing is False"
)
if path[-1] in d or create_missing:
d[path[-1]] = value
if path[-1] in target_copy or create_missing:
target_copy[path[-1]] = value
return target
def nested_delete(target: dict, *path: str) -> Dict[str, Any]:
def nested_delete(target: Dict[str, Any], *path: str) -> Dict[str, Any]:
"""Delete the key by its path
### Args:
* target (`dict`): Dictionary to perform modifications on
* target (`Dict[str, Any]`): Dictionary to perform modifications on
### Raises:
* `KeyError`: Key is not found under path provided
@@ -69,16 +71,16 @@ def nested_delete(target: dict, *path: str) -> Dict[str, Any]:
### Returns:
`Dict[str, Any]`: Changed dictionary
"""
d = target
target_copy: Dict[str, Any] = target
for key in path[:-1]:
if key in d:
d = d[key]
if key in target_copy:
target_copy = target_copy[key]
else:
raise KeyError(f"Key '{key}' is not found under path provided ({path})")
if path[-1] in d:
del d[path[-1]]
if path[-1] in target_copy:
del target_copy[path[-1]]
else:
raise KeyError(f"Key '{path[-1]}' is not found under path provided ({path})")

View File

@@ -2,5 +2,14 @@
"locale": "en",
"bot": {
"bot_token": "sample_token"
},
"cache": {
"type": "memcached",
"memcached": {
"uri": "127.0.0.1:11211"
},
"redis": {
"uri": "redis://127.0.0.1:6379/0"
}
}
}

28
tests/test_cache.py Normal file
View File

@@ -0,0 +1,28 @@
from pathlib import Path
from libbot.cache.classes import Cache
from libbot.cache.manager import create_cache_client
try:
from ujson import JSONDecodeError, dumps, loads
except ImportError:
from json import JSONDecodeError, dumps, loads
from typing import Any, Dict
import pytest
@pytest.mark.parametrize(
"engine",
[
"memcached",
"redis",
],
)
def test_cache_creation(engine: str, location_config: Path):
with open(location_config, "r", encoding="utf-8") as file:
config: Dict[str, Any] = loads(file.read())
cache: Cache = create_cache_client(config, engine)
assert isinstance(cache, Cache)

View File

@@ -18,5 +18,6 @@ deps =
-r{toxinidir}/requirements/pycord.txt
-r{toxinidir}/requirements/pyrogram.txt
-r{toxinidir}/requirements/speed.txt
-r{toxinidir}/requirements/cache.txt
commands =
pytest --basetemp={envtmpdir} --cov=libbot