BDEPEND=python_targets_python3_10? ( dev-lang/python:3.10 ) python_targets_python3_11? ( dev-lang/python:3.11 ) python_targets_python3_12? ( dev-lang/python:3.12 ) python_targets_python3_13? ( dev-lang/python:3.13 ) >=dev-python/gpep517-16[python_targets_python3_10(-)?,python_targets_python3_11(-)?,python_targets_python3_12(-)?,python_targets_python3_13(-)?] >=dev-python/setuptools-78.1.0[python_targets_python3_10(-)?,python_targets_python3_11(-)?,python_targets_python3_12(-)?,python_targets_python3_13(-)?] DEFINED_PHASES=compile configure install prepare test DEPEND=sci-ml/sentencepiece DESCRIPTION=Text tokenizer for Neural Network-based text generation EAPI=8 HOMEPAGE=https://github.com/google/sentencepiece INHERIT=distutils-r1 IUSE=python_targets_python3_10 python_targets_python3_11 python_targets_python3_12 python_targets_python3_13 KEYWORDS=~amd64 LICENSE=Apache-2.0 RDEPEND=sci-ml/sentencepiece python_targets_python3_10? ( dev-lang/python:3.10 ) python_targets_python3_11? ( dev-lang/python:3.11 ) python_targets_python3_12? ( dev-lang/python:3.12 ) python_targets_python3_13? ( dev-lang/python:3.13 ) REQUIRED_USE=|| ( python_targets_python3_10 python_targets_python3_11 python_targets_python3_12 python_targets_python3_13 ) SLOT=0 SRC_URI=https://github.com/google/sentencepiece/archive/refs/tags/v0.2.0.tar.gz -> sentencepiece-0.2.0.tar.gz _eclasses_=distutils-r1 3641642ea51cdadc1d1ea0bb8404ae89 flag-o-matic e8de74bac929ba17427e740e95707d00 multibuild 4650a65187015567b4e041bb9bfdb364 multilib b2a329026f2e404e9e371097dda47f96 multiprocessing 1e32df7deee68372153dca65f4a7c21f ninja-utils 2df4e452cea39a9ec8fb543ce059f8d6 out-of-source-utils dbf9e34ee8964084651e25907fa8f52c python-r1 0e15b2ab9cfc87d7474678201d6bca38 python-utils-r1 974462533ec80cd68e5f366ad8a86002 toolchain-funcs f9d71a6efe9d083aec750dd13968e169 _md5_=e0475d1a7598f5e01b83d8a647fad9aa