Page Menu
Home
Software Heritage
Search
Configure Global Search
Log In
Files
F9346013
No One
Temporary
Actions
View File
Edit File
Delete File
View Transforms
Subscribe
Mute Notifications
Award Token
Flag For Later
Size
4 KB
Subscribers
None
View Options
diff --git a/swh/indexer/language.py b/swh/indexer/language.py
index 913108a..6389e2c 100644
--- a/swh/indexer/language.py
+++ b/swh/indexer/language.py
@@ -1,142 +1,148 @@
# Copyright (C) 2016-2017 The Software Heritage developers
# See the AUTHORS file at the top-level directory of this distribution
# License: GNU General Public License version 3, or any later version
# See top-level LICENSE file for more information
import io
from pygments.lexers import guess_lexer
from chardet.universaldetector import UniversalDetector
from .indexer import BaseIndexer
def _cleanup_classname(classname):
"""Determine the language from the pygments' lexer names.
"""
return classname.lower().replace(' ', '-')
def _read_raw(raw_content, size=2048):
"""Read raw content in chunk.
"""
bs = io.BytesIO(raw_content)
while True:
chunk = bs.read(size)
if not chunk:
break
yield chunk
def _detect_encoding(raw_content):
"""Given a raw content, try and detect its encoding.
"""
detector = UniversalDetector()
for chunk in _read_raw(raw_content):
detector.feed(chunk)
if detector.done:
break
detector.close()
return detector.result['encoding']
def compute_language(raw_content):
"""Determine the raw content's language.
Args:
raw_content (bytes): content to determine raw content
Returns:
Dict with keys:
- lang: None if nothing found or the possible language
- decoding_failure: True if a decoding failure happened
"""
try:
encoding = _detect_encoding(raw_content)
content = raw_content.decode(encoding)
lang = _cleanup_classname(
guess_lexer(content).name)
return {
'lang': lang
}
except Exception:
return {
'lang': None
}
class ContentLanguageIndexer(BaseIndexer):
"""Indexer in charge of:
- filtering out content already indexed
- reading content from objstorage per the content's id (sha1)
- computing {mimetype, encoding} from that content
- store result in storage
"""
CONFIG_BASE_FILENAME = 'indexer/language'
ADDITIONAL_CONFIG = {
'tool': ('dict', {
'name': 'pygments',
'version': '2.0.1+dfsg-1.1+deb8u1',
+ 'max_content_size': 10240,
}),
}
def __init__(self):
super().__init__()
self.tool_name = self.config['tool']['name']
self.tool_version = self.config['tool']['version']
+ self.max_content_size = self.config['tool']['max_content_size']
def filter_contents(self, sha1s):
"""Filter out known sha1s and return only missing ones.
"""
yield from self.storage.content_language_missing((
{
'id': sha1,
'tool_name': self.tool_name,
'tool_version': self.tool_version
} for sha1 in sha1s
))
def index_content(self, sha1, raw_content):
"""Index sha1s' content and store result.
Args:
sha1 (bytes): content's identifier
raw_content (bytes): raw content in bytes
Returns:
A dict, representing a content_mimetype, with keys:
- id (bytes): content's identifier (sha1)
- lang (bytes): detected language
"""
+ l = len(raw_content)
+ if self.max_content_size <= l:
+ raw_content = raw_content[0:self.max_content_size]
+
result = compute_language(raw_content)
result.update({
'id': sha1,
'tool_name': self.tool_name,
'tool_version': self.tool_version,
})
return result
def persist_index_computations(self, results, policy_update):
"""Persist the results in storage.
Args:
results ([dict]): list of content_mimetype, dict with the
following keys:
- id (bytes): content's identifier (sha1)
- lang (bytes): detected language
policy_update ([str]): either 'update-dups' or 'ignore-dups' to
respectively update duplicates or ignore them
"""
self.storage.content_language_add(
results, conflict_update=(policy_update == 'update-dups'))
File Metadata
Details
Attached
Mime Type
text/x-diff
Expires
Fri, Jul 4, 3:40 PM (1 w, 6 d ago)
Storage Engine
blob
Storage Format
Raw Data
Storage Handle
3277274
Attached To
rDCIDX Metadata indexer
Event Timeline
Log In to Comment