Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Submit feedback
Contribute to GitLab
Sign in
Toggle navigation
O
OpnSense
Project
Project
Details
Activity
Releases
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
Kulya
OpnSense
Commits
4b9c1e0b
Commit
4b9c1e0b
authored
Mar 30, 2016
by
Ad Schellevis
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
style fixes fetchACLs
parent
a4c5f2c7
Changes
1
Show whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
32 additions
and
27 deletions
+32
-27
fetchACLs.py
src/opnsense/scripts/proxy/fetchACLs.py
+32
-27
No files found.
src/opnsense/scripts/proxy/fetchACLs.py
View file @
4b9c1e0b
...
...
@@ -36,17 +36,18 @@ import os.path
import
tarfile
import
gzip
import
zipfile
import
StringIO
import
syslog
from
ConfigParser
import
ConfigParser
acl_config_fn
=
(
'/usr/local/etc/squid/externalACLs.conf'
)
acl_target_dir
=
(
'/usr/local/etc/squid/acl'
)
acl_config_fn
=
'/usr/local/etc/squid/externalACLs.conf'
acl_target_dir
=
'/usr/local/etc/squid/acl'
acl_max_timeout
=
30
class
Downloader
(
object
):
""" Download helper
"""
def
__init__
(
self
,
url
,
timeout
):
""" init new
:param url: source url
...
...
@@ -60,7 +61,7 @@ class Downloader(object):
""" fetch (raw) source data into tempfile using self._source_handle
"""
try
:
f
=
urllib2
.
urlopen
(
self
.
_url
,
timeout
=
self
.
_timeout
)
f
=
urllib2
.
urlopen
(
self
.
_url
,
timeout
=
self
.
_timeout
)
# flush to temp file
self
.
_source_handle
=
tempfile
.
NamedTemporaryFile
()
while
True
:
...
...
@@ -72,7 +73,7 @@ class Downloader(object):
self
.
_source_handle
.
seek
(
0
)
f
.
close
()
except
(
urllib2
.
URLError
,
urllib2
.
HTTPError
,
IOError
)
as
e
:
syslog
.
syslog
(
syslog
.
LOG_ERR
,
'proxy acl: error downloading
%
s'
%
self
.
_url
)
syslog
.
syslog
(
syslog
.
LOG_ERR
,
'proxy acl: error downloading
%
s'
%
self
.
_url
)
self
.
_source_handle
=
None
def
get_files
(
self
):
...
...
@@ -90,17 +91,16 @@ class Downloader(object):
if
tf_file
.
isfile
():
yield
tf_file
.
name
,
tf
.
extractfile
(
tf_file
)
except
IOError
as
e
:
syslog
.
syslog
(
syslog
.
LOG_ERR
,
'proxy acl: error downloading
%
s (
%
s)'
%
(
self
.
_url
,
e
))
syslog
.
syslog
(
syslog
.
LOG_ERR
,
'proxy acl: error downloading
%
s (
%
s)'
%
(
self
.
_url
,
e
))
elif
len
(
self
.
_url
)
>
4
and
self
.
_url
[
-
3
:]
==
'.gz'
:
# source is in .gz format unpack
try
:
gf
=
gzip
.
GzipFile
(
mode
=
'r'
,
fileobj
=
self
.
_source_handle
)
yield
os
.
path
.
basename
(
self
.
_url
),
gf
except
IOError
as
e
:
syslog
.
syslog
(
syslog
.
LOG_ERR
,
'proxy acl: error downloading
%
s (
%
s)'
%
(
self
.
_url
,
e
))
syslog
.
syslog
(
syslog
.
LOG_ERR
,
'proxy acl: error downloading
%
s (
%
s)'
%
(
self
.
_url
,
e
))
elif
len
(
self
.
_url
)
>
5
and
self
.
_url
[
-
4
:]
==
'.zip'
:
# source is in .zip format, extract all into a single string
target_data
=
dict
()
with
zipfile
.
ZipFile
(
self
.
_source_handle
,
mode
=
'r'
,
compression
=
zipfile
.
ZIP_DEFLATED
)
as
zf
:
...
...
@@ -127,6 +127,7 @@ class DomainSorter(object):
""" Helper class for building sorted squid domain acl list.
Use as file type object, close flushes the actual (sorted) data to disc
"""
def
__init__
(
self
,
filename
=
None
,
mode
=
None
):
""" new sorted output file, uses an acl record in reverse order as sort key
:param filename: target filename
...
...
@@ -147,7 +148,7 @@ class DomainSorter(object):
"""
sets
=
255
for
i
in
range
(
sets
):
target
=
chr
(
i
+
1
)
target
=
chr
(
i
+
1
)
setid
=
int
(
i
/
(
sets
/
self
.
_num_targets
))
if
setid
not
in
self
.
_buckets
:
self
.
_buckets
[
setid
]
=
tempfile
.
NamedTemporaryFile
()
...
...
@@ -173,7 +174,7 @@ class DomainSorter(object):
"""
target
=
key
[
0
]
if
target
in
self
.
_sort_map
:
self
.
_sort_map
[
target
]
.
write
(
'
%
s
%
s
%
s
\n
'
%
(
key
,
self
.
_seperator
,
value
))
self
.
_sort_map
[
target
]
.
write
(
'
%
s
%
s
%
s
\n
'
%
(
key
,
self
.
_seperator
,
value
))
else
:
# not supposed to happen, every key should have a calculated target pool
pass
...
...
@@ -224,7 +225,7 @@ class DomainSorter(object):
continue
if
self
.
is_domain
(
line
):
# prefix domain, if this domain is different then the previous one
if
prev_line
is
None
or
'.
%
s'
%
line
not
in
prev_line
:
if
prev_line
is
None
or
'.
%
s'
%
line
not
in
prev_line
:
f_out
.
write
(
'.'
)
f_out
.
write
(
line
)
prev_line
=
line
...
...
@@ -234,12 +235,13 @@ def filename_in_ignorelist(filename):
""" ignore certain files from processing.
:param filename: filename to inspect
"""
if
(
filename
.
lower
()
.
split
(
'.'
)[
-
1
]
in
[
'pdf'
,
'txt'
,
'doc'
])
:
if
filename
.
lower
()
.
split
(
'.'
)[
-
1
]
in
[
'pdf'
,
'txt'
,
'doc'
]
:
return
True
elif
(
filename
.
lower
()
in
(
'readme'
,
'license'
,
'usage'
,
'categories'
)
):
elif
filename
.
lower
()
in
(
'readme'
,
'license'
,
'usage'
,
'categories'
):
return
True
return
False
def
main
():
# parse OPNsense external ACLs config
if
os
.
path
.
exists
(
acl_config_fn
):
...
...
@@ -248,27 +250,27 @@ def main():
os
.
mkdir
(
acl_target_dir
)
else
:
# remove index files
for
filename
in
glob
.
glob
(
'
%
s/*.index'
%
acl_target_dir
):
for
filename
in
glob
.
glob
(
'
%
s/*.index'
%
acl_target_dir
):
os
.
remove
(
filename
)
# read config and download per section
cnf
=
ConfigParser
()
cnf
.
read
(
acl_config_fn
)
for
section
in
cnf
.
sections
():
target_filename
=
acl_target_dir
+
'/'
+
section
if
cnf
.
has_option
(
section
,
'url'
):
target_filename
=
acl_target_dir
+
'/'
+
section
if
cnf
.
has_option
(
section
,
'url'
):
# collect filters to apply
acl_filters
=
list
()
if
cnf
.
has_option
(
section
,
'filter'
):
for
acl_filter
in
cnf
.
get
(
section
,
'filter'
)
.
strip
()
.
split
(
','
):
if
cnf
.
has_option
(
section
,
'filter'
):
for
acl_filter
in
cnf
.
get
(
section
,
'filter'
)
.
strip
()
.
split
(
','
):
if
len
(
acl_filter
.
strip
())
>
0
:
acl_filters
.
append
(
acl_filter
)
# define target(s)
targets
=
{
'domain'
:
{
'filename'
:
target_filename
,
'handle'
:
None
,
'class'
:
DomainSorter
}}
targets
=
{
'domain'
:
{
'filename'
:
target_filename
,
'handle'
:
None
,
'class'
:
DomainSorter
}}
# only generate files if enabled, otherwise dump empty files
if
cnf
.
has_option
(
section
,
'enabled'
)
and
cnf
.
get
(
section
,
'enabled'
)
==
'1'
:
download_url
=
cnf
.
get
(
section
,
'url'
)
if
cnf
.
has_option
(
section
,
'enabled'
)
and
cnf
.
get
(
section
,
'enabled'
)
==
'1'
:
download_url
=
cnf
.
get
(
section
,
'url'
)
acl
=
Downloader
(
download_url
,
acl_max_timeout
)
all_filenames
=
list
()
for
filename
,
line
in
acl
.
download
():
...
...
@@ -296,11 +298,12 @@ def main():
continue
if
filetype
in
targets
and
targets
[
filetype
][
'handle'
]
is
None
:
targets
[
filetype
][
'handle'
]
=
targets
[
filetype
][
'class'
](
targets
[
filetype
][
'filename'
],
'wb'
)
targets
[
filetype
][
'handle'
]
=
targets
[
filetype
][
'class'
](
targets
[
filetype
][
'filename'
],
'wb'
)
if
filetype
in
targets
:
targets
[
filetype
][
'handle'
]
.
write
(
'
%
s
\n
'
%
line
)
targets
[
filetype
][
'handle'
]
.
write
(
'
%
s
\n
'
%
line
)
# save index to disc
with
open
(
'
%
s.index'
%
target_filename
,
'wb'
)
as
idx_out
:
with
open
(
'
%
s.index'
%
target_filename
,
'wb'
)
as
idx_out
:
index_data
=
dict
()
for
filename
in
all_filenames
:
if
len
(
filename
.
split
(
'/'
))
>
2
:
...
...
@@ -313,7 +316,7 @@ def main():
for
filetype
in
targets
:
if
targets
[
filetype
][
'handle'
]
is
not
None
:
targets
[
filetype
][
'handle'
]
.
close
()
elif
cnf
.
has_option
(
section
,
'enabled'
)
and
cnf
.
get
(
section
,
'enabled'
)
!=
'1'
:
elif
cnf
.
has_option
(
section
,
'enabled'
)
and
cnf
.
get
(
section
,
'enabled'
)
!=
'1'
:
if
os
.
path
.
isfile
(
targets
[
filetype
][
'filename'
]):
# disabled, remove previous data
os
.
remove
(
targets
[
filetype
][
'filename'
])
...
...
@@ -321,5 +324,7 @@ def main():
# no data fetched and no file available, create new empty file
with
open
(
targets
[
filetype
][
'filename'
],
'wb'
)
as
target_out
:
target_out
.
write
(
""
)
# execute downloader
main
()
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment