This is a only exception while all other libraries are dynamically
linked.
The "TODO" does not make any sense as libcap provides both dynamic
and static libraries.
There is not reason to explain why it needs to be statically linked
'especially' as dynamic linking works very well.
Then just make it dynamically-linked by default, which also matches
the behaviour on autotools.
Compiled and tested on aarch64, no errors found.
Signed-off-by: Tianling Shen <i@cnsztl.eu.org>
* moved dev to version 3.1.1
* introduced n2n-portfwd tool
* moved port-forwarding source to tool
* adjusted tiny things, seemingly works with ./configure's --enable-miniupnp and --enable-natpmp
* adjusted for Windows
* wished for better typing skills
* applied some finishing touch
* typo
* moved dev to version 3.1.1
* laid ground for n2n-route tool
* adapted code style
* overhauled n2n-route's tool program logic and removed route code from edge
* added missing initialization of federation's purgeable field
* lifted un/purgeable confusion
* added warning about removed -n cli option
* realized that Windows does not offer inet_aton()
* removed -n option documentation from edge's man page
* slightly simplified n2n-route program logic
* applied more logic changes to n2n-route tool
* added 'info' read command to edge's management port
* corrected indention
* added Linux route control to n2n-route tool
* temporarily restricted n2n-route tool to Linux only
* We must be over the routing!
* pulled default gateway change detection into main loop to cover mobile use, and added devstr_t type
* corrected use of new UNPURGEABLE - so far gone unnoticed
* addresses possiible address issue
* i broke it
* reverted bad ideas
* added command line options, help text, and prevented vpn gateway being used for supernode/peer traffic routes
* added option to manually provide default gateway, also verbosity options
* getting there
* added option to limit networks to be routed (-n), adapted documentation
* fine-tuned minor things
* Fix cmake informational log messages
* If the correct library name is used find_library works better
* Re-enable remaining find_library users
* Reorder cmake to make libpcap detection work
For a moment there, I was starting to think that I might have found
a good feature in cmake, but then it turned out that it had specific
ordering requirements and strange cryptic functions that error out
in mysterious ways.
If I am going to have an antique build engine with special quirks that
I must learn in order to use it, I may as well choose the one that is
not also trying to be clever and hiding its internal magic from me.
Which means that I still prefer Makefiles - they are more debuggable.
* Ensure all options start off for cmake too
* Attempt to make the build time config logic in the cmake process match the makefile process
* Add a simple muscle memory helper for cmake
* Ask them what they are smoking, then tell them to put it out
* Avoid occasional stale package cache errors
* UPnP port redirection is supported.
* compile fixes
* compile fix
* optimize reconnection code
* prepared upnp threadification to counter main loop stall at supernode change
* NAT-PMP port forwarding support, temporarily merge codes to resolve conflicts.
* make compile fix
* prepared threadification in more detail
* adopted threadification to new file setup
* cleaned up
* renamed functions and data structures
* fixes
* differentiated between miniupnp and natpmp and added corresponding lib support to makefile
* name
* commented unused header includes
* comments
* license
* fixes
* fixes
* fixes
* NAT-PMP is already available.
* added CLI parameter to disable port forwarding if required
* preliminary made use of multithreading
* adjusted log level
* added man page documentation
* def'ed conf
* made pmpnat adjustments
Co-authored-by: fengdaolong <fengdaolong@gmail.com>
* Make it clear which hash we are testing
* Coverage for all pearson hash variants
* Initial coverage for auth.c
* Add some more auth.c coverage
* Ensure new bins are added to both build systems
* Ensure new source files are fully linted
* Factor build packages out into a more maintainable list
* Create a location for scripts to live
* Provide a make target to return the source dir as close as reasonable to the original distributed state
* Add a code lint step, checking the coding style
* Change test harness as recommended by shellcheck
* Ensure we actually have the linter tool installed
* Use the correct directory for cmake to run the tests
* Adjust for the older uncrustify in the current github ubuntu-latest
* Make one file pass the linter
* Integrate the lint with the existing test workflow
* Add files with minimal changes needed to the linter
* Add more files with minimal changes needed to the linter
* Dont build binaries if we fail the lint test
* Update the phony targets with the lint steps
* Ensure the flake8 package is installed in the new lint workflow job
* Use the makefile to drive the packages needed to install for linting
* No need to add dependancies on lint, just rely on the workflow status to show failure
* Update the scripts dir README to reflect current assumptions
* Rename and briefly document the indent.sh script
* Fix the ignore to ignore the right Makefile
* Rename the test_harness script to make it clear it is a shell script
* Provide a master lint make target and add a shell script lint tool
* Elminate stray tabs
* Drop include/auth.h from linter - there are inconsistant results with function definitions when using the current uncrustify rules
* Ensure that recent code additions pass the linter
* Include some of the more obviously correct lint fixes to edge_utils.c
* Refactor edge JSON api into its own source file
* Use shorter names for static management functions
* Implement a JSON RPC way of managing the verbosity
* Tidy up help display in n2nctl script
* Make note of issue with implementing the stop command
* Implement a JSON RPC call to fetch current community
* Make n2nhttpd time value be more self-contained
* Make n2nhttpd order more closely match the existing management stats output
* Wire up status page to the verbosity setting
* Add JSON versions of the remainder of the edge management stats
* Add new file to cmake
* Properly define management handler
* Only update the last updated timestamp after a successful data fetch
* Function and types definition cleanup
* Force correct type for python scripts mgmt port
* Implement initial JSON API for supernode
* Fix whitespace error
* Use helper function for rendering peers ip4 address
* Proxy the auth requirement back out to the http client, allowing normal http auth to be used
* Ensure that we do not leak the federation community
* Use the same rpc method name and output for both edge and supernode for peers/edges
* Allow n2nctl to show raw data returned without resorting to tricks
* Make n2nctl pretty printer understandable with an empty table
* Use the full name for supernodes RPC call
* Use same RPC method name (but some missing fields) for getting communities from both edge and supernode
* Add *_sup_broadcast stats to edge packet stats output
* Refacter the stats into a packetstats method for supernode RPC
* Even if I am not going to prettyprint the timestamps, at least make all the timestamps on the page the same unit
* Simplify the RPC handlers by flagging some as writable and checking that in the multiplexer
* Remove invalid edges data
* Avoid crash on bad data to verbose RPC
* Avoid showing bad or inconsistant protocol data in communities RPC
* Minor clarification on when --write is handled
* Make linter happy
* Fix changed method name in n2nhttpd
* Move mainloop stop flag into the n2n_edge_t structure, allowing access from management commands
* Implement edge RPC stop command
* Move mainloop stop flag into the n2n_sn_t structure, allowing access from management commands
* Implement supernode RPC stop command
* Allow multiple pages to be served from mini httpd
* Extract common script functions into a separate URL
* Handle an edge case in the python rpc class
With a proper tag-based demultiplexer, this case should be a nop,
but we are single-threaded and rely on the packet ordering in this
library.
* Add n2nhttpd support to query supernode using urls prefixed with /supernode/
* Handle missing values in javascript table print
* Add another less filtering javascript key/value renderer
* Add a supernode.html page to the n2nhttpd
* Address lint issue
* Mention the second html page on the Scripts doc
* Remove purgable column from supernode edges list - it looks like it is rarely going to be set
* Add a simple one-line example command at the top of the API documentation
* Acknowledge that this is not the most efficient protocol, but point out that it was not supposed to be
* Make it clear that the n2nctl script works for both edge and supernode
* Fight with inconsistant github runner results
* Turn off the /right/ coverage generator
* Add a simple test framework
* Add a code coverage report example oneliner
* Move the coverage report into a separate directory
* Add a github action to run tests and publish a branch with the coverage report
* Fix: Missing job separator
* Fix: remember to actually run configure
* Fix: Gotta autogen before I configure
* Dont try to upload coverage report unless this is a push
* Clearly show the git ref tested in the coverage report
* Add a test for the various transforms
* Add tests for the elliptic curve and pearson hash
* Ensure we ignore new generated output
* Remove unneeded boilerplate from the compression tests
* Add an example of a test of the encoded wire packets
* Ensure that correctly testable data is output even when zstd is not compiled
* Factor test runner out into its own script and attempt to add it to the cmake file
* Tell cmake about a new object file
* Stop trying to make Cmake work...
* Stop trying to make cmake work, round 2
* In the middle of a thousand lines of cmake output was one important one - windows could not find assert() - try again
* Try again to plumb the tests into cmake
* Add missing library to our superset install line
* Fix build error when libcap-dev is installed
* Switch to using artifact uploads instead of pages to store/show the coverage report
* Fix artifact upload yaml
* Upload coverage report to codecov
* Fix codecov - clearly it doesnt do a recursive search for coverage files
* Fix codecov - my hopeful use of a list of directories didnt work
* Fix codecov - unfortunately, it doesnt just consume the coverage data and needs us to generate the gcov output
* Fix codecov - nope, it still doesnt recursively search
* Fix codecov - it really helps if I run the gcov data generator
* Add a simple matrix build
* Fix older ubuntu versions of gcovr that do not support the '--html-title' option
* Ensure we use gcover options that are identical on older ubuntu
* Improve coverage generation and required build packages
* Move the makefiles towards a clearer arrangement
* Break circular make dependancy and start using implicit rules for most compilation
* Fix old cmake to compile again after renames
* Add new file 'network_traffic_filter.c/.h"
* Add feature to drop or accept specific packet transmit over edge network interface by rules.
* fix CMakeLists.txt typo
* Update Rule String Format
* replace -F (filter) with -R (rule) for traffic restrictions.
* Update edge help (-h) message. Update documents.
* Pass Build on CircleCI and local Windows 10 VS2019
* Fix cmake build failed because of sn_selection
* fix operate void* cause vs2016 build failed
* Fix typo to pass build on windows.
* add inttypes.h for n2n_typedefs.h to pass windows build
* modify headers to pass build on windows.
* Add new file 'network_traffic_filter.c/.h"
* Add feature to drop or accept specific packet transmit over edge network interface by rules.
* fix CMakeLists.txt typo
* Update Rule String Format
* replace -F (filter) with -R (rule) for traffic restrictions.
* Update edge help (-h) message. Update documents.
* add cmake export commands
* add extra edge description (hint) field
* add peer2peer description registry
* fixed bug: peer2peer not pass the dev_desc filed
* update -I option in the helper text
Co-authored-by: Longtao Wang <longtao.wang@cortexlabs.ai>