Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
29 changes: 13 additions & 16 deletions .github/workflows/scorecards-analysis.yml
Original file line number Diff line number Diff line change
Expand Up @@ -17,39 +17,36 @@ jobs:
permissions:
# Needed to upload the results to code-scanning dashboard.
security-events: write
actions: read
contents: read
id-token: write

steps:
- name: "Checkout code"
uses: actions/checkout@v3.5.3
with:
persist-credentials: false
uses: actions/checkout@b4ffde65f46336ab88eb53be808477a3936bae11 # v4.1.1

- name: "Run analysis"
uses: ossf/scorecard-action@e38b1902ae4f44df626f11ba0734b14fb91f8f86 # v1.0.1
uses: ossf/scorecard-action@0864cf19026789058feabb7e87baa5f140aac736 # v2.3.1
with:
results_file: results.sarif
results_format: sarif
# Read-only PAT token. To create it,
# follow the steps in https://github.com/ossf/scorecard-action#pat-token-creation.
repo_token: ${{ secrets.SCORECARD_READ_TOKEN }}
# Publish the results to enable scorecard badges. For more details, see
# https://github.com/ossf/scorecard-action#publishing-results.
# For private repositories, `publish_results` will automatically be set to `false`,
# regardless of the value entered here.
# Scorecard team runs a weekly scan of public GitHub repos,
# see https://github.com/ossf/scorecard#public-data.
# Setting `publish_results: true` helps us scale by leveraging your workflow to
# extract the results instead of relying on our own infrastructure to run scans.
# And it's free for you!
publish_results: true

# Upload the results as artifacts (optional).
# https://docs.github.com/en/actions/advanced-guides/storing-workflow-data-as-artifacts
# Optional.
- name: "Upload artifact"
uses: actions/upload-artifact@0b7f8abb1508181956e8e162db84b466c27e18ce
uses: actions/upload-artifact@26f96dfa697d77e81fd5907df203aa23a56210a8 # v3
with:
name: SARIF file
path: results.sarif
retention-days: 5

# Upload the results to GitHub's code scanning dashboard.
- name: "Upload to code-scanning"
uses: github/codeql-action/upload-sarif@v2 # v1.0.26
- name: "Upload SARIF results"
uses: github/codeql-action/upload-sarif@cdcdbb579706841c47f7063dda365e292e5cad7a # v1
with:
sarif_file: results.sarif
2 changes: 1 addition & 1 deletion src/detect.c
Original file line number Diff line number Diff line change
Expand Up @@ -1218,7 +1218,7 @@ static bool DetectRunTxInspectRule(ThreadVars *tv,
} else if ((inspect_flags & DE_STATE_FLAG_FULL_INSPECT) == 0 && mpm_in_progress) {
TRACE_SID_TXS(s->id, tx, "no need to store no-match sig, "
"mpm will revisit it");
} else if (inspect_flags != 0 || file_no_match != 0) {
} else {
TRACE_SID_TXS(s->id, tx, "storing state: flags %08x", inspect_flags);
DetectRunStoreStateTx(scratch->sgh, f, tx->tx_ptr, tx->tx_id, s,
inspect_flags, flow_flags, file_no_match);
Expand Down
12 changes: 6 additions & 6 deletions src/util-streaming-buffer.c
Original file line number Diff line number Diff line change
Expand Up @@ -123,7 +123,7 @@ StreamingBufferBlock *SBB_RB_FIND_INCLUSIVE(struct SBB *head, StreamingBufferBlo
* \brief does data region intersect with list region 'r'
* Takes the max gap into account.
*/
static inline bool RegionsIntersect(const StreamingBuffer *sb, const StreamingBufferConfig *cfg,
static inline bool RegionsIntersect(const StreamingBufferConfig *cfg,
const StreamingBufferRegion *r, const uint64_t offset, const uint64_t re)
{
/* create the data range for the region, adding the max gap */
Expand Down Expand Up @@ -158,7 +158,7 @@ static StreamingBufferRegion *FindFirstRegionForOffset(const StreamingBuffer *sb

StreamingBufferRegion *p = NULL;
for (; r != NULL; r = r->next) {
if (RegionsIntersect(sb, cfg, r, offset, data_re) == true) {
if (RegionsIntersect(cfg, r, offset, data_re) == true) {
*prev = p;
return r;
}
Expand All @@ -182,7 +182,7 @@ static StreamingBufferRegion *FindLargestRegionForOffset(const StreamingBuffer *
SCLogDebug("checking: %p/%" PRIu64 "/%" PRIu64 ", offset %" PRIu64 "/%" PRIu64, r,
r->stream_offset, reg_re, offset, data_re);
#endif
if (!RegionsIntersect(sb, cfg, r, offset, data_re))
if (!RegionsIntersect(cfg, r, offset, data_re))
return candidate;

if (r->buf_size > candidate->buf_size) {
Expand All @@ -200,7 +200,7 @@ static StreamingBufferRegion *FindRightEdge(const StreamingBuffer *sb,
const uint64_t data_re = offset + len;
StreamingBufferRegion *candidate = r;
for (; r != NULL; r = r->next) {
if (!RegionsIntersect(sb, cfg, r, offset, data_re)) {
if (!RegionsIntersect(cfg, r, offset, data_re)) {
SCLogDebug(
"r %p is out of scope: %" PRIu64 "/%u/%" PRIu64, r, offset, len, offset + len);
return candidate;
Expand Down Expand Up @@ -1433,11 +1433,11 @@ static StreamingBufferRegion *BufferInsertAtRegion(StreamingBuffer *sb,
data_offset + data_len);
ListRegions(sb);

if (RegionsIntersect(sb, cfg, &sb->region, data_offset, data_offset + data_len)) {
if (RegionsIntersect(cfg, &sb->region, data_offset, data_offset + data_len)) {
SCLogDebug("data_offset %" PRIu64 ", data_len %u intersects with main region (next %p)",
data_offset, data_len, sb->region.next);
if (sb->region.next == NULL ||
!RegionsIntersect(sb, cfg, sb->region.next, data_offset, data_offset + data_len)) {
!RegionsIntersect(cfg, sb->region.next, data_offset, data_offset + data_len)) {
SCLogDebug(
"data_offset %" PRIu64
", data_len %u intersects with main region, no next or way before next region",
Expand Down